2-Mary Dalrymple, John J. Lowe, Louise Mycock - The Oxford Reference Guide To Lexical Functional Grammar-Oxford University Press (2019)
2-Mary Dalrymple, John J. Lowe, Louise Mycock - The Oxford Reference Guide To Lexical Functional Grammar-Oxford University Press (2019)
The Oxford
Reference Guide
to Lexical Functional
Grammar
MARY DALRYMPLE
JOHN J. LOWE
LOUISE MYCOCK
1
3
Great Clarendon Street, Oxford, ox2 6dp,
United Kingdom
Oxford University Press is a department of the University of Oxford.
It furthers the University’s objective of excellence in research, scholarship,
and education by publishing worldwide. Oxford is a registered trade mark of
Oxford University Press in the UK and in certain other countries
© Mary Dalrymple, John J. Lowe, and Louise Mycock 2019
The moral rights of the authors have been asserted
First Edition published in 2019
Impression: 1
All rights reserved. No part of this publication may be reproduced, stored in
a retrieval system, or transmitted, in any form or by any means, without the
prior permission in writing of Oxford University Press, or as expressly permitted
by law, by licence or under terms agreed with the appropriate reprographics
rights organization. Enquiries concerning reproduction outside the scope of the
above should be sent to the Rights Department, Oxford University Press, at the
address above
You must not circulate this work in any other form
and you must impose this same condition on any acquirer
Published in the United States of America by Oxford University Press
198 Madison Avenue, New York, NY 10016, United States of America
British Library Cataloguing in Publication Data
Data available
Library of Congress Control Number: 2018965641
ISBN 978–0–19–873330–0
Printed and bound by
CPI Group (UK) Ltd, Croydon, CR0 4YY
Links to third party websites are provided by Oxford in good faith and
for information only. Oxford disclaims any responsibility for the materials
contained in any third party website referenced in this work.
Preface
This book is based on work first published by Dalrymple (2001). Some chapters
from that book appear in this one, although the entire text has been overhauled
and revised. Some of the chapters in Part II of this book are entirely new. All in all,
this is not a new edition of the 2001 book, but a new book which includes parts of
the previous one.
As usual, the LFG community has been incredibly supportive of our work
in producing this book, and we are grateful to the many people who provided
comments, feedback, and support. We begin by conveying our thanks to Luke
Carr, Jamie Findlay, and Miltiadis Kokkonidis for helpful comments on Dalrymple
(2001).
We would like to single out three heroic individuals for special praise for their
dedication and thoroughness in reading and commenting on multiple versions of
the second edition as it took shape. Bozhil Hristov, Adam Przepiórkowski, and
Amanda Thomas each read through several versions of the entire book, giving
valuable comments and feedback each time. We have been particularly impressed
by Adam’s ability to detect problems from the microscopic to the macroscopic level.
We hope they will feel that the final version of the book reflects their hard work
and the helpful comments and suggestions that they made.
We are also grateful to those of our readers who commented in detail on entire
sections of the book. For detailed and helpful comments on the chapters in Part I,
we are grateful to Ash Asudeh, Alex Biswas, Ken Kahn, Leslie Lee, Joey Lovestrand,
Agnieszka Patejuk, Liselotte Snijders, and Martin Trpovski. For comments on Part
II, we are grateful to Ash Asudeh, Alex Biswas, Leslie Lee, and Joey Lovestrand.
We are also grateful to readers who provided comments on individual chapters
of the book. In Part I: Miriam Butt for Chapter 1 (Background and theoretical
assumptions); I Wayan Arka, Agnieszka Patejuk, and Ida Toivonen for Chapter 2
(Functional structure); Ida Toivonen for Chapter 3 (Constituent structure); Oleg
Belyaev and Ron Kaplan for Chapter 5 (Describing syntactic structures); and Ron
Kaplan for particularly detailed and helpful comments on Chapter 6 (Syntactic
relations and syntactic constraints). In Part II: Miriam Butt for Chapter 7 (Beyond
c-structure and f-structure: linguistic representations and relations) and Chapter
9 (Argument structure and mapping theory); Tina Bögel, Stephen Jones, and
Aditi Lahiri for Chapter 11 (Prosodic structure); and Miriam Butt, Louisa Sadler,
and Andy Spencer for Chapter 12 (The interface to morphology). In Part III:
Doug Arnold, Ash Asudeh, and Andy Morrison for Chapter 13 (Modification);
Dag Haug for Chapter 14 (Anaphora) and Chapter 15 (Functional and anaphoric
control); Doug Arnold, Oleg Belyaev, Dag Haug, John Lamping, Louisa Sadler,
and Vijay Saraswat for Chapter 16 (Coordination); Doug Arnold, Ash Asudeh,
Ron Kaplan, and Louisa Sadler for Chapter 17 (Long-distance dependencies); Ash
Asudeh, Aaron Broadwell, Kersti Börjars, Ron Kaplan, Peter Sells, Nigel Vincent,
and Andy Way for Chapter 18 (Related research threads and new directions); and
xviii preface
Liselotte Snijders for the bibliography. This book has been long in the making, and
we apologize if we have omitted reference to anyone whose comments we have
benefited from but who is not mentioned here.
For administrative support, we gratefully acknowledge the Faculty of Linguis-
tics, Philology, and Phonetics at the University of Oxford. We thank Julia Steer and
Vicki Sunter of OUP for very helpful advice, and for their patience over the long
course of the gestation of this book.
Mary Dalrymple is grateful to the following organizations for research support:
2011–14: Ministerio de Ciencia e Innovación, Gobierno de España: “The syntax
and information structure of unbounded dependencies,” principal investigator
Prof. Alexandre Alsina, Universitat Pompeu Fabra, Barcelona); 2012: British
Academy Small Grant “Plural semantics in Austronesian”; 2012–13: Leverhulme
Research Fellowship “Plurals: Morphology and Semantics,” RF-2012-295;
2017–2018: Centre for Advanced Study at the Norwegian Academy of Science
and Letters, “SynSem: From Form to Meaning—Integrating Linguistics and
Computing,” principal investigators Prof. Dag Haug and Prof. Stephan Oepen.
She is also grateful to Ken Kahn for all kinds of support, and to her sister Matty
for being an inspiration as a writer.
John Lowe is grateful to the following organizations for research support: 2012–
15: Leverhulme Early Career Fellowship (ECF-2012-081); 2013–14: Ministerio de
Ciencia e Innovación, Gobierno de España: “The syntax and information structure
of unbounded dependencies,” principal investigator Prof. Alexandre Alsina, Uni-
versitat Pompeu Fabra, Barcelona); 2016: Jill Hart Fund for Indo-Iranian Philology,
University of Oxford. He is also grateful to his family for being a source of strength
and inspiration, especially to Helen, Henry, and Wilfred.
Louise Mycock is grateful to the following for research support: 2013–14: Min-
isterio de Ciencia e Innovación, Gobierno de España: “The syntax and informa-
tion structure of unbounded dependencies,” principal investigator Prof.Alexandre
Alsina, Universitat Pompeu Fabra, Barcelona). She is also grateful to her family
for their ongoing support: her husband Chi Lun Pang, her grandparents Iris and
Brian, her dad Dave, her sisters Kathryn and Fiona and their partners, her nephew
Charlie, and her nieces Lily and Maisie.
List of Abbreviations
We have modified the glosses from our source materials for consistency with the
Leipzig Glossing Rules (Bickel et al. 2015), a simplified version of which we use
throughout the book. We use the following abbreviations:
1 first person
2 second person
3 third person
abs absolutive
acc accusative
active active
av active voice
aux auxiliary
b Chechen gender class B
caus causative
comp complementizer
d Chechen gender class D
dv dative voice
en Catalan en
erg ergative
excl exclusive
f feminine
fut future
fv final vowel
gen genitive
hi Catalan hi
incl inclusive
indf indefinite
inf infinitive
iprf imperfect
iv instrumental voice
link Tagalog linker
loc locative
m masculine
n neuter
nfut nonfuture
xx list of abbreviations
nom nominative
nonpl nonplural
nonsg nonsingular
npst nonpast
ov objective voice
pass passive
pfv perfective
pl plural
postneg postnegation
pot potential
prep preposition
prf perfect
prs present
pst past
q question particle
refl reflexive
scv Chechen simultaneous converb
sg singular
vm Hungarian verbal modifier
wp Chechen witnessed past
1
Background and theoretical
assumptions
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
2 background and theoretical assumptions
The existing transformational framework would not have led to the prediction that
transformations would operate in this way. Since transformations were not con-
strained as to the output structure they produced, it was surprising that they would
produce structures like those that the basic grammar could otherwise generate.
This important finding had wide-reaching implications: the basic phrase structure
of languages is invariant, and the application of particular transformations does
not alter this basic phrase structure.
Why should so many transformations have been structure-preserving in this
sense? Bresnan (1978) made the key observation: all structure-preserving trans-
formations can be reformulated as lexical redundancy rules. According to this view,
operations on the abstract syntactic argument structure of a lexical item produce
a new syntactic argument structure, with a surface form that is realized in an
expected way by a basic phrase structure grammar. This allowed an abstract and
uniform crosslinguistic characterization of argument alternations like the active-
passive relation, while also allowing for a theory of crosslinguistic similarities and
differences in the phrasal expression of the different alternations.
With this, the need emerged for a theory allowing simultaneous expression
of both the phrasal constituency of a sentence and its more abstract functional
syntactic organization. The formal insights leading to the development of Lexi-
cal Functional Grammar arose originally from the work of Woods (1970), who
explored methods for representing the surface constituent structure of a sentence
together with more abstract syntactic information. Building on this work, Kaplan
(1975a,b, 1976) realized that placing certain constraints on the representation of
abstract syntactic structure and its relation to surface phrasal structure would
lead to a simple, formally coherent, and linguistically well-motivated grammatical
architecture. Based on these formal underpinnings, the relation of the abstract
functional syntactic structure of a sentence to its phrase structure could be fully
explored. More information about the historical development of the theory is
provided by Dalrymple et al. (1995a) and Bresnan et al. (2016).
functional structure, and the relation between them. A clear understanding of the
concepts presented in Chapter 5 is essential for the discussion in the rest of the
book. Chapter 6 is best thought of as a compendium of relatively more advanced
formal tools and relations, and may be most profitably used as a reference in
understanding the analyses presented in the rest of the book.
The second part of the book explores nonsyntactic levels of linguistic struc-
ture and the modular architecture of LFG. Chapter 7 (Beyond c-structure and
f-structure: Linguistic representations and relations) sets the scene for our explo-
ration of other linguistic levels and their relation to constituent structure and
functional structure, presenting LFG’s projection architecture and outlining how
different grammatical levels are related to one another. Chapter 8 (Meaning and
semantic composition) introduces the LFG view of the syntax-semantics interface
and semantic representation, according to which the meaning of an utterance
is determined via logical deduction from a set of premises associated with the
subparts of the utterance. Chapter 9 (Argument structure and mapping theory)
discusses the content and representation of argument structure, its relation to
syntax, and its role in determining the syntactic functions of the arguments of a
predicate. Chapter 10 (Information structure) introduces the level of information
structure, the structuring of an utterance in context, and explores the relation of
information structure to other linguistic levels. Chapter 11 (Prosodic structure)
introduces the level of prosodic structure, which analyzes the string in parallel with
constituent structure, but with respect to prosodic units rather than phrasal units.
Chapter 12 (The interface to morphology) discusses the place of morphology in
the architecture of LFG, showing how a realizational theory of morphology can be
integrated in an LFG setting.
The third part of the book illustrates the concepts of the theory more explicitly
by presenting a series of sketches of the syntax and semantics of a range of repre-
sentative linguistic phenomena. We present the syntactic aspects of the analyses
separately from the semantic aspects, so readers who are not interested in formal
semantic analysis should still be able to profit from the syntactic discussion in these
chapters. In this part, we often leave aside analysis of the information structure,
prosody, and morphology of these phenomena, though we sometimes include an
analysis of these other aspects as well, in line with the increasing awareness of the
importance of adopting a holistic approach and taking account of the interplay
of linguistic modules in a full account of the data. Chapter 13 (Modification)
discusses the syntax and semantics of modifiers, particularly concentrating on
modification of nouns by adjectives. Chapter 14 (Anaphora) presents a theory
of the syntax and semantics of anaphoric binding, including both intrasentential
and intersentential anaphora. Chapter 15 (Functional and anaphoric control)
discusses constructions involving control, where the referent of the subject of a
subordinate clause is determined by lexical or constructional factors. Chapter 16
(Coordination) presents an analysis of aspects of the syntax and semantics of coor-
dination, and Chapter 17 (Long-distance dependencies) discusses long-distance
dependencies in topicalization, relative clause formation, and question formation.
The final chapter of the book, Chapter 18 (Related research threads and new
directions), discusses LFG-based work in areas not covered elsewhere in the book,
as well as new developments in the theory of LFG, including work in historical
how to use the book 5
• For the semantics chapter in Part II (Chapter 8) and the semantics sections of
the chapters in Part III, Gamut (1991a,b) and Partee et al. (1993: Chapter 7)
provide useful background.
• Chapter 10 discusses information structure, its representation, and its place
in the overall LFG architecture. There is some discussion of information
structure in Chapter 17, but it should be possible to follow almost all of the
discussion in Chapter 17 even without familiarity with the material presented
in Chapter 10. For an overview and introduction to information structure, see
Lambrecht (1994) and Erteschik-Shir (2007: Chapters 1–3).
• The content and representation of prosodic structure is discussed in
Chapter 11, but does not figure in the analyses presented in Part III. For
an introduction to the concepts discussed in Chapter 11, see Selkirk (1984),
Nespor and Vogel (2007), and Ladd (2008).
• The analyses presented in Part III also do not include morphological analysis,
and so the morphology chapter in Part II (Chapter 12) can be skipped by
6 background and theoretical assumptions
those who are not concerned with morphology and its interface with the rest
of the grammar. Spencer (2004) and Haspelmath and Sims (2011) provide a
solid introduction to morphology, and Stewart (2015) provides an overview
of contemporary morphological theories. Stump (2001: Chapter 1) is an
introduction to issues in morphological theory with a focus on the word-
and-paradigm model, providing a theoretical underpinning for the family of
realizational theories which that chapter adopts.
LFG assumes two different ways of representing syntactic structure, the constituent
structure or c-structure and the functional structure or f-structure. Within the
overall system of linguistic structures, these constitute two subsystems. Functional
structure is the abstract functional syntactic organization of the sentence, famil-
iar from traditional grammatical descriptions, representing syntactic predicate-
argument structure and functional relations like subject and object. Constituent
structure is the overt, more concrete level of linear and hierarchical organization
of words into phrases.
Section 2.1 presents motivation for the categories and information appearing
in functional structure and outlines some common characteristics of functional
structure categories. Section 2.2 demonstrates that grammatical functions are best
treated as primitive concepts, as they are in LFG, rather than defined in terms
of morphological or phrase structure concepts. Section 2.3 shows that syntac-
tic subcategorization requirements, the array of syntactic arguments required by
a predicate, are best stated in functional terms. The formal representation of
functional structure and constraints on f-structure representations are discussed
in §2.4. Finally, §2.5 provides an overview of the content and representation of
functional structure features.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
10 functional structure
However, the subject phrase in such a question may not appear in sentence-initial
position if the subordinate clause begins with a word like that, though this is
possible for the object phrase:1
In fact, the subject-object distinction is only one aspect of a rich set of distinc-
tions among grammatical functions. Keenan and Comrie (1977: 66) propose a
more fine-grained analysis of abstract grammatical structure, the Keenan-Comrie
Hierarchy or Accessibility Hierarchy, stating that “the positions on the Accessibility
Hierarchy are to be understood as specifying a set of possible grammatical dis-
tinctions that a language may make.” They claim that the hierarchy is relevant
for various linguistic constructions, including relative clause formation, where
it restricts the grammatical function of the argument in the relative clause that
is interpreted as coreferent with the modified noun. The border between any
two adjacent grammatical functions in the hierarchy can represent a distinction
between acceptable and unacceptable relative clauses in a language, and different
languages can set the border at different places on the hierarchy:2
In some languages, the hierarchy distinguishes subjects from all other grammatical
functions: only the subject of a relative clause can be relativized, or interpreted as
coreferent with the noun modified by the relative clause. Other languages allow rel-
ativization of subjects and objects in contrast to other grammatical functions. This
more fine-grained hierarchical structure allows further functional distinctions to
emerge.
Keenan and Comrie speculate that their hierarchy plays a role in the analysis
of other grammatical constructions besides relative clause formation, and indeed
Comrie (1974) applies the hierarchy in an analysis of grammatical functions in
causative constructions. In fact, the Keenan-Comrie hierarchy closely mirrors the
“relational hierarchy” of Relational Grammar, as given by Bell (1983), upon which
much work in Relational Grammar is based:
The labels obj𝜃 and obl𝜃 represent families of relations indexed by thematic roles,
with the 𝜃 subscript representing the thematic role associated with the argument.
For instance, objtheme is the member of the group of thematically restricted obj𝜃
functions that bears the thematic role theme, and oblsource and oblgoal are
members of the obl𝜃 group of grammatical functions filling the source and goal
thematic roles. These clearly stated relations between thematic roles and grammat-
ical functions reflect a broader generalization about grammatical functions within
the LFG framework. As Bresnan et al. (2016: 9) put it, grammatical functions are
the “relators” of phrase structure positions to thematic roles. Structural expressions
of a particular grammatical function vary crosslinguistically, but a characteristic
mapping to argument structure exists in all cases.
3 The grammatical function hierarchy given in (5) lists only grammatical functions that play a
role within the clause. We discuss the grammatical function poss, the function borne by prenominal
genitives within the English noun phrase, in §2.1.10.
12 functional structure
(7) a. The girl handed the baby a toy on Tuesday in the morning.
b. ⃰David saw Tony George Sally.
Alternation: Only arguments can alternate with subjects and objects; see
Needham and Toivonen (2011):
4 These tests do not give clear-cut results in all cases. For discussion of contentious cases, see §9.10.5.
grammatical functions and functional structure 15
In (10) and (11), the reflexive’s antecedent is the subject Jon, which is an argument
of the same verbal predicate. By contrast, when the reflexive appears as part of a
modifier phrase, as in (12), the sentence is ungrammatical.
In contrast, reordering arguments may affect the meaning of the sentence, focusing
attention on one or another argument, but does not alter the conditions under
which the sentence is true.
(14) a.⃰Which famous professor did Kim climb K2 without oxygen in order to
impress ?
b. Which famous professor did Kim attempt to impress by climbing K2
without oxygen?
This generalization is not as robust as those discussed above, since as Pollard and
Sag point out, it is possible to extract a phrase from some modifiers:
5 Relational grammar (Perlmutter and Postal 1983) also recognizes this basic division of grammatical
functions into “term relations” and “oblique relations.” Terms are also sometimes referred to as “core
functions” (Andrews 2007a; Bresnan et al. 2016).
16 functional structure
Control: Kroeger (1993) shows that in Tagalog, only a term can be the con-
trollee in the participial complement construction, and only a term can be a con-
troller in the participial adjunct construction. The sentences in (17) are examples
of the participial complement construction. The controllee can be a term (17a,b),
but never a nonterm (17c):
The clausal complement that Chris was yawning bears the grammatical function
comp in (19):
The xcomp function differs from the comp function in not containing an overt
subj internal to its phrase; xcomp is an open function, whose subj is determined
by means of lexical specifications on the predicate that governs it, as discussed in
§2.1.7. What is the termhood status of the xcomp and comp arguments?
Zaenen and Engdahl (1994) classify xcomp as a kind of oblique in their
analysis of the linking of sentential and predicative complements, though without
providing explicit evidence in support of this classification. Oblique arguments are
nonterms, and so if Zaenen and Engdahl are correct, xcomp would be classified as
a nonterm.
Word order requirements on infinitival and finite complements in English
provide some support for this position. Sag (1987) claims that in English, term
phrases always precede obliques:
Since the xcomp to be unhappy is not required to precede the oblique phrase to
Sandy but can appear either before or after it, Sag’s diagnostic indicates that the
xcomp must also be an oblique. Similar data indicate that comp is also an oblique
phrase:
Semantically unrestricted functions like subj and obj can be associated with any
thematic role, as Fillmore (1968) shows:
18 functional structure
The examples in (24) show that the subj of different verbs can be associated with
different thematic roles: agent in a sentence like He hit the ball, goal in a sentence
like He received a gift, and so on. Similar examples can be constructed for obj.
In contrast, members of the semantically restricted family of functions obj𝜃 and
obl𝜃 are associated with a particular thematic role. For example, the objtheme
function is associated only with the thematic role of theme, and the oblgoal
is associated with goal. Languages may differ in the inventory of semantically
restricted functions they allow. For example, English allows only objtheme :
Other thematic roles cannot be associated with the second object position, under-
lined in these examples:
2.1.5 SUBJ
The subject is the term argument that ranks the highest on the Keenan-Comrie
relativization hierarchy. As discussed in §2.1.1, Keenan and Comrie’s hierarchy is
applicable to a number of processes including relativization: if only one type of
argument can participate in certain processes for which a grammatical function
hierarchy is relevant, that argument is often the subject.
There is no lack of tests referring specifically to the subject function.
Agreement: The subject is often the argument that controls verb agreement in
languages in which verbs bear agreement morphology; indeed, Moravcsik (1978)
proposes the following language universal:
There is no language which includes sentences where the verb agrees with a constituent
distinct from the intransitive subject and which would not also include sentences where the
verb agrees with the intransitive subject. (Moravcsik 1978: 364)
Honorification: Matsumoto (1996) calls this the most reliable subject test in
Japanese. Certain honorific forms of verbs are used to honor the referent of the
subject:
The verb form o-V ni narimasu is used to honor the subject sensei ‘teacher’. It can-
not be used to honor a nonsubject, even if the argument is a “logical subject”/agent.
Like example (29), example (30) is grammatical because the honorific verb form is
used to honor the subject sensei ‘teacher’:
A proper name is used only when the referent is not superior to the speaker in
age or social rank: example (31) is unacceptable because Jon is not an appropriate
target of honorification, and sensei is not the subject:
grammatical functions and functional structure 21
Bell (1983: 154 ff.) shows that the same is true in Cebuano.
This is only a sampling of the various tests for subjecthood. Many other tests could,
of course, be cited (see, for example, Andrews 2007a; Falk 2006; Li 1976; Zaenen
1982; Zaenen et al. 1985).
The question of whether every clause has a subj is controversial. The Subject
Condition7 was discussed by Bresnan and Kanerva (1989), who attribute it origi-
nally to Baker (1983) (see also Levin 1987; Butt et al. 1999):
Though the Subject Condition seems to hold in English, and perhaps in many
other languages as well, there are languages in which the requirement does not so
clearly hold. For example, German impersonal passives, as in (35), are traditionally
analyzed as subjectless clauses:
However, Berman (1999, 2003) claims that clauses like (35) contain an unpro-
nounced expletive subject and thus that the Subject Condition is not violated. She
argues that in a German impersonal passive construction the verbal agreement
morphology (specifically the 3sg form) introduces a subject that has only person
and number agreement features: that is, an expletive subject without semantic
content. In this way, every finite clause in German satisfies the Subject Condition.
Kibort (2006) discusses Polish constructions which, similar to the German
construction in (35), appear to be subjectless. These clauses contain one of a small
number of inherently impersonal predicates, such as słychać ‘hear’, as in (36a), or
an intransitive predicate which has been passivized, as in (36b), which is similar to
the German example in (35). In (36a), the lack of a syntactic subject is reflected
in the form of the verb used: the verb słychać ‘hear’ has the same form as the
infinitive. Since this is a nonfinite verb form, there is no agreement morphology
to introduce subject features, in contrast to the German construction analyzed
by Berman. (36b) contains the passivized form of an intransitive predicate, było
sypane ‘has been spreading’, which bears default impersonal agreement (3sg.n).
The agent can optionally appear as an oblique argument:
On the basis of such data, Kibort (2006) argues that truly subjectless constructions
exist in Polish. It may be, then, that the Subject Condition is a language-particular
requirement imposed by some but not all languages, rather than a universal
requirement.
An alternative, more fine-grained view of subject properties and the notion of
subjecthood is explored in detail by Falk (2006). The core of Falk’s proposal is
that the grammatical function subject represents the conflation of two distinct
functions: the element that expresses the most prominent argument of the verb
(gf)
% and the element that connects its clause to other clauses in a sentence (pivot).
Prominence in the case of gf % is defined in terms of the mapping between the
thematic hierarchy (comprising thematic roles such as agent, theme, location) and
the relational hierarchy (comprising grammatical functions). Falk (2006: 39) thus
defines gf
% as being “the element with the function of expressing as a core argument
the hierarchically most prominent argument,” and argues that a subset of “subject
grammatical functions and functional structure 23
However, pivot and not gf % is the target of relativization in relative clause forma-
tion: the relativized argument is the single argument of the intransitive predicate
‘angry’ or the object of ‘kill’, but not the gf
% of ‘take’.
8 Falk attributes the examples in (37) and (38) to Manning (1996b). We present them in simplified
form, leaving out morphological detail.
24 functional structure
Falk (2006) proposes that gf% plays a role in all languages, in the sense that
all languages have clauses including gf% as an argument. However, he holds that
the requirement for all clauses to include gf
% as an argument holds only for some
languages, similar to Kibort’s (2006) claims in relation to the Subject Condition.
In particular, Falk (2006: 170) proposes that certain intransitive Acehnese verbs
(those with agentive subjects, such as “go”) require a gf, % while certain others
(those with nonagentive subjects, such as “fall”) require an object, not gf.
% pivot
has a different status: Falk tentatively proposes that some languages, including
Choctaw/Chickasaw and Warlpiri, lack a pivot altogether. Such languages would
be expected to lack pivot-sensitive constructions such as long-distance dependen-
cies and functional control constructions.
In order to maintain consistency with the vast majority of previous work in the
LFG framework and because the distinction between gf % and pivot is not crucial
to the analyses we present, we will continue to use subj. However, this does not
mean that we reject the important distinctions involving different aspects of the
subj role that Falk explores.
In (39), the morpheme -terir encodes third person plural agreement with the obj
a resensei ‘teachers’.
However, Mohanan goes on to show that many phrases in Malayalam that are obj
are not marked with acc case. That is, every phrase in Malayalam that is acc is an
obj, but not all objs are acc.
Relativization: Givón (1997: §4.4.3) notes that only subjects and objects can
be relativized in Kinyarwanda, and only objects can be relativized with a gap;
relativization of subjects requires the use of a resumptive pronoun.
2.1.6.1 Multiple objects In many languages it is possible to have more than one
phrase bearing an object function in a sentence. English is one such language:
Zaenen et al. (1985) discuss Icelandic, another language with multiple object
functions, and note the existence of asymmetries between the two kinds of objects.
For instance, the primary object can be the antecedent of a reflexive contained in
the secondary object:
However, the secondary object cannot antecede a reflexive contained in the pri-
mary object:
In contrast, as Bresnan and Moshi (1990) show, languages like Kichaga allow
multiple thematically restricted objects with roles other than theme:9
This example contains three object markers, representing a locative object (17obj),
an instrumental object (8obj), and a patient object (7obj). According to Bresnan
and Moshi’s analysis, in this example the instrumental obj is the unrestricted
obj; the locative and patient arguments bear thematically restricted obj functions
objloc and objpatient . Bresnan and Moshi provide much more discussion of obj𝜃
in Kichaga and other Bantu languages.
While the division of the object functions into thematically unrestricted obj and
a family of thematically restricted object functions (obj𝜃 ) represents the main-
stream approach in LFG, the notion of object and its possible subtypes remains
relatively understudied, as Börjars and Vincent (2008) point out. In a paper which
considers object as a pretheoretical construct as well as the concept obj within LFG,
Börjars and Vincent (2008) argue that the object role is not associated with specific
semantic content but should rather be viewed as a “semantically inert” grammatical
function.
The classification of a book as a direct object in both (46) and (47) may have a
semantic rather than a syntactic basis: there may be a tendency to assume that
a book must bear the same grammatical function in each instance because its
thematic role does not change. The LFG view differs: in example (46), the phrase
her bears the obj function, while in example (47), the phrase a book is the obj.
Within the transformational tradition, evidence for the LFG classification of
English objects came from certain formulations of the rule of passivization, which
applies uniformly to “transform” an object into a subject:
10 Dryer assumes a more complicated crosslinguistic typology of object functions than is generally
accepted in LFG. His richer typology turns out to be best explained in terms of different strategies for
relating thematic roles to object grammatical functions, as described in Chapter 9.
28 functional structure
this argument is the patient in the nonapplied form, and the beneficiary in the
applied form of the verb:
The patient argument alternates with the obj marker in the nonapplied form, and
the beneficiary argument alternates with the obj marker in the applied form:
This and other evidence is best explained by assuming that the patient argument
in (49a) and (50a) and the beneficiary argument in (49b) and (50b) each bear the
nonrestricted/primary obj function, while the patient argument in (49b) and (50b)
bears the restricted/secondary obj𝜃 function and behaves differently. In other
words, the syntactic behavior of the arguments in examples (49) and (50) is best
explained by reference to a distinction between obj and obj𝜃 , not between direct
and indirect objects.
Although obj𝜃 can be regarded as a secondary object function, this does not
mean that obj𝜃 appears only when an obj argument is also required. Drawing on
patterns involving passivization, causativization, and raising, Çetinoğlu and Butt
(2008) present data which support an analysis of certain Turkish transitive verbs
as subcategorizing for a subj and an obj𝜃 . Similarly, Dahlstrom (2009) proposes
that certain verb stems in the Algonquian language Meskwaki subcategorize for a
subj and an obj𝜃 . Dahlstrom shows that the obj𝜃 argument of these verbs does
not share properties with obj; for instance, it cannot undergo lexical processes
such as antipassivization (which suppresses an obj argument), and it does not
trigger agreement on the verb. In both Turkish and Meskwaki, the verbs involved
are lexically specified as taking an obj𝜃 argument. As Dahlstrom (2009) points
out, these data provide evidence in favor of retaining obj and obj𝜃 as distinct
grammatical functions, contrary to the possibility raised by Börjars and Vincent
(2008) of collapsing the two. Dalrymple and Nikolaeva (2011) also rely on the
obj/obj𝜃 distinction in their work on differential object marking and information
structure. Based on data from a range of languages, Dalrymple and Nikolaeva
(2011) claim that in certain languages a monotransitive predicate can take either
a topical obj argument or a nontopical obj𝜃 ; that is, the argument is either obj
or obj𝜃 depending on the argument’s information structure role.11 Without the
11 Chapter 10 discusses information structure and the distinction between topics and non-topics.
grammatical functions and functional structure 29
distinction between obj and obj𝜃 , this generalization would be lost. Maintaining
the distinction between these two grammatical functions therefore appears to be
justified.
The comp function is the function of clausal complements, familiar from tra-
ditional grammatical description. A comp can be declarative, interrogative, or
exclamatory (Quirk et al. 1985):
The xcomp function is an open complement function, the one borne by a phrase
like to yawn in the examples in (53). In those examples, the subj of the xcomp is
also an argument of the matrix verb; in both of the examples in (53), it is David:
Like xcomp, the xadj function is an open function, whose subj must be specified
externally; unlike xcomp, xadj is a modifier, not a governable grammatical func-
tion. In example (54), the subj of the xadj stretching his arms is also the subj of
the matrix clause, David:
12 Arka and Simpson (1998, 2008) propose that some control constructions in Balinese involve an
open subj function xsubj: for instance, in the Balinese equivalent of I tried to take the medicine, the
infinitive phrase to take the medicine can bear the subj function, with its subj controlled by the term
argument I. Falk (2005) presents an alternative analysis of the relevant data based on his proposal that,
as well as xcomp, two other governable grammatical functions exist: xobj𝜃 and xobl𝜃 . We do not
explore either of these alternatives further here.
30 functional structure
(55) David bet [Chris] [five dollars] [that she would win].
Most importantly, there is evidence for a split in the syntactic behavior of clausal
complements in a number of languages; Dalrymple and Lødrup (2000) propose
that this evidence is best explained by analyzing some clausal complements in these
“mixed languages” as obj, and some as comp.
For example, in Swedish, the clausal complement of a verb such as antar
‘assumes’ bears the obj function; it can be pronominalized and can topicalize, as
shown in examples (56a–c) from Engdahl (1999):
In contrast, the complement clause of a Swedish verb such as yrkade ‘insisted’ bears
the comp function, and does not display these properties:
As Dalrymple and Lødrup (2000) show, other languages also show a similar split in
the behavioral properties of clausal complements, with some clausal complements
patterning with nominal obj arguments and others exhibiting behavior typical of
comp arguments. Thus, they argue, the comp grammatical function cannot be
eliminated from grammatical description, since many clausal complements must
be analyzed as bearing the comp function.
By contrast, Alsina et al. (2005) present data from Spanish, Catalan, and Malay-
alam in support of discarding comp entirely and instead reinterpreting clausal
complements as having one of the grammatical functions obj, obj𝜃 , or obl𝜃 .13
According to this approach, any difference between non-clausal and clausal com-
plements bearing the same grammatical function follows solely from their differ-
ence in phrase structure category. In support of this claim, Alsina et al. (2005)
argue that there are syntactic differences among clausal complements that are
not accounted for on Dalrymple and Lødrup’s view. For instance, in Catalan it is
necessary to distinguish two types of clausal complements which alternate with
different pronominal clitics, en or hi, as shown in examples (58) and (59):
13 In their analysis of control phenomena in Balinese, Arka and Simpson (2008) make a similar
proposal in relation to xcomp.
32 functional structure
Alsina et al. (2005) claim that the shared properties of the complements of such
verbs are best captured by an analysis which treats them all as having the same
grammatical function, obl𝜃 . The result is a significant simplification, as each
verb has a single subcategorization frame and redundancy is reduced: convèncer
‘convince’ subcategorizes for sentential or prepositional oblde , and estar d’acord
‘agree’ subcategorizes for sentential or prepositional oblen .
The status of comp is still poorly understood. Dalrymple and Lødrup (2000)
provide data which indicate that comp and obl𝜃 behave differently in a number of
respects in German, and cannot be collapsed: comp clauses cannot be topicalized,
while obl𝜃 phrases can; comp clauses cannot appear in the middle field, while
obl𝜃 phrases can; and comp phrases cannot be coordinated with obl𝜃 phrases.
In contrast, Forst (2006) discusses data which indicate that German complement
clauses are best analyzed as bearing one of the grammatical functions obj, obl𝜃 , or
obj𝜃 , and that the comp grammatical function is not needed in German. Lødrup
(2012) provides a new argument for the relevance of comp, arguing that the comp
grammatical function is needed not only in the analysis of clausal complements,
but also for certain nominal complements in Norwegian: that is, that there are
nominal as well as clausal comps. Until convincing arguments can be made that all
comps in languages such as English, German, and Norwegian can be reanalyzed
in terms of other grammatical functions, comp cannot be abandoned on the basis
of being redundant.
2.1.8 PREDLINK
predlink is a closed grammatical function originally introduced by Butt et al.
(1999) to analyze predicative constructions that include a linking or copular verb,
such as those in (61):
According to this analysis, the copular predicate in these examples selects for a subj
and a predlink. predlink has also been employed in the analysis of some verbless
sentences that express similar meanings to the examples in (61) (see Attia 2008).
The predlink analysis can be contrasted with the traditional approach, which
assumes that the copular predicate selects xcomp, an open argument whose subj is
specified externally to the phrase. The xcomp analysis is problematic in those cases
where the constituent identified as the xcomp does not otherwise appear to have an
internal subject phrase, or has a different subject from that of the copular predicate.
This is the case with noun phrases (61a), adjective phrases (61b), prepositional
phrases (61c), and especially with clauses (61d). As Butt et al. (1999: 70) point
out, if the xcomp analysis is to be maintained for all predicative constructions, it
is necessary to assume that a phrase belonging to one of the relevant categories,
such as a nurse in (61a), has two sets of subcategorization requirements depending
on whether it is being used predicatively, and thus selects a subj argument, as in
(61a), or not. Under a predlink analysis, no such issue arises because predlink is
a closed function—all of its relevant grammatical dependents are included; none
are externally specified.
However, when it can be argued that the subjects of both the copula and the non-
verbal complement are the same, for instance when the post-copular complement
exhibits agreement with the subject of the main predicate as in the French examples
in (62), an xcomp analysis has been argued to capture the relation between the two
in a more straightforward manner:14
14 See Dalrymple et al. (2004a) and Attia (2008) for a different perspective.
34 functional structure
The thematic role of the oblgoal argument is marked by the preposition to. It is
not possible for more than one oblique argument to have the same thematic role:
Butt et al. (1999) provide more discussion of oblique phrases with idiosyncratic
prepositional marking.
grammatical functions and functional structure 35
2.1.10 POSS
We follow Bresnan et al. (2016) and many others in analyzing the genitive phrase
which appears in the prenominal position in an English nominal phrase as bearing
the grammatical function poss:15
Like the subj function, arguments with the poss function can bear any of a number
of semantic roles, including possessor (as in example 67), agent (as in examples
68b,c), and other roles. poss arguments often share other syntactic characteristics
of subjects: for example, poss and subj often behave similarly with respect to
anaphoric binding constraints. In §2.1.5, we saw that an identifying characteristic
of subjects in Hindi is their inability to serve as the antecedent of a pronoun in the
same clause. This was shown in example (32), repeated here:
15 Dalrymple (2001) analyzes prenominal genitives as spec rather than poss. We do not adopt that
analysis here, though we will make use of the feature spec as a feature of nominal f-structures: see §2.5.7
for a discussion of the functional features of nominal phrases, including def and spec.
36 functional structure
As Sulger shows, possessors are also able to bind reflexives within the nominal
phrase:
Indeed, Chisarik and Payne (2003) and Sulger (2015) analyze possessor arguments
as subj rather than poss, arguing that this provides a more adequate and general
picture of nominal and clausal syntactic relations. Laczkó (1995) discusses similar-
ities between subj and poss in Hungarian, though he does not argue for collapsing
the distinction between the two; see Falk (2002a) for additional arguments in favor
of maintaining a distinction between subj and poss.
Bresnan et al. (2016: 242) propose that poss is a governable grammatical
function for nominals. On this view, most or all nouns are ambiguous, either
requiring (73a) or disallowing (73b) a poss phrase:
Bresnan et al. (2016: 315–16) treat poss as a derived argument, in the sense of
Needham and Toivonen (2011); for discussion of derived arguments, see §9.10.5.2.
Though we are sympathetic with this view, we do not treat nouns as ambiguous,
and we do not represent the poss argument as one of the governable grammatical
functions of a noun.
Some languages have more than one way of expressing the possession relation
within the noun phrase. For example, the possessor in English can be expressed
either by a prenominal genitive or by a postnominal of-phrase; these are sometimes
referred to as “short” and “long” possessive phrases (Falk 2002a), or as “genitive”
and “of-oblique” possessors (Chisarik and Payne 2003):
16 Chisarik and Payne (2003) present arguments in favor of analyzing these as a new grammatical
function ncomp; we do not adopt this analysis here.
17 In much LFG work, including Bresnan and Mchombo (1987), Dalrymple (2001), Falk (2001b),
and Bresnan et al. (2016), the syntacticized discourse functions topic and focus are used as overlay
functions. However, this approach has been rejected as inadequate in other work on displaced con-
stituents, notably by Alsina (2008), who proposes an overlay feature op (“operator”) instead of discourse
functions like topic and focus, and Asudeh (2004, 2011, 2012), who proposes an overlay feature
udf (“unbounded dependency function”). We use the feature dis rather than op or udf: the term
“unbounded dependency” generally refers to the relation between a filler and a gap, as in example (75)
but not example (76), and the term “operator” is widely used in semantics but with a different (though
related) meaning.
38 functional structure
In both (76a) and (76b), there is an anaphoric dependency between the left-
dislocated constituent Beans, with the overlay function dis, and the pronoun them
with the primary function obj. Details of the analysis of anaphoric dependencies
can be found in Chapter 14.
Some LFG work (Falk 2001b, for instance) has also treated subj as being an
overlay function, the only overlay function which is also an argument. However,
this work predates Falk’s major work on subjecthood (Falk 2006), discussed in
§2.1.5, in which he proposes that subj be abandoned and replaced with two sepa-
rate functions: gf
% and pivot. gf % is a primary (nonoverlay) governable grammatical
function, while pivot is a secondary overlay function. This approach has the
advantage of eliminating from the original set of grammatical functions the only
overlay function that is an argument, subj. We use subj rather than gf % or pivot in
this book (see §2.1.5), but we follow Falk (2006) in classifying subj as a primary
nonoverlay grammatical function.
Falk (2006: 74) proposes that pivot is an overlay function which relates elements
which are “shared” between clauses, defining pivot as “the element with the
function of connecting its clause to other clauses in the sentence.” Like the overlay
function dis, a syntactic element with the function pivot must be identified with a
primary function; see Falk (2006) for details. To give an example, in Dyirbal (Dixon
1994: 15) pivot is invariably identified with either the intransitive subject or the
transitive object; in (77), the pivot obj of the verb bura ‘see’ in the first clause is
identified with the pivot gf% of the verb banaga ‘return’ in the second clause:
Falk (2006) speculates that not all languages make use of the function pivot; as
noted in §2.1.5, Choctaw/Chickasaw and Warlpiri seem to be pivotless languages,
lacking pivot-sensitive constructions such as long-distance dependencies and
functional control constructions.
As mentioned in §2.1.5, we set aside pivot in this book. The only overlay
function we employ is dis.
b. Accusative object:
Stúlkan kyssti drengina.
girl.nom kissed boys.acc
‘The girl kissed the boys.’
c. Dative subject:
Bátnum hvolfdi.
boat.dat capsized
‘The boat capsized.’
d. Genitive subject:
Verkjanna gætir ekki.
pains.gen is.noticeable not
‘The pains are not noticeable.’
In sum, the relation between grammatical function and case is complex. Even when
there is a close relation between case and grammatical function, as discussed in
§2.1.6, a clear and explanatory description of casemarking and other morphosyn-
tactic properties is best obtained by reference to abstract functional properties.
It would be difficult to find a language less like English in its phrase structure
configurational properties. Thus, Warlpiri would seem to be a good candidate to
test the hypothesis that evidence for grammatical functions can be found only in
English-like configurational languages.
However, as Hale (1983) demonstrates, languages like Warlpiri do show evi-
dence of abstract grammatical relations, just as English-like configurational lan-
guages do. Hale discusses person marking, control, and interpretation of reflexive/
reciprocal constructions, showing that constraints on these constructions are not
statable in terms of surface configurational properties. Simpson and Bresnan
(1983) and Simpson (1991) provide further evidence that properties like control
in Warlpiri are best stated in terms of abstract functional syntactic relations. In
particular, Simpson and Bresnan (1983) examine the karra-construction, in which
the subject of a subordinate clause with affix karra is controlled by the subject of
the matrix clause:
not suffer from this difficulty, however, since the hypothesized relation between the
surface order of arguments in a sentence and the order of semantic composition is
more complex.
A more subtle problem does arise, however. It is not clear that such an approach
can make certain distinctions that are necessary for syntactic analysis: in particular,
it does not seem possible to distinguish between predicates that take the same
number of arguments with the same phrasal categories. For example, any two-
argument verb that requires a nominal subject and a clausal complement should
behave like any other such verb. However, it has been claimed that there are
languages in which some clausal complements bear the obj function, while others
bear the comp function, as discussed in §2.1.7. In a theory like LFG, this distinction
is reflected in a difference in the grammatical function of the complement; some
complements are obj, and others are comp. However, it is not clear how such
a distinction can be made in a theory that does not allow explicit reference to
grammatical functions.
Dowty (1982: 107) also argues against theories which, like LFG, assume that
grammatical functions are undefined primitives by claiming that in his approach
“grammatical relations play an important role in the way syntax relates to composi-
tional semantics.” This statement is a non sequitur. In LFG, grammatical functions
are primitive concepts and also play an important role in compositional semantics
and the syntax-semantics interface. Indeed, this is the topic of Chapter 8 and the
chapters in Part III (see also Bresnan 1982a: 286).
Leaving aside these difficulties, there is a strong degree of similarity between
theories that define grammatical functions in terms of abstract properties such as
order of semantic combination and theories like LFG, in which grammatical func-
tions are not definable in terms of phrasal or argument structure. For both types of
theories, grammatical functions are abstract and are analyzed independently from
phrasal and other structures.
2.3 Subcategorization
At a minimum, the idiosyncratic information that must be specified for a word is
its pronunciation and its meaning. Research has shown that the syntactic behavior
of a word can be partially predicted from its meaning; this is because a number
of regularities govern the relation between the meaning of a predicate and the
grammatical functions of its arguments, as we discuss in detail in Chapter 9. LFG
and other linguistic theories define and capitalize on this relation in their theory
of syntactic subcategorization.
LFG assumes that syntactic subcategorization requirements of predicates are
stated at the f-structure level, in functional rather than phrasal terms. Predicates
require a set of arguments bearing particular semantic roles. These roles are asso-
ciated with grammatical functions according to a theory of argument mapping,
to be discussed in Chapter 9. In turn, these grammatical functions are realized
at the level of constituent structure in a variety of ways, as required by particular
languages: in some languages, grammatical functions are associated with particular
subcategorization 43
However, Maling shows that the true criterion at work in these examples is
not based on phrase structure category, but is semantic in nature: only gradable
predicates, those that can hold to a greater or lesser degree, are acceptable as
complements of seem (see also Bresnan et al. 2016: 295–301, and references cited
there). Many prepositional phrases do not express gradable predicates, accounting
for the unacceptability of example (82b). However, prepositional phrases that
denote gradable predicates are acceptable as complements of seem:
Further, as Maling shows, adjective phrases that are not gradable predicates are
unacceptable as complements of seem. In the following examples, the adjective
irrational as a description of a mental state is gradable and can be used as the
complement of seems, while as a technical mathematical term it is not gradable
and cannot be used:
19 A function is a special kind of relation which assigns a unique value to its argument. For example,
the relation between a person and his or her age is a function, since every person has exactly one age.
The relation between a person and his or her children is not a function, since some people have no
children and some people have more than one child.
20 We use the term feature as a synonym of attribute, so that feature-value pair and attribute-value
pair mean the same thing. Bresnan et al. (2016) use the term feature in a way that differs from ours: for
them, a feature is an attribute-value pair in which the value is simple, such as [def +] or [vtype fin].
21 In some literature, particularly in Head-Driven Phrase Structure Grammar (see, for example,
Pollard and Sag 1994), the objects that are represented in LFG as structures like (85) are instead
represented via diagrams such as:
attribute1
• value1
•
• value2
attribute2
Attributes are labeled arcs in the diagram, and values are nodes. A sequence of attributes, a path through
the f-structure, corresponds to the traversal of several labeled arcs. A possible source of confusion
for those trained within the Head-Driven Phrase Structure Grammar framework is that the same
formal notation used to represent LFG functional structures in examples like (85) is used to represent
constraints on structures in HPSG. What is depicted in (85) is not a constraint; it is a formal object.
functional structure representation 45
& '
(85) attribute1 value1
attribute2 value2
This f-structure does not contain all the syntactic information that the man
contributes. We assume here and elsewhere that the full f-structure representation
for the examples we exhibit contains at least the information shown, but may also
contain other information not relevant to the particular point under discussion.
The f-structure in (86) contains two attributes: pred and def. The value of def
is +, indicating a positive value for the definiteness feature def. This is a simple
atomic value, with no internal structure.
For the sentence The man yawned, we have the following f-structure:
(87)
pred ‘yawn!subj"’
tense pst
'
y &
pred ‘man’
subj m
def +
As (87) shows, f-structures can themselves be values of attributes: here, the value of
the attribute subj is the f-structure for the subject of the sentence. We can annotate
f-structures with labels for subsequent reference; in (87), we have annotated the
subj f-structure with the label m and the f-structure for the sentence with the label
y. These labels are not a part of the f-structure, but provide a convenient way of
referring to an f-structure and its parts.
English ones for semantic forms would be equally satisfactory, though generally
less clear for an English-speaking audience.
The list of grammatical functions mentioned in a semantic form is called the
argument list. We discuss its role in determining wellformedness constraints on
f-structures in §2.4.6. In (87), the argument list for the verbal predicate yawn
indicates that it subcategorizes for a subject only.
It is not just verbs that have arguments, of course. A preposition which con-
tributes semantic content has its own pred value which includes an argument list.
It forms a separate embedded f-structure with its argument, as shown in (88):
It is also possible for two different attributes to have the same f-structure as their
value. Here the situation is slightly more complex. Recall that an f-structure is a
functional structure representation 47
set of pairs of attributes and values: f-structures, like other sets, obey the Axiom of
Extension, which states that two sets are the same if and only if they have the same
members (Partee et al. 1993: §8.5.8). Thus, two f-structures are indistinguishable
if they contain the same attribute-value pairs.22
Notationally, it is in some cases clearer to represent two identical f-structures
separately, repeating the same f-structure as the value of the two attributes:
& '
(91) a1 v1
attribute1 a2 v2
& '
a1 v1
attribute2
a2 v2
Care must be taken if a semantic form, the value of the attribute pred, is repeated.
Since each instance of a semantic form is unique, a repeated semantic form must
be explicitly marked with an index to indicate identity; see §5.2.2.1 for more
discussion of this point. If no such index appears, the two semantic forms are
assumed to be different.
In other cases, it may be easier and more perspicuous not to repeat the f-
structure, but to use other notational means to indicate that the same f-structure
appears as the value of two different attributes. We can accomplish this by drawing
a line from one occurrence to another, a common practice in LFG literature; this
notation conveys exactly the same information as in (91):
& '
(92) a1 v1
attribute1 a2 v2
attribute2
(93)
22 This view of f-structures is different from the view of similar structures in Head-Driven Phrase
Structure Grammar (Pollard and Sag 1994; Sag et al. 2003; Levine 2017); the attribute-value structures
of HPSG are graphs, not set-theoretic objects. On the HPSG view, two attribute-value structures can
contain the same attributes and values and can nevertheless be different structures. To state the same
point in a different way: HPSG relies on a type-token distinction in attribute-value structures (Shieber
1986), meaning that two attribute-value structures are of the same type if they have the same set of
attributes and values, but may be different tokens of that type. In the set-theoretic view of LFG, the
Axiom of Extension precludes a type-token distinction, so two f-structures that have the same attributes
and values are not distinguished.
48 functional structure
2.4.4 Sets
Sets are also valid structures, and may appear as values of attributes. Sets are
often used to represent structures with an unbounded number of elements.23 For
instance, there is in principle no limit to the number of modifiers that can appear
with any phrase, and so the value of the adj attribute is the set of all modifiers that
appear. Formally, sets are enclosed in curly brackets, with the elements of the set
inside the brackets:
In (94), only a single modifier appears, but other sentences may contain more
modification:
Any valid structure can be an element of a set: for example, some sets can have
atomic values as their elements. As we discuss in §2.5.7, sets of atomic values have
been used to represent the values of the pers, gend, and case attributes in analyses
of feature resolution and feature indeterminacy.
23 The f-structure remains a function from attributes to values even if the value of an attribute is a
set. A structure like (a), in which an attribute has as its value a set with three members, is a function,
while a structure like (b), in which the same attribute has three different values simultaneously, is not a
function.
5 67 87 87 8 9:
(a) Function: attribute1 a1 v1 a2 v2 a3 v3
7 8
attribute1 a1 v1
7 8
a2 v2
(b) Not a function: attribute1
7 8
attribute1 a3 v3
functional structure representation 49
In the following example, the conjuncts of the coordinate subject David and Chris
are each singular (with the value sg for the num feature in the index feature
bundle) but the coordinate structure as a whole is a plural phrase. Thus, the set
bears the nondistributive attribute index with pl as the value for the num feature:25
Notice that nondistributive attributes such as index may appear as attributes of the
conjuncts in a set as well as attributes of the set itself: in example (97), each of the
24 The proposal to treat sets representing coordinate structures as hybrid objects is originally due
to John Maxwell (p.c.), and is foreshadowed in work on coordination in LFG by Peterson (1982) and
Andrews (1983a).
25 We motivate a more complex representation of the number feature in §2.5; in this section we make
the simplifying assumption that num has atomic values like sg (singular) and pl (plural).
50 functional structure
conjuncts David and Chris has a singular index value, while the set as a whole has
a plural index. In contrast, conjuncts can have distributive attributes like pred,
but a set cannot have a distributive attribute.
26 Recall from §2.1.2 that the governable grammatical functions are subj, obj, obj𝜃 , obl𝜃 , xcomp,
and comp.
27 Optional dependents (for example, instruments and benefactives) are discussed in §9.10.5.2.
functional structure representation 51
It is also common for the argument list to be represented in the following way,
where (↑ subj) represents the subject f-structure and (↑ obj) represents the object
f-structure, as explained in §5.3.1:
These notational variants are used equivalently, though technically the variant
shown in (101) is incorrect, since it contains the uninstantiated f-structure
metavariable ↑ (which we introduce in §5.3.1); here, we choose the more succinct
representation in (99) to save space and make the f-structures more readable.
There is a difference between grammatical functions that appear inside the
angled brackets and those that appear outside. In (99), the functions subj and
obj are semantic as well as syntactic arguments of devour, contributing to its
meaning as well as filling syntactic requirements. In contrast, the semantically
empty expletive subject it of a verb like rain makes no semantic contribution—it
is athematic. This intuitive difference is reflected in the formal requirement that
arguments of a predicate that appear inside angled brackets must contain a pred
attribute whose value is a semantic form; this is not required for arguments outside
angled brackets. Thus, the subj function appears outside the angled brackets of the
argument list of the semantic form of rain, since there is no semantic role associated
with it:
(102) It rained.
5 :
pred ‘rain!"subj’
Note that it is not a violation of any condition for more than one predicate to
govern an f-structure with a semantic form. In fact, this is a common situation with
“raising” verbs like seem, whose subject is also the subject of its xcomp argument
(see Chapter 15). Since the subject of seem is a syntactic but not a semantic
argument of the seem predicate, the subj in the value of the pred attribute of seem
appears outside the angled brackets:
The line connecting the f-structure for David to the subj position of seem indicates
that the same f-structure is the value of two different attributes, as discussed in
§2.4.3 of this chapter: it is both the subj of seem and the subj of yawn. Complete-
ness is satisfied for both predicates: each requires a subj, and this requirement is
satisfied for each verb.
Following Kaplan and Bresnan (1982), the Completeness requirement can be
formally defined as follows:
(104) Completeness:
An f-structure is locally complete if and only if it contains all the governable
grammatical functions that its predicate governs. An f-structure is complete
if and only if it and all its subsidiary f-structures are locally complete.
Chapter 8 provides further discussion of the place of the pred feature in the theory
of the syntax-semantics interface and its role in ensuring syntactic wellformedness.
It is usual but not necessary for the argument list of a predicate to mention single
grammatical functions and not lists of functions or paths through the f-structure.
In some treatments of subcategorized oblique phrases, however, the argument list
of a predicate contains longer expressions, such as oblon obj; see, for example,
Levin (1982) and Falk (2001b):
(110) Coherence:
An f-structure is locally coherent if and only if all the governable gram-
matical functions that it contains are governed by a local predicate. An
f-structure is coherent if and only if it and all its subsidiary f-structures
are locally coherent.
(111) Consistency:
In a given f-structure a particular attribute may have at most one value.
The subj noun phrase the boys is a plural (pl) phrase, but the verb yawns requires
its subject to be singular (sg). These two requirements cannot be simultaneously
met: the value of the attribute num must be either sg or pl, and it cannot have both
values at the same time.
(113) Simple atomic values for the feature pers: 1 (first person), 2 (second
person), 3 (third person)
However, atomic values such as these do not underpin a predictive theory of feature
resolution in coordination. For example, a theory of syntactic feature resolution
functional features 55
should predict that coordinating a second person pronoun with a third person
pronoun produces a coordinate phrase with a second person feature in every
language (Corbett 2000), and that other resolution patterns are impossible:
Similarly, the use of atomic features does not allow for a formally explicit theory
of natural classes of features and syncretism (Baerman et al. 2005). For example,
Baerman et al. (2002) present the following paradigm for present tense verbs in
Rongpo,28 exemplifying with the verb ‘go’:
The same verb form gyən is used with second and third person singular subjects.
Representation of the value of the pers feature as a simple atomic value such
as 2 or 3 does not allow for the formal expression of properties (such as “non-
first-person”) that the two values have in common. As Baerman et al. (2005:
Chapter 3) point out, other languages display syncretisms involving first/third
person and first/second person, and so the complex value of the pers feature
must allow reference to these commonalities as well. We discuss proposals for the
complex representation of the value of the pers feature in §2.5.7.2.
It is also often assumed that the num (number) feature has simple atomic values
such as sg (singular), du (dual), and pl (plural):
(116) Simple atomic values for the feature num: sg (singular), du (dual), pl
(plural)…
(118) Simple atomic values for the feature case: nom (nominative), acc
(accusative), dat (dative), erg (ergative), abs (absolutive), gen (genitive),
loc (locative)…
Complex values for the case feature allow indeterminate forms to satisfy multiple
conflicting case requirements simultaneously imposed by different predicates, as
we discuss in §2.5.7.5.
functional features 57
29 For definition and discussion of markedness and the representation of feature values, see Jakobson
(1932), Greenberg (1966a), Moravcsik (1988), Blevins (2000), and references cited in those works.
58 functional structure
b. (set-based representation)
−
+
c.
−
(feature structure representation)
..
.
On Falk’s view, tensed forms are associated with a tense feature, participial forms
are associated with a part feature, and the infinitive is identified by the absence of
both of these features.
One issue that arises for such approaches is the treatment of feature under-
specification. In many approaches, a word form which can have any of several
values for a feature is treated by leaving the value for the feature underspecified
functional features 59
or unspecified. For example, the verb form put is the present tense form, the past
tense form, or the infinitive form of the verb:
(122) a. David and Chris put their dishes in the dishwasher every day. [present
tense]
b. David and Chris put their dishes in the dishwasher yesterday. [past tense]
c. David and Chris don’t want to put their dishes in the dishwasher today.
[infinitive]
It is, then, necessary to encode information about the form of words at f-structure,
in order to allow such requirements to be stated. Word form is encoded by means
of features such as form; the f-structure for It rained is:
(124)
Notice that the subj argument of rain appears outside the angled brackets in the
semantic form, since it is syntactically required but does not make a semantic
contribution (§2.4.6.1). The f-structure of the subject contains the attribute form
with value it, satisfying the special requirements of the verb rained.
This means that the pronoun it makes one of two different f-structure contribu-
tions on each occasion of its use: one of them, shown in (124), has a form attribute
with value it, while the other is a standard referential pronoun containing a pred
attribute with value ‘pro’, as in a sentence like David saw it. It is generally assumed
60 functional structure
that the only words which have an f-structure with a form feature are those whose
presence is syntactically required by some predicate in the language but which
do not make a semantic contribution. For example, there is no English verb that
requires the presence of an argument with the form hippopotamus, and so there is
no f-structure for the word hippopotamus which contains a form feature and no
pred feature.
Some verbs place constraints on the form of the complementizer of their
complement clause; for example, Quirk et al. (1972: §15.6) claim that the verb
justify allows the complementizer whether, but not the complementizer if, in its
comp argument:
(125) You have to justify whether/ ⃰if your journey is really necessary.
It is the value of the compform feature that is constrained by the verb justify,
which disallows the value if for the compform feature in its comp. We return
to a discussion of this example in §5.2.5.
In their comparative grammars of English, French, and German, Butt et al.
(1999) propose a number of additional specialized form features, including pron-
form for pronouns, spec-form for nominal specifiers like the, conj-form and
preconj-form for conjunctions like and and preconjunctions like either or both,
prt-form for particles (including German verbal particles as well as English
particles like up), and comp-form (which we call compform). In the compu-
tational setting of Butt et al.’s work, the linguistic analysis of a sentence is input
to a computer application that performs further processing for some practical
purpose. In such a setting, features such as these are often included in the analysis
in case they are useful for such applications, rather than for purely syntactic
reasons. In the analyses we present, we introduce only f-structure features which
functional features 61
we can show to be clearly motivated by syntactic patterns and which take part in
syntactic selection requirements. It may be that further research will reveal the
need for additional form features at f-structure such as those that Butt et al. (1999)
propose.
In §11.4.1, we will introduce an fm string feature whose value is the form of
the unit of the string under analysis. We will see that the fm feature bears a close
relation to the f-structure form feature, but is relevant for relations in the string
rather than f-structural relations and constraints.
(127) Non-standard f-structure for A girl handed a toy to the baby according to
Kaplan and Bresnan (1982, (45)):
’
According to this early analysis, the value of the pcase attribute is to, which
is also the grammatical function of the phrase to the baby (see Butt et al. 1999
for a different but related treatment). In later work, two different attributes have
generally been assumed: the form of the preposition is instead recorded as the
value of the pform feature, as we describe below, while the value of the pcase
attribute is specified as the particular oblique grammatical function contributed
by the prepositional phrase. For example, in (128), the value of the pform
attribute is to, and the value of the pcase attribute is the grammatical function
oblgoal , which is also the grammatical function of the prepositional phrase
to the baby:
62 functional structure
(128) F-structure for A girl handed a toy to the baby, standard treatment of pcase
feature:
’
In §6.2, we describe the constraints which allow the value of the pcase attribute to
specify the grammatical function borne by the oblique phrase.30
F-structural specification of the form of particles or prepositions is necessary
because predicates can place idiosyncratic requirements on the form of their
prepositional complements or particles (§2.1.9.2):
The pform feature records the value of the form of the particle or preposition so
that it can be constrained and checked by the predicate:31
30 As we discuss in §§6.1.3 and 6.2, the pcase analysis is closely related to the constructive case
theory of Nordlinger (1998, 2000).
31 See Toivonen (2003) for an in-depth discussion of particles in Germanic languages, and an
alternative proposal for the f-structure of English particle verbs.
functional features 63
2.5.5.1 vtype and tense We adopt the proposal of Payne and Börjars (2015)
for the vtype and tense features and their values. On their proposal, all verb forms
have a vtype feature, and tensed forms also have a tense feature:32
help):
The vtype feature has the values listed in (132) in English; additional values may
be required for other languages, depending on the richness of the verbal paradigm:
(132) Values for the vtype feature: fin (finite), prs.ptcp (present participle),
pst.ptcp (past participle), inf (bare infinitive)
For English, the tense feature has at least the values past (pst) and present (prs).
There is some controversy over the status of English will as a future tense marker
(Huddleston 1995); we do not take a position on this issue here, though we note
that fut is a possible value for the tense feature in many languages:
(133) Values for the tense feature: pst (past), prs (present), fut (future)
32 Payne and Börjars use the attribute vform rather than vtype for this feature. For consistency with
the Leipzig glossing rules (Bickel et al. 2015), we provide the value prs.ptcp for the present participial
form of a verb, rather than Payne and Börjars’s ger-part.
64 functional structure
2.5.5.2 mood and aspect Besides the tense feature, aspect and mood con-
stitute the other two dimensions of tense-aspect-mood systems (often abbreviated
as TAM). We adopt Butt et al.’s (1999) treatment of the mood feature, and a variant
of their analysis for the aspect feature.
Butt et al. (1999) propose the values ind (indicative), sbjv (subjunctive), and
imp (imperative) for the feature mood in English, French, and German; to these,
we add the value interr (interrogative):
(134) Values for the mood feature: ind (indicative), sbjv (subjunctive), imp
(imperative), interr (interrogative)
For other languages, additional values may be required: for example, Ancient
Greek has an optative (opt) mood.
The treatment of aspect is more complex. Butt et al.’s analysis of English,
French, and German makes use of an aspect feature which they abbreviate as perf.
Butt et al. (1999) do not distinguish between perfect and perfective aspect in their
representation of the aspect feature; we follow standard practice in using prf for
perfect aspect, and pfv for perfective aspect. We also follow Butt et al. in using a
single feature, which we represent as prf, in the analysis of aspectual distinctions
in English, French, and German.
According to Butt et al.’s analysis of English, the feature prf appears with value +
for sentences with the perfect auxiliary, and does not appear in sentences without
the perfect auxiliary:
5 7 8:
(135) a. David has eaten a banana: aspect prf +
Butt et al. also propose a feature prog (progressive), since in English, a sentence
can have both perfect and progressive aspect:
33 Here and in general, it is worth keeping in mind that the f-structure is a purely syntactic
representation, and need not encode fine-grained semantic distinctions. The distinction between near
and remote past may not be syntactically significant in any language, even when it is important
semantically; whether or not the full range of semantically relevant distinctions is encoded in the
f-structure is one aspect of the general issue of the relation between syntax and meaning, discussed
in §2.5.1.
functional features 65
+
(136) David has been eating bananas:
+
A representation like this works well for languages like English, in which two
separate dimensions of the aspect feature must be represented in the same
f-structure. We provide the representation in (137) for such languages, similar to
the representation proposed by Butt et al. (1999):
Progressive: +
+
Perfect progressive: +
(138) Atomic values for the aspect feature: prf (perfect), iprf (imperfect), pfv
(perfective), ipfv (imperfective), prog (progressive)
2.5.5.3 voice Many voice systems encode a two-way distinction between active
and passive voice; to represent this distinction, Sadler and Spencer (2001) propose
a feature voice with values active (active) and pass (passive). Often, active voice
is treated as the unmarked, default value for the voice feature: if we adopt this view
in Sadler and Spencer’s setting, active voice can be represented by the absence of the
voice feature, and passive voice can be represented by the presence of the voice
feature with the value pass. Additional values for the voice feature are necessary
in the analysis of languages with antipassive voice (antip) and middle voice (mid)
(Fox and Hopper 1993).
As we discuss in Chapter 9, voice alternations are often viewed in terms of
“promotion” and “demotion” of arguments, with the active voice taken as basic,
and the passive voice involving “demotion” of the subject to oblique status, and
“promotion” of a nonsubject argument to subject status. Many languages have
voice systems which do not work in this way: in so-called “symmetric voice
systems” (Foley 1998), voice alternations involve singling out one argument as
the subj, but without demoting the other core arguments or altering their core
status. Such systems often have several voices, each named after the role of the
argument that is singled out as the subj. For example, Kroeger (1993) discusses
five voices in Tagalog: actor voice (sometimes called “active voice”: av), objective
voice (ov), dative/locative voice (dv), instrumental voice (iv), and benefactive
voice (bv). Interestingly, Arka (2003) shows that Balinese has a three-way voice
alternation, with demotional and symmetric voice alternations coexisting in the
same language: Balinese distinguishes agentive voice (similar to actor voice in
Tagalog, and also represented as the value av), a non-demotional objective voice
(ov) in which the agent remains a core argument, and a demotional passive voice
(pass) in which the agent is demoted to oblique status. Kroeger (1993), Arka
(2003), and Falk (2006) provide an in-depth discussion of these “symmetric voice”
languages, their clausal syntactic organization, and their analysis within the theory
of argument mapping, the relation between grammatical functions and semantic
roles; we discuss argument mapping in Chapter 9.
In languages with direct and inverse voice, the status of the arguments of a
verbal predicate and their position on an animacy/saliency hierarchy determines
the voice of a clause: for example, a first person agent acting on a third person
patient must be expressed with direct voice (dir), but a third person agent acting
on a first person patient must be expressed in inverse voice (inv) (see §5.2.10 for
a discussion of related patterns in Lummi). Arnold (1994) provides an analysis of
argument mapping and inverse voice in Mapudungan.
Our inventory of values for the voice feature must, then, encompass values
that are relevant in the analysis of demotional voice alternations such as the
active/pass alternation as well as voice alternations in symmetric voice languages
and languages with direct and inverse voice. Additional values may be needed
in the analysis of symmetrical voice languages with voice distinctions other than
those attested in Tagalog and Balinese:
(139) Values for the voice feature: active (active), pass (passive), antip
(antipassive), mid (middle), av (actor/agentive voice), ov (objective voice),
dv (dative voice), bv (benefactive voice), iv (instrumental voice), dir
(direct voice), inv (inverse voice)
functional features 67
2.5.6 Negation
The presence of negation affects syntactic patterns such as case assignment and
negative concord (see Przepiórkowski and Patejuk 2015 for Polish), and so nega-
tion must be represented explicitly at f-structure. Clausal negation is represented
by a feature neg with value + by King (1995), Butt et al. (1999), and Falk (2001b),
among others:
A related feature, neg-form, is proposed by Butt et al. (1999: 142–3) in the analysis
of bipartite clausal negation forms such as French, with the initial component
contributing a neg feature, and the postnegation component contributing the
value of neg-form:
(141) F-structure for French bipartite negation ne. . . . pas according to Butt et al.
(1999):
David n’ a pas mangé de soupe.
neg have postneg eaten of soup
‘David did not eat any soup.’
’
+
Butt et al. (1999: 143) propose that the neg feature is also used in the treatment
of constituent negation, where only a subconstituent is negated and not the entire
clause:
(142) F-structure for David found the ball not in the box [but on the floor…]
according to Butt et al. (1999):
’
’
+
68 functional structure
However, a single neg feature may not be adequate in the analysis of multiple
negation examples such as (143), from Przepiórkowski and Patejuk (2015):
(144) F-structure for John doesn’t like Mary with neg feature:
’
+
To ensure the proper treatment of examples with multiple negation, the ParGram
Consortium (2017) proposes to treat at least some instances of clausal negation
adverbially, with an f-structure representing clausal negation in the adj set:
(145) F-structure for John doesn’t like Mary with negation as an adjunct:
’
(146) F-structure for John doesn’t not like Mary with negation as an adjunct:
’
34 Bresnan et al. (2016: Chapter 8) also group pers, num, and gend features together as components
of the value of the index feature, as we do here. They provide a different treatment of index and
agreement features in their discussion of anaphoric binding in their Chapter 10, where they use the
cover term agr for pers, num, and gend features, and provide a variable such as i or j for the value of
the attribute index.
35 Alsina and Arsenijević (2012a,b) argue against the index/concord distinction and in favor of
a single type of syntactic agreement feature; Wechsler and Zlatić (2012) and Hristov (2013a) provide
counterarguments to Alsina and Arsenijević’s position.
70 functional structure
(147) num ...
concord gend ...
case ...
pers ...
index num ...
gend ...
As shown in (147), both the index feature bundle and the concord feature bundle
contain num and gend features. Since index and concord features are specified
separately, the values of these features can differ: for example, a form can have
singular concord but plural index, or vice versa. Wechsler and Zlatić (2003:
51) provide the Bosnian/Croatian/Serbian36 example in (148) to illustrate this
possibility:
The noun deca ‘children’ controls feminine singular agreement on the determiner
and adjective, but third person plural agreement on the verb. Wechsler and Zlatić
propose that deca is unusual in having feminine singular concord features but
neuter plural index features:
(149) index and concord features for deca, Wechsler and Zlatić (2003):
36 Due to the complexity of the political and linguistic divisions in the former Yugoslavia we use
the name Bosnian/Croatian/Serbian to refer to the primary languages of Bosnia and Herzegovina,
Croatia, Montenegro, and Serbia, without prejudice as to how many distinct languages or varieties this
encompasses and without implying primacy of one variety or language over others.
functional features 71
2.5.7.2 Person The pers feature is often represented with a simple atomic
value:37
(151) Simple atomic values for the feature pers: 1 (first person), 2 (second
person), 3 (third person)
37 Wechsler (2004) proposes additional atomic values for languages like French, which have a
distinction in the first person between a singular pronoun (je) and a plural or ‘authorial’ singular
pronoun (nous), and in the second person between an informal singular pronoun (tu) and a plural or
formal singular pronoun (vous). According to Wechsler’s analysis, values for the pers feature in French
are 1s (first person singular), 1a (first person associative or “authorial” first person), 2s (second person
singular informal), and 2a (second person associative or formal); third person is represented by the
absence of a first or second person feature.
72 functional structure
There are two sorts of motivations for a more complex representation of the value
of the pers feature. First, as noted in §2.5.1, the pers feature resolves in predictable
ways in coordination: for English and many other languages, the resolution rules
are as in (152). In a formal theory of syntactic feature resolution, complex values
allow resolution to be modeled in a way that accounts for these patterns:
Secondly, as noted in §2.5.1, values of the pers feature pattern together in agree-
ment relations, and a complex representation allows for the definition of natural
classes and commonalities in values of the pers feature. We discuss analyses
focusing on each of these phenomena in turn.
Dalrymple and Kaplan (2000) propose a theory of feature resolution which
defines the value of the pers feature as a closed set38 containing the atoms s
(mnemonic for speaker) and h (mnemonic for hearer):
(153) Values for the pers feature according to Dalrymple and Kaplan (2000):
{s, h}: first person
{h}: second person
{ }: third person
Dalrymple and Kaplan further propose that feature resolution can be modeled as
set union:39 the pers value of a coordinate structure is determined by taking the
set union of the pers values of the conjuncts. The table in (154) represents exactly
the same pattern of person agreement as the table in (152):
(154) Resolution of the pers feature via set union (Dalrymple and Kaplan 2000):
{s, h} (first) ∪ {h} (second) = {s, h} (first)
{s, h} (first) ∪ { } (third) = {s, h} (first)
{h} (second) ∪ { } (third) = {h} (second)
{ } (third) ∪ { } (third) = { } (third)
Other languages have a richer system of personal pronouns. For example, Fula and
many other languages exhibit a distinction between inclusive and exclusive first
person pronouns: the referent of an inclusive first person pronoun includes the
hearer, while the referent of an exclusive first person pronoun excludes the hearer.
For such languages, Dalrymple and Kaplan propose the following refinement to
the feature values in (153):
(155) Values for the pers feature in languages with a first person inclusive/
exclusive distinction according to Dalrymple and Kaplan (2000):
{s, h}: first person inclusive
{s}: first person exclusive
{h}: second person
{ }: third person
In their analysis of the pers feature, Vincent and Börjars (2007) propose that
the pers feature has the inventory of feature values in (155) in all languages,
distinguishing inclusive from exclusive pronouns in the first person, whether or
not there is a morphological inclusive/exclusive distinction in the pronominal
system.
In §2.5.1, we discussed theoretical issues that bear on the determination of the
value of features like pers. The issue of universality is whether all languages share
the same representation and structure for the values of a feature like pers. For
Vincent and Börjars, universality is a high priority: they propose as a working
hypothesis that “in grammatical sub-systems with a clear referential basis like per-
son, the definitions of the persons should be constant across languages” (Vincent
and Börjars 2007: 301–2). The issue of the relation between syntax and meaning
is whether and to what extent the complex value of a syntactic feature like pers
reflects details of the meaning of the form with which it is associated. In a language
like English, the first person plural pronoun we can be used to refer to the speaker
and the hearer (an inclusive reading) or to the speaker and another person (an
exclusive reading); prioritizing a transparent relation between syntax and meaning
entails that this meaning distinction should be encoded as a syntactic ambiguity
in the f-structural representation, even if there is no morphological or syntactic
evidence for this ambiguity. On the basis of these motivations, Vincent and Börjars
(2007) argue that the more complex system in (155) is universal, and relevant for
languages like English in addition to languages like Fula.
Sadler (2011) observes that set-based analyses of complex feature values can be
directly translated to equivalent feature-based analyses, with a positive value rep-
resenting the presence of an element in the set, and a negative value representing
its absence (see §6.3.4). Otoguro (2015) proposes a feature-based analysis of the
pers feature in his analysis of pers agreement in Germanic and Romance:
2.5.7.3 Gender Many LFG analyses assume simple atomic values for the gend
feature:
(157) Simple atomic values for the feature gend: m (masculine), f (feminine), n
(neuter) . . .
(160) Icelandic:
masculine & masculine = masculine
feminine & feminine = feminine
neuter & neuter = neuter
masculine & feminine = neuter
masculine & neuter = neuter
feminine & neuter = neuter
Dalrymple and Kaplan (2000) propose that the value of the gend feature is a
closed set containing members like m and f, and that, as in their analysis of person
resolution, gender resolution is modeled as set union. For Icelandic, they propose
the gend values in (161):
(161) Values for the gend feature in Icelandic according to Dalrymple and
Kaplan (2000):
masculine {m}
feminine {f}
neuter {m, f}
Given these values, modeling gender resolution as set union produces the correct
result, as can be seen by comparing the tables in (160) and (162):
(162) Resolution of the gend feature as set union (Dalrymple and Kaplan 2000):
{m} (masculine) ∪ {m} (masculine) = {m} (masculine)
{f} (feminine) ∪ {f} (feminine) = {f} (feminine)
{m, f} (neuter) ∪ {m, f} (neuter) = {m, f} (neuter)
{m} (masculine) ∪ {f} (feminine) = {m, f} (neuter)
{m} (masculine) ∪ {m, f} (neuter) = {m, f} (neuter)
{f} (feminine) ∪ {m, f} (neuter) = {m, f} (neuter)
Vincent and Börjars (2007) also propose an analysis using sets in the representation
of the value of the gend feature, but object to the gender values that Dalrymple
and Kaplan present, and to the use of set union to model gender resolution. They
observe that Dalrymple and Kaplan’s analysis of gender violates the markedness
criterion discussed in §2.5.1, according to which the formally more marked value
of a feature should behave as the more marked member in an opposition of feature
values. In Icelandic, neuter is the unmarked gender value; however, in Dalrymple
and Kaplan’s system, neuter is formally the most marked, since it is represented as
a set with two members (m and f), while the masculine and feminine values each
contain one member and so are less formally marked.
Vincent and Börjars (2007: 297) propose as a desideratum for the formal
representation of complex feature values that “for whatever morphosyntactic sub-
system we examine, we must be able to show by independent tests that the member
76 functional structure
of that system which is assigned the representation { } acts as the unmarked form.”
They propose the values given in (163) for the gend feature in Icelandic:
(163) Values for the gend feature in Icelandic according to Vincent and Börjars
(2007):
masculine {m}
feminine {f}
neuter { }
Here the unmarked form is the neuter value, represented as the empty set { }, as
desired. As Vincent and Börjars observe, these values allow for a pleasingly simple
generalization about gender resolution in Icelandic and many other languages: for
conjuncts of different genders, agreement is with the unmarked form, represented
as { }. Crucially, however, under these assumptions it is necessary to model gender
resolution by set intersection (∩) rather than set union (∪):
(164) Resolution of the gend feature as set intersection (Vincent and Börjars
2007):
{m} (masculine) ∩ {m} (masculine) = {m} (masculine)
{f} (feminine) ∩ {f} (feminine) = {f} (feminine)
{ } (neuter) ∩ { } (neuter) = { } (neuter)
{m} (masculine) ∩ {f} (feminine) = { } (neuter)
{m} (masculine) ∩ { } (neuter) = { } (neuter)
{f} (feminine) ∩ { } (neuter) = { } (neuter)
As mentioned above and discussed in more detail in §6.3.4, Sadler (2011) demon-
strates that set-based analyses of complex feature values have a ready translation to
feature-based analyses. She provides a feature-based analysis of Icelandic gender
which conforms to Vincent and Börjars’s observations about markedness if we
assume that a negative value for a feature corresponds to the unmarked value for
that feature.
2.5.7.4 Number In much LFG work, the number feature is assumed to have
atomic values such as sg, du, paucal, and pl:
(165) Simple atomic values for the feature num: sg (singular), du (dual), paucal
(paucal), pl (plural)
However, patterns involving constructed number are often taken as evidence that
the value of the num feature is not atomic, but a complex value whose form
may be constrained in different ways by different parts of the sentence. As briefly
discussed in §2.5.1, Sadler (2011) provides an analysis of the constructed dual in
Hopi which assumes a complex value for the num feature. Example (117), repeated
here, illustrates the Hopi patterns:
functional features 77
Sadler proposes the following complex values for the num feature:
Arka (2011, 2012) also proposes a feature-based analysis for constructed number
in Marori, and Jones (2015a,b) proposes a feature-based analysis for the com-
plex number system of Meryam Mir. Interestingly, Nordlinger (2012) presents a
detailed examination of the number system of Murrinh-Patha, arguing that the
complex patterns of number marking in that language are not best analyzed in
terms of a feature-based treatment of the num feature such as Sadler, Arka, and
Jones propose.
We know of no set-based proposals for the value of the number feature, and
indeed it is not clear that it would be possible to formulate an analysis of the
number feature of coordinate noun phrases that could take advantage of the
set-based operations of union or intersection that have been proposed for pers
resolution (§2.5.7.2) or gend resolution (§2.5.7.3). If we assume a set-based rep-
resentation of the num feature, and we assume that the num value for a coordinate
structure is determined by set union or set intersection, as for pers and gend, we
would incorrectly expect a coordinate structure with two singular conjuncts to be
singular: no matter what set we choose to represent the singular value for the num
feature, the union of any set with itself is that set, and the intersection of any set
with itself is also that set. In fact, however, a coordinate structure such as David
and Chris, with singular conjuncts, has plural number and not singular number;
set union and intersection do not provide a straightforward foundation for a theory
of number in coordinate structures. Dalrymple (2012) provides a general overview
of issues related to the treatment of the num feature and the representation of its
value in LFG.
The Bosnian/Croatian/Serbian example discussed in §2.5.7.1 shows that the
same noun phrase can control different values for the num feature on different
agreement targets. Mittendorf and Sadler (2005) present two alternative analyses of
number mismatch in Welsh NPs: one analysis assumes a complex f-structure, and
the other relies on the index/concord distinction and the possibility for the num
feature to be specified differently for index and concord. Hristov (2012) proposes
an analysis of nouns which seem to exhibit both singular and plural properties,
such as British English “company,” which require singular determiners but can
appear with plural verbs (as in the attested example This company are superbly
managed: Hristov 2012: 174). Wechsler (2011) proposes that some agreement
mismatches can be driven by the absence of syntactic agreement features on the
controller (see also Wechsler and Hahm 2011); this is his Agreement Marking
Principle, informally stated in (169):
In support of this view, Wechsler (2011) discusses the French examples in (170), in
which the verb êtes displays syntactic (second person plural) index agreement with
functional features 79
the pronominal subject vous ‘you’ whether the subject is semantically singular (as
in example 170a, exemplifying the polite plural use of the pronoun) or semantically
plural (as in 170b), while the predicate adjective displays semantic agreement
(singular in 170a, plural in 170b). Wechsler proposes that the French pronoun vous
does not bear concord features, and so there are no syntactic features to control
agreement with the predicate adjective; in this situation, the Agreement Marking
Principle correctly predicts that the adjective agreement features reflect semantic
features of the controller:
2.5.7.5 Case A great deal of work in LFG has been done on the morphosyntax
of case. Butt and King (2004c) discuss the kinds of syntactic relations that can
be marked by case: semantic case and quirky case can appear on either terms or
obliques, and structural case and default case appear on terms. Otoguro (2006)
discusses the theoretical status of case, illustrating his discussion with examples
from Icelandic, Hindi-Urdu, and Japanese. For a thorough and wide-ranging
overview of case and its role in grammar, see Butt (2006); for a detailed discussion
of case in an LFG setting, see Butt (2008).
We have seen that features such as index must be treated as nondistributive,
since a coordinate structure has its own value for index features which is deter-
mined by rules referring to the value of the index features of the conjuncts. A
distinguishing characteristic of nondistributive features, including the components
of the index feature, is that for any resolving feature and any combination of
features of the conjuncts, there is a means of determining the resolved value of
the feature for the coordinate structure. As a component of the concord feature
80 functional structure
bundle, the case feature behaves differently, in that there are no resolution rules
for case: that is, there is no sense in which different values for the case feature
systematically combine to determine a resolved case value for the coordinate
structure.
In fact, coordinate structures in which the conjuncts have mismatching case
values are often unacceptable, and this is predicted if we treat case as a distributive
feature, as we discuss in §6.3.2.2. For example, if we assume that the concord
feature is distributive and that subjects must bear nominative case, we predict that
each conjunct in a coordinate subject phrase must specify nominative case in its
concord feature bundle. However, the situation is in fact more complex: Peterson
(2004a) and Hristov (2012: Chapter 5; 2013b) argue that case in English is best
treated as a nondistributive feature, in light of attested examples exhibiting case
mismatch such as . . .[him and I] have this fight, right?, from the British National
Corpus, and Patejuk (2015) provides further discussion of case mismatches in
coordinate structures in Polish. In this book, we provisionally treat concord and
its component case feature as a distributive feature, but further research may
show the need for treating it instead as a nondistributive feature, and analyzing
its behavior in coordination by means of constraints on the relation between
the concord value of the coordinate structure and the concord values of the
conjuncts.
Simple values such as those listed in (172) are often assumed for the case feature:
(172) Simple atomic values for the feature case: nom (nominative), acc
(accusative) erg (ergative), abs (absolutive), gen (genitive), loc (locative),
dat (dative), abl (ablative), ins (instrumental). . .
This simple representation does not allow for a treatment of case indeterminacy
(Groos and van Reimsdijk 1979; Zaenen and Karttunen 1984), where a form with
underspecified case can satisfy more than one case requirement at the same time.
The German nouns Männer/Männern ‘men’ and Papageien ‘parrots’ differ in the
cases that they express:
(173) a. Papageien
‘parrots.nom/acc/dat/gen’
b. Männer
‘men.nom/acc/gen’
c. Männern
‘men.dat’
As noted in §2.5.1, Papageien but not Männer or Männern can appear as the
shared object of coordinated verbs if one verb requires accusative case and the
other requires dative case (174c). Example (174) shows that the verb findet ‘find’
requires an accusative object (174a), while the verb hilft ‘helps’ requires a dative
object (174b) (Pullum and Zwicky 1986):
functional features 81
Dalrymple and Kaplan (2000) propose to represent the value of the case feature of
a nominal as a set whose members are the case values that the nominal can express.
A verb taking an accusative object requires acc to be a member of the case set of
its object, and similarly for a verb taking a dat object:40
This treatment is adequate for simple cases of indeterminacy, but Dalrymple et al.
(2009) show that it does not make correct predictions for some more complex
cases. They examine the transitivity problem (where a modifier restricts the pos-
sibilities for case expression of its head noun) and the second-order indeterminacy
problem (where a predicate imposes indeterminate requirements on an argument
which may itself be indeterminately specified); see Dalrymple et al. (2009) for
detailed discussion of these issues, and a demonstration that a feature-based
approach with underspecification solves both of these problems. Dalrymple et al.
propose the following representation for the determinate dative case value for a
noun like Männern:
40 The symbol ∈ is the set-membership predicate: a ∈ S means that a is a member of the set S.
See §6.3.
82 functional structure
(176) Determinate dat case for Männern according to Dalrymple et al. (2009):
nom −
case
acc −
gen −
dat +
Predicates specify a positive value for the case that they require: for example, a
predicate requiring dative case specifies [dat +] for its argument. Indeterminate
nouns like Männer and Papageien are underspecified; they can be assigned a value
of + for one or more case features, as long as the value of the feature is not already
specified as −. For clarity of exposition, we explicitly represent all case features in
(177) and (178), with a blank value for features with no assigned value:
Complex values for the case feature (whether set-based or feature-based) bear a
superficial resemblance to the complex values of the pers, num, and gend features;
importantly, however, they express very different linguistic intuitions, despite their
functional features 83
formal similarity. For indeterminate features like case, a complex value such as
{nom,acc} or [nom +, acc +] enumerates a set of alternative possibilities which
may be simultaneously realized, allowing an indeterminate form to simultaneously
fulfill conflicting case requirements. In contrast, the specification of the dual value
of the num feature as [sg +, pl +] does not entail that dual nouns are in any
sense simultaneously sg and pl, or alternately sg or pl with respect to different
predicates; instead, complete patterns of values of the pers, num, and gend
features holistically represent the values for these features.
2.5.7.6 Nominal specification and quantification The function spec has been
used in the analysis of articles such as the and a as well as quantifiers and
occasionally possessors, thus grouping together elements which “specify” rather
than modify the head in a nominal phrase. Sadler (2000) and Falk (2002a) note that
treating both possessors and articles as contributing a value of the spec attribute
makes it difficult to analyze languages which allow both an article and a possessor
in the same nominal phrase; Falk (2002a) provides example (179) to show that the
definite article is obligatory in Romanian noun phrases containing a possessor:41
The article and the possessor cannot both contribute values for the same spec
attribute in examples like (179): either the article or the possessor must contribute
a value for an attribute other than spec. As discussed in §2.1.10, we adopt the poss
grammatical function for possessors.
Similarly, English noun phrases can contain both an article and a quantifier:
Therefore, we also need separate attributes for the contributions of articles and
quantifiers.
We make use of the features def, deixis, and spec for nominal f-structures. In
English, the value for the feature def is either + or −:
Demonstrative determiners like this and those are definite, contributing the value
+ for the def feature, as well as contributing a deixis feature. The English deictic
system makes only a two way proximal/distal distinction. The deictic systems of
other languages are more complex, including additional values such as medial:
(182) Values for the feature deixis: prox (proximal), medial (medial), dist
(distal) …
There is an important difference between articles like the and quantifiers like every
or all: articles cannot be modified ( ⃰almost the), but quantifiers can be (almost
every). This means that a quantifier must contribute not an atomic value for a
feature, but an f-structure with its own pred attribute and semantic form value,
in which an adj attribute can also appear:
(183) a. every:
b. almost every:
Therefore, the value for the feature spec is an f-structure whose semantic form
represents the quantifier:
(184) Values for the feature spec: f-structures with semantic form ‘all’, ‘every’,
‘both’…
(185) Values for the feature prontype: rel (relative pronoun), wh (interroga-
tive pronoun), pers (personal pronoun), refl (reflexive pronoun), recp
(reciprocal pronoun)
The features in (187) have been argued to have complex values, at least in some
languages. For details of the complex value of the feature, see the indicated section.
As discussed in §2.5.1, we often use atomic abbreviations for these complex values
when the details of the representation of the feature are not important to the issues
under discussion, and it is these abbreviations which appear in (187):
Features can also be cross-classified in several ways. Section 2.5.7.1 discussed two
kinds of nominal features, index and concord features:
The features num and gend appear in both feature bundles, and in fact they
can take on different values in each feature bundle: for example, as shown in
example (148) of this chapter, the Bosnian/Croatian/Serbian noun dobra ‘children’
has singular concord, but plural index.
In §2.4.5, we demonstrated the need to distinguish betwen distributive and
nondistributive features in coordinate structures. We assume that the only nondis-
tributive features are index (specifying the resolved pers, num, and gend features
of coordinate noun phrases), adj (specifying modifiers of coordinate structures
in examples like David yawned and sneezed simultaneously, where simultaneously
modifies the coordinated verbs yawned and sneezed), and conj and preconj
(specifying the form of the conjunction and preconjunction). We return to a
discussion of distributive and nondistributive features in §6.3.2 and Chapter 16:
Example f-structures for a range of simple English nominal phrases are given in
(190). Recall that sg and pl are abbreviations standing for complex values for the
num feature, as discussed in §2.5.7.4, and similarly for the simple abbreviatory
values presented in (190k) for the case, pers, and gend features. We omit
concord features in most of the examples:
(190) a dog:
−
b. the dog:
+
c. this dog:
+
functional features 87
d. that dog:
+
e. every dog:
f. all dogs:
j. David’s dog:
k. his dog:
We have seen that there is a large degree of unity in the abstract functional syntactic
structure of languages. In contrast, phrasal structure varies greatly: some languages
allow phrases with no lexical heads, and some have no such categories; some
languages have a VP constituent, and others do not; and so on. In this chapter,
we will discuss the organization of overt phrasal syntactic representation, the
constituent structure or c-structure. We will explore commonalities in constituent
structure across the world’s languages, and talk about how languages can differ
in their phrasal organization. We will also consider the relation between phrasal
syntactic structure and the surface linear order of syntactic units; we consider the
surface linear order to be a separate level of representation, the syntactic string or
s-string.
Section 3.1 begins by discussing some traditional arguments for constituent
structure representation. Many of these arguments turn out to be flawed, since
the theory of phrase structure has a different status in LFG than in theories in
which grammatical functions are defined configurationally and abstract syntac-
tic (and even semantic and information structural) relations are represented in
phrase structure terms. Some arguments for particular phrase structure theories
and configurations are based on phenomena which in LFG are better treated
in other structures, such as f-structure, and do not constitute good arguments
for constituent structure at all. For this reason, we must examine the status of
arguments for and against particular phrase structures or phrase structure theories
particularly carefully. Section 3.2 proposes some valid criteria within LFG for
phrase structure determination.
Research on constituent structure has revealed much about the universally
available set of categories and how they can combine into phrases. We adopt a view
of phrase structure that incorporates insights primarily from the work of Kroeger
(1993), King (1995), Sadler (1997), Sells (1998), Toivonen (2003), and Bresnan
et al. (2016). We assume lexical categories like N and V, as well as functional
categories like I and C. These categories appear as the heads of phrases like NP
and IP; the inventory of constituent structure categories that are crosslinguistically
available and the theory of the organization of words and categories into phrases
are explored in §3.3. The general theory of constituent structure organization is
exemplified in §3.4, where we provide more specific discussion of the constituent
structure organization of clauses and the role of the functional categories I and C in
clausal structure. In §3.5, we discuss the relation between hierarchical constituent
structure and surface linear order.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
90 constituent structure
If there were no intervening representations between Sentence and words, the gram-
mar would have to contain a vast (in fact, infinite) number of conversions of the form
Sentence→X, where X is a permissible string of words. However, we find that it is possible
to classify strings of words into phrases in such a way that sentence structure can be stated
in terms of phrase strings, and phrase structure in terms of word strings, in a rather simple
way. Further, a phrase of a given type can be included within a phrase of the same type, so
that a finite number of conversions will generate an infinite number of strings of words.
Chomsky’s first point is that the same sequence of categories may appear in
more than one environment, and any adequate grammar must characterize this
regularity. For instance, the phrase the dachshund can appear in many positions in
a sentence:
Generalizations about the structure of the phrase the dachshund and the phrases
that can be substituted for it are captured by assuming that the dachshund is a
noun phrase and can appear wherever other noun phrases can appear. This is
intuitively appealing, since it captures the intuition that the dachshund (and the
black dachshund) are phrasal units in a sense that dachshund is or petted the are
not. Further, it simplifies the linguistic description: the different ways in which
a noun phrase can be formed do not have to be separately enumerated for each
environment in which a noun phrase can appear.
Chomsky’s second point is that a phrase may contain subconstituents of the
same type: a clause can contain a subordinate clause, a noun phrase can contain
other noun phrases, and so on. Our linguistic descriptions therefore need the same
kind of recursive character.
traditional arguments for constituent structure 91
This formal motivation for a level of constituent structure analysis and repre-
sentation is buttressed by a range of diagnostics for phrase structure constituency.
However, current syntactic theories vary greatly in their criteria for determining
the validity of these diagnostics, and tests that are accepted in some theoretical
frameworks are not recognized as valid in other frameworks. For instance, Chom-
sky (1981) and work growing out of the transformationally based theory presented
there (see, for example, Webelhuth 1995; Rizzi 1997; Chomsky 2001; Adger 2003)
propose that abstract syntactic properties and relations are represented in terms of
phrase structure trees, and that relations between these trees are statable in terms of
the movement of constituents (“move-α,” Move, or Internal Merge). In fact, in such
work not only syntactic relations, but also semantic and information structural
properties and relations, are represented in phrase structure terms. Given such a
theory, the criterion for constituenthood is fairly straightforward: any unit that
plays a role in abstract syntactic structure, semantic structure, or information
structure is a phrase structure constituent, since it must be represented as a unit
and must be eligible to undergo movement.
Such criteria have no relevance in a theory like LFG. We have seen that abstract
syntactic structure is represented by functional structure, not in phrasal terms;
further, phrase structural transformations and movement rules play no role in
LFG theory. We will see in Part II (Chapter 7 to Chapter 12) that semantic and
information structural features, as well as other nonsyntactic aspects of grammar,
are likewise represented by separate structures in LFG. LFG’s constituent structure
trees represent tangible phrasal configurations, not more abstract relations. Thus,
many criteria commonly proposed within the transformational tradition to iden-
tify phrase structure constituents turn out to be irrelevant in an LFG setting.
In this light, it is interesting to carefully examine one detailed proposal for
testing constituency within a transformationally based theory. Radford (2009:
58–69) discusses a set of tests for constituency, including the following:
Test (i), coordination, proves not to be very successful. It is well known that many
strings that are shown not to be constituents by other criteria can be coordinated:
(4) a. David gave [a flower] [to Chris] and [a book] [to Pat].
b. [David] [likes], and [Chris] [dislikes], carrots.
92 constituent structure
At the same time, sequences that are not constituents can be omitted:
(6) Chris can write better plays than David (can write) novels.
Even if, as Radford argues, this example involves two separate omitted constituents,
can and write, the usefulness of this test for determining constituency is seri-
ously undermined. The same problem applies to test (iii), substitution. Many
constituents cannot be replaced by a proform:
(8) a. Chris can eat one pie, but two pies he won’t be able to eat.
b. Yes, he will be able to do it.
(9) a. Chris ran into David yesterday.
b. They decided to see a movie.
Likewise, test (iv) is not fully reliable. While sentence fragments usually involve
constituents, there are constituents that cannot appear as sentence fragments (10b),
and nonconstituents that can (10c):
Although none of these tests are fully successful (as noted by Radford himself),
some prove to be distinctly less useful than others. This attests to the difficulty of
formulating unequivocal criteria for determining constituent structure units.
This argument can be repeated for many other verb-second languages, at least
for the majority of data: Kroeger (1993) discusses verb-second phenomena in
Germanic languages and in Tagalog; Zaenen and Kaplan (2002) and Cook and
Payne (2006) discuss the German data in more detail; and Simpson (1991, 2007)
discusses similar phenomena in Warlpiri.
Kaplan and Zaenen (1989b) propose that the distribution of the adverb depends on
the presence or absence of a VP constituent: an adverb cannot appear as daughter
of VP. Sentences with an auxiliary verb contain a VP constituent:
The adverb can appear as a daughter of S (or I" , in Sells’s 2005 analysis), but not
as a daughter of VP, accounting for the ungrammaticality of examples (13b,c). In
contrast, sentences with no auxiliary verb have no VP, allowing for the wider range
of possible adverb positions shown in (14).
falling intonation on its right edge; conversely, then, focus intonation may be
viewed as indicating alignment with a right-edge phrasal boundary. Here, the
feature +F represents this right-edge focus intonation, and the subscript focus
marks the focused constituent:2
See Zybatow and Mehlhorn (2000) for further discussion of focus intonation in
Russian. While this example clearly demonstrates alignment between prosodic
features such as intonation and syntactic groupings, we assume (with Lahiri and
Plank 2010) that such alignment is not necessary and, indeed, is not often found;
our approach to the formal relation between syntax and prosody will be presented
in Chapter 11.
Clitic placement: Zwicky (1990) proposes other “edge tests,” tests that pick
out the first or last word in a phrase. He argues that the distribution of the
possessive clitic in an English possessive noun phrase is best described by referring
to the right edge of the possessive phrase:
In this example, the possessor is my friend from Chicago, and the placement of
the clitic ’s shows that my friend from Chicago is a syntactic constituent (and not,
for example, two separate constituents my friend and from Chicago). The syntax of
the English possessive clitic is discussed in more detail by Payne (2009) and Lowe
(2016b).
Chomsky (1986) assumes that N, P, V, and A are major lexical categories; foll-
owing Jackendoff (1977), we also assume that Adv is a major lexical category.
These major lexical categories can head phrases of the corresponding category:
In addition to these major lexical categories, some LFG work assumes a set of
more crosslinguistically variable minor lexical categories. For example, Dalrymple
(2001), Butt et al. (2002a), and Forst et al. (2010) assume the category Part, the
category of English particles such as up in examples like (21):
Similarly, negative particles such as English not are sometimes assigned to a minor
category Neg (see, for example, Sells 2000a on negation in Swedish). Another minor
category is CL, which has been used to represent the category of clitics in some lan-
guages, for example in Romance languages (Grimshaw 1982a; Schwarze 2001a), in
Bosnian/Croatian/Serbian (O’Connor 2002a; Bögel et al. 2010), in Sanskrit (Lowe
2011, 2014a), and in Pashto (Lowe 2016a). Sequences of CL clitics that appear as
single unit (a “clitic cluster”) in a clause are sometimes grouped under a super-
ordinate “CCL” node (Bögel et al. 2010; Lowe 2011, 2014a; Lowe and Belyaev
2015).3
In contrast, some authors assume that there are no minor categories, and that
the inventory of c-structure categories is fixed. For example, Toivonen (2003) and
3 We use the term “clitic” here in purely syntactic terms; words that are treated as “clitics” on purely
prosodic grounds (such as “prosodic deficiency”) are not necessarily “clitics” in syntactic terms, and may
in principle be part of any major lexical or functional c-structure category. See Lowe (2014a, 2016a) for
further discussion of these issues.
constituent structure organization and relations 97
Bresnan et al. (2016) assume only four lexical categories, V, P, N, and A (treating
adverbs as part of A). Under such assumptions, apparent minor categories must be
analyzed as subtypes of major categories. English particles of the sort seen in (21),
for instance, are all homophonous with prepositions; it is therefore possible to treat
them as belonging to the category P. The c-structure possibilities for English not
partially match those of other English adverbs like never; not may therefore be
analyzed as a member of the category Adv (as in example (47) on page 108).
In fact, it is not necessary to propose a distinct minor c-structure category in
order to account for different c-structure constraints on different subsets of a single
c-structure category (for example, constraints on particles as opposed to preposi-
tions, or the negative not as opposed to other adverbs). The theory of nonprojecting
categories (Toivonen 2003), discussed in §3.3.4, permits a distinction between
members of the same c-structure category that project full phrasal structure and
those that do not. In addition, complex categories, discussed in §3.3.5, enable us
to model differences in c-structure constraints between words of any category that
differ in respect of some morphological or other feature.
In this work, we do not make use of any minor categories; all the words that
we analyze can be assigned to one of the five lexical categories stated in (19),
or else to one of the functional categories introduced in §3.3.3 below.4 In doing
this, however, we make no claim as to the inventory of c-structure categories in
Universal Grammar, and leave open the possibility that minor lexical categories
may exist in some languages.
4 However, we do assume a non-X" -theoretic category Cnj for conjunctions such as and; we postpone
discussion of conjunctions and coordinate structures to Chapter 16, where we provide a thorough
discussion of the syntax and semantics of coordination.
98 constituent structure
(22) X"
X0 YP ···
The X" node may dominate any number of daughter phrases; we do not assume
that constituent structure trees must be binary branching.
The X" category may be sister to a series of specifier phrases (ZP . . . ) and forms
an XP phrasal constituent with X" as its head:
(23) XP
ZP ··· X
Some languages allow only a single specifier phrase; other languages (for example,
languages like Russian, in which more than one question phrase can appear in
sentence-initial specifier position) allow multiple specifiers.
We also assume that an XP category can directly dominate a single X0 category,
as long as that is the only node it dominates. This permits us to simplify our
representation by omitting intermediate X" nodes that have no sisters and only one
X0 daughter:
(24) XP
X0
Besides these X" -theoretic structures, other permissible phrase structure config-
urations have been proposed. It is generally assumed that it is possible to adjoin
a maximal phrase to either a maximal (XP) or a nonmaximal (X" ) projection,
so-called “Chomsky-adjunction.” In example (25a), YP is adjoined to XP, and in
example (25b), YP is adjoined to X" :
(25) a. XP b. X"
YP XP YP X"
constituent structure organization and relations 99
Since we do not assume binary branching, it is possible to adjoin more than one
phrase at a time. In example (26), YP1 and YP2 are adjoined to XP:
(26) XP
YP1 YP2 XP
The adjoined phrase is often a modifier. Although the structures seen in examples
(25) and (26) are the most widely accepted structures that involve adjunction, they
are not the only types of adjunction that have been proposed. Another type, which
we accept in this work, is the adjunction of nonmaximal, nonprojecting categories
to other categories, which we discuss in §3.3.4.
3.3.3.1 The functional category I As Kroeger (1993) notes, the clausal syntax of
a number of languages that are genetically unrelated and typologically quite differ-
ent nevertheless organizes itself around a finite verbal element, either an auxiliary
or a main verb, which appears in a specified (often second) position; languages
as different as English, German, Warlpiri, and Tagalog all have a distinguished
position in which a finite main or auxiliary verb appears. The position in which
this element appears is called I (originally for INFL, i.e. inflection). The idea that
the functional category I is the head of a finite clause was originally proposed by
Falk (1984) in his LFG-based analysis of the English auxiliary system, and has since
been incorporated, in different ways, into both transformational and nontransfor-
mational analyses of clausal structure (Chomsky 1986, 1993; Pollock 1989; Kroeger
1993; King 1995; Nordlinger 1998; Cinque 1999, 2003, 2004; Bresnan et al. 2016).
Languages can differ as to which lexical categories can fill the I position. In
English, the tensed auxiliary appears in I, but nonauxiliary verbs may not appear
there (see the discussion in §3.4.1.1):
NP I"
N I VP
David is V
yawning
100 constituent structure
In other languages, such as Russian (King 1995), all finite verbs appear in I, as
shown in example (44) on page 107.
NP I"
N VP
David V"
V CP
knows C"
C IP
that NP I"
N VP
Chris V
yawned
In other languages, the rules may differ. King (1995: Chapter 10) provides a
thorough discussion of I and C in Russian, and Sells (1998, 2001b, 2005) discusses
the structure of IP in Icelandic and Swedish.
3.3.3.3 The functional category D Following work by Brame (1982) and Abney
(1987), many researchers have assumed that at least some instances of the category
traditionally labeled “noun phrase” are more accurately treated as a determiner
phrase or DP. According to this theory, the head of a phrase like the boy is the
determiner the, as shown in example (29). This contrasts with the more traditional
structure, shown in example (30), in which the noun is the head:
5 As example (28) shows, we do not assume that a phrase like IP must dominate an I head; we discuss
phrase structure optionality in §3.3.7.
constituent structure organization and relations 101
D" D N"
D NP the N
the N boy
boy
The projection DP was first assumed in an LFG context by Bresnan (1997, 2000)
and by Sadler (1997), who presents an analysis of Welsh pronominal clitics as Ds
within a more complete DP analysis; Börjars (1998) discusses the internal structure
of nominal phrases and the status of D in a range of languages.
DPs are now widely utilized in LFG work on many languages. The internal
structure of nominal phrases is complex, however, varying considerably between
languages. For example, Laczkó (2007) argues that Hungarian NPs can appear as a
complement of D, as in (29), but that (unlike English) a (possessive) DP can appear
as a specifier of NP, between the determiner and the noun. This is shown in (31),
based on Laczkó’s analysis of the Hungarian phrase az én kalapjaim ‘my hats’:6
(31) az én kalap-ja-i-m
the I hat-poss-pl-1sg
‘my hats’
DP
D"
D NP
az
DP N"
the
N
D
kalap-ja-i-m
én hat-poss-pl-1sg
I
In some languages, there is little empirical motivation for a category DP, while in
others there is much evidence. In this work, we will not make detailed proposals
concerning the internal syntactic structure of noun phrases; for simplicity, we will
use the traditional category NP for noun phrases like the boy wherever possible
(as in 30), using DP only where clearly motivated.
6 Laczkó’s (2007) proposal is actually somewhat more complicated, in that he assumes a complex
structure below the n. For simplicity, we omit the details of this structure here.
102 constituent structure
(32) lar.ke ne
boy erg
KP
K"
NP K
N ne
erg
lar.ke
boy
KPs are also assumed for Persian by Nemati (2010). In contrast, Spencer (2005a)
argues that Hindi ne is a postposition, analyzed as a nonprojecting P adjoined to
NP, as shown in (39).
Another functional category sometimes assumed is QP, for phrases headed by
a quantifier Q. Guo et al. (2007) assume a QP in their analysis of Chinese noun
phrases with classifiers, and Spector (2009) assumes a QP for Hebrew nominal
phrases. Wescoat (2007) assumes QP dominating DP for quantified noun phrases
in French:
Q DP
tout D
all
D NP
le N
the
personnel
personnel
constituent structure organization and relations 103
Occasionally other functional categories are proposed in order to account for the
complex structure of noun phrases in certain languages. For example, Laczkó
(2007) assumes a NumberP in Hungarian to account for the position of number
words in nominal phrases, and Bögel et al. (2008) and Bögel and Butt (2013)
propose an “Ezafe Phrase,” EzP, to account for the Urdu ezafe-construction, in
which a particle e (the ezafe) connects a modifying noun or adjective with a
preceding head NP. See Laczkó (1995) for an extensive discussion of the functional
and phrasal syntax of noun phrases within the LFG framework.
D N"
an N0
A0 N0
Adv0 A0 person
extremely happy
7 A “small” construction in Sadler and Arnold’s sense is different from a “small clause” (Stowell 1981),
a tenseless phrase consisting of a subject and a predicate.
8 There is a highly restricted set of exceptions to this generalization, which are not accounted for by
the structure proposed here. As noted by Arnold and Sadler (2013: 65), a restricted set of prenominal
adjectives can take phrasal complements, as long as those complements do not contain referential or
quantificational nouns. Examples include the eager to please student, a larger than average house, and a
difficult to solve problem. In some cases, these adjectival phrases have become conventionalized and
are treated as units in the syntax, but this cannot explain all such examples. A more wide-ranging
104 constituent structure
In subsequent work, Toivonen (2003) proposes that some words may be inher-
ently nonprojecting; that is, they never project full phrasal categories. Such non-
projecting words may be of any lexical category, and they adjoin to an X0 word just
like Sadler and Arnold’s “small” categories. Toivonen refers to projecting words as
X0 in the usual way, but distinguishes nonprojecting words with the notation ! X
(following Asudeh 2002c). The plain label X is used by Toivonen as a cover term
for X0 or !
X. Toivonen (2003: 2) proposes the structure in (36) for Swedish particle
verbs (ihjäl ‘to death’ is a particle):
NP I"
N0 I0 VP
V0 P
! N0
exception to the generalization made here is that almost any type of phrase can be used as a prenominal
modifier, for example I’m having one of those I’m-so-sick-of-this-bloody-job-that-I-could-scream days
(Arnold and Sadler 2013: 65 ex. 93). The particular intonation required with such examples suggests
that they require a separate account. Bresnan and Mchombo (1995: 194) propose that phrases can be
innovatively lexicalized, and thus used as if single words.
constituent structure organization and relations 105
(37) X0
X0 !
Y ···
Arnold and Sadler (2013) argue that nonprojecting categories must also be permit-
ted to adjoin to other nonprojecting categories, as in the following configuration:
(38) X
!
!
X !
Y
Spencer (2005a) argues that nonprojecting categories may also adjoin to full
phrasal categories; for example, he proposes that Hindi-Urdu postpositions are
nonprojecting !
Ps that adjoin to NP. His proposal implies the following analysis
of the Hindi-Urdu phrase lar.ke ne ‘boy erg’ (for a contrasting analysis, compare
example (32) on page 102):
(39) lar.ke ne
boy erg
NP
NP P
!
N ne
erg
lar.ke
boy
We do not assume that there are any necessary restrictions on what categories
nonprojecting words may adjoin to.
From now on in this book, we use X as an abbreviation for X0 ; nonprojecting
words are always referred to as !
X.
As we show in §5.1.4, this pattern can be generalized to other values for finiteness
by allowing the parameters of the rule to be instantiated in different ways, placing
fine-grained constraints on how different subtypes of categories can be realized.
Falk (2003, 2008) builds on Frank and Zaenen’s proposals in his analysis of the
English auxiliary system. Complex categories are also discussed by Kuhn (1999)
and Crouch et al. (2008).
As Simpson points out, if the V formed a VP constituent with the object phrase,
we would expect example (43) to be grammatical; in fact, however, it is not
grammatical:
Simpson (1991) presents a number of other arguments to show that Warlpiri has
no VP, although it does have a V" constituent.
In the case of functional categories, evidence is often less directly available;
here too, however, categories are only presumed to exist if they are motivated by
direct evidence. For instance, Laczkó (2016) argues that there is no evidence in
Hungarian for the existence of the category IP or I.
AdvP C"
Adv IP
kogda I"
when
I VP
rodilsja NP
born
N
Lermontov
Lermontov
108 constituent structure
3.4.1.1 The IP phrase In English, the tensed auxiliary verb appears in I, and the
rest of the verb complex appears inside the VP. Evidence that the English VP forms
a constituent comes from the fact that, like other constituents, it can be preposed
for information structure reasons:
(45) a. David wanted to win the prize, and [win the prize] he will.
b. . . . there is no greater deed than to die for Iran. And [dying] they are, . . .
(from Ward 1990)
It is not possible to prepose only the auxiliary and verb, since they do not form a
constituent:
(46) ⃰David wanted to win the prize, and [will win] he the prize.
clausal organization 109
Since the sentential negation morpheme not must be preceded by a tensed auxiliary
" right-adjoined to the tensed
verb, we assume that it is a nonprojecting adverb (Adv)
verb in I:
NP I"
N I VP
David I "
Adv V
is not yawning
It is not possible in (modern) English for the negation marker to follow a nonaux-
iliary verb (in contrast with French; see Pollock 1989):
Thus, in English only a tensed auxiliary verb can appear in I, and main verbs,
whether tensed or nontensed, appear inside the VP. When there is no auxiliary
verb, the I position is not filled and the IP is “headless”:
NP I"
N VP
David V
yawned
NP I"
N I VP
V VP
been V
yawning
110 constituent structure
In contrast to English, only nontensed verbs appear within the VP in Russian. King
(1995) provides several pieces of evidence supporting this claim. A VP constituent
that contains a nontensed verb can be scrambled to a higher clause in colloquial
Russian:9
A similar construction involving a tensed verb is not grammatical, since the tensed
verb is not a constituent of VP:
King concludes from this and other evidence, based on coordination and the
distribution of the negative marker ne, that nontensed verbs are of category V in
Russian. In contrast, all tensed verbs in Russian appear in I position (King 1995),
as shown in example (53):
I"
I VP
prislal NP V"
sent
N NP
muž N
husband
den’gi
money
C"
C IP
Is NP I"
N VP
David V
yawning
NP C"
N C IP
What is NP I"
N VP
David V
eating
maximal phrase XP of the same category. Headed categories like XP and X" are
called endocentric (Bloomfield 1933), and the tendency for languages to make
use of endocentric categories and for grammatical functions to be determined by
configurational syntactic relations is what Bresnan et al. (2016) call the principle of
endocentricity.
In contrast, some languages allow an exocentric category, one that has no
head: the category S (Bresnan 1982a; Kroeger 1993; Austin and Bresnan 1996;
Nordlinger 1998). According to Bresnan et al. (2016), languages making use of
the category S are lexocentric, meaning that syntactic functions are determined
by features associated with words rather than by configurational structure. S is a
constituent structure category that can contain a predicate together with any or all
of its arguments and modifiers, including the subject; for this reason, Austin and
Bresnan (1996) call languages with the category S, including Tagalog, Hungarian,
Malayalam, and Warlpiri, “internal subject” languages. As an exocentric category,
the category S can dominate a series of either lexical or phrasal constituents. Some
authors permit the category S even in thoroughly endocentric languages such as
English (Bresnan et al. 2016), but we restrict its use to those languages which show
clear evidence for exocentric structure.
According to Austin and Bresnan (1996), the phrase structure of a simple
Warlpiri sentence is as follows:10
NP I"
N I S
kurdu-jarra-rlu kapala NP V NP
child-du-erg prs
N wajilipi-nyi N
chase-npst
maliki wita-jarra-rlu
dog-abs small-du-erg
The Warlpiri I" consists of an I, in which the clitic auxiliary appears, and an
S complement. A phrasal constituent of any category appears as the specifier
daughter of IP; in (56), a noun phrase fills the specifier of IP. Importantly, there
are no syntactic constraints on the order of words in the sentence, so long as the
auxiliary appears in second position (Simpson 1991; Austin and Bresnan 1996).11
10 Legate (2002), Laughren (2002), and Simpson (2007) show that many Warlpiri sentences are in
fact more complicated than this, insofar as various projections exist above the IP, but this does not affect
the phrasal configuration below the IP.
11 The order of words is highly constrained according to other, namely information structural,
factors (Simpson 2007).
clausal organization 113
The daughters of S can in principle appear in any order, and no element obligatorily
appears as daughter of S, since any phrasal constituent can appear in the specifier
of IP position.
In Chapter 4, we discuss the close correlation between specifier and complement
constituent structure positions as defined by X" theory and the grammatical
functions of the phrases that appear in those positions. For example, in many
languages the subject must appear in the specifier position of IP. In contrast, the
relation between constituent structure positions dominated by S and their abstract
functional role does not obey the same X" -theoretic constraints. Thus, in languages
which, like Warlpiri, make use of the exocentric category S, constituent structure
position is often not an indicator of grammatical function. Instead, grammatical
functions are often marked morphologically, by means of case endings, with a
concomitant tendency to freer word order.
The category S has also been proposed for languages that have relatively fixed
word order. Sadler (1997) proposes the following clausal structure for Welsh, a
VSO language:
I S
Gwnaeth NP VP
do.3sg.pst
N V NP
hi weld
y draig
3sg.f see
the dragon
On Sadler’s analysis, the tensed verb in Welsh is always of category I. The comple-
ment of I is S; S dominates the subject NP and a predicate XP, which may be a VP,
as in example (57), or a maximal phrase of different category, for example NP or
AP in copular constructions (on copular constructions, see §5.4.5). Similar clausal
structures are assumed by Kroeger (1993) and Asudeh (2012: 157–63) for Irish,
and by Falk (2001b: 51–2) for Hebrew.
A further functional category E (for Expression or External-Topic), originally
proposed by Banfield (1973), has been proposed in analyses of left-dislocated
phrases, such as external topics that are intonationally and syntactically separated
from the rest of the clause. In (58), David is an external topic (and not a vocative),
and the syntactic separation of David from the rest of the sentence is clear from
the fact that a resumptive pronoun appears and a gap is not allowed:
Within LFG, the category E is used by King (1995), building on work by Aissen
(1992), in her analysis of Russian c-structure, and also by Jaeger and Gerassimova
(2002) for Bulgarian, Simpson (2007) for Warlpiri, and Lowe (2012) for Sanskrit.
However, it is often possible to account for external topics and similarly positioned
phrases by assuming adjunction to a clausal node, without recourse to an addi-
tional phrase-structure category. In example (58), for example, the NP David may
be analyzed as adjoined to CP. We have no need for E in this work, and so will make
no use of it.
This may appear trivial, but it is important for several reasons. Lowe (2016b)
adapts the theory of “Lexical Sharing” developed by Wescoat (2002, 2005, 2007,
2009), according to which it is possible for single items in the s-string (Wescoat’s
“linear structure”) to be associated with two terminal nodes in the c-structure. In
French, for example, some sequences of preposition followed by determiner are,
unremarkably, sequences of two separate words, but certain other sequences of
preposition followed by determiner are single words:
(60) a. à la fille
to the girl
b. ⃰à le garçon
to the boy
the s-string 115
c. au garçon
to.the boy
P DP P DP
D NP D NP
N N
à la fille au garçon
to the girl to.the boy
The arrows between the s-string and the c-structure represent the fact that these
are not constituency relations, but a “projection” from one syntactic level (the
s-string, encoding linear order) to another (the c-structure, encoding hierar-
chical relations). This is compatible with Carnie’s (2012b: 93) observation that
the relation between words (s-string units) and the c-structure is different from
the relation between c-structure nodes.12 The s-string and its relation to c-structure
are discussed in more detail in Chapter 11.
Lexical Sharing is also assumed by Broadwell (2007, 2008) in his analyses of
Zapotec and Turkish, by Alsina (2010) in his analysis of Catalan, by Belyaev
(2014) in his analysis of Ossetic, and by Lowe (2015d) in his analysis of Sanskrit
compounds. Mismatches between the s-string and c-structure may also be relevant
to the phenomenon of “prosodic inversion,” examined by Halpern (1995) and, in
an LFG setting, by Kroeger (1993), Newman (1996), Austin and Bresnan (1996),
Nordlinger (1998), Bögel et al. (2010), Bögel (2010), and Lowe (2011). Lowe
(2016a), building on a suggestion by Dalrymple and Mycock (2011), argues that the
observed distinction between what Halpern calls “surface order” (the order of the
words in a sentence) and “syntactic order” (their order in the constituent structure
tree) is best captured as a difference in the order of items between the s-string and
c-structure, rather than a distinction between syntax and prosody, as is assumed
12 We do not adopt Carnie’s convention of representing this difference by omitting the line between
the s-string unit and the c-structure.
116 constituent structure
in many previous works. Other authors who investigate different aspects and uses
of the s-string include Mycock (2006) and Asudeh (2006, 2009).
Like other linguistic levels, the s-string is subject to wellformedness constraints,
which may differ on a language-by-language basis. These constraints will be
discussed in more detail in §6.11 and Chapter 11. In §11.4.1, we will see that a
considerably more refined treatment of the string enables us to analyze interactions
involving syntax, prosody, and other aspects of the linguistic structure of an utter-
ance. For certain phenomena, then, the s-string, as a separate syntactic component,
is an essential component of analysis. Nevertheless, in the vast majority of cases,
as in examples (57) and (61), the relation between the s-string and c-structure
is straightforward. In the following chapters, we explicitly represent the s-string
as separate from the c-structure only where necessary; otherwise, we assume the
separate s-string and the arrows connecting s-string units to terminal c-structure
nodes only implicitly, and we maintain the simpler representation (which does not
include arrows), as illustrated in the previous sections of this chapter.
LFG hypothesizes that constituent structure and functional structure are mutually
constraining structures and that the relation between these structures is governed
by constraints associated with words and phrasal configurations. We turn now to
an investigation of the relation between these two facets of linguistic structure: in
this chapter, we explore universally valid generalizations regarding the correlation
between phrasal positions and grammatical functions.
Section 4.1 discusses the formal representation of the relation between c-
structure and f-structure. Section 4.2 explores the relation between c-structure and
f-structure: how c-structure phrases and their heads relate to f-structure, and the
c-structure/f-structure realization of arguments and modifiers. Next, we exam-
ine apparent mismatches between units at c-structure and those at f-structure;
Section 4.3 shows that these cases have a natural explanation within LFG. In
§4.4, we discuss the Lexical Integrity Principle, the concept of wordhood, and the
possibly complex contribution of words to functional structure, and in §4.5, we
discuss the principle of Economy of Expression.
(1) V
’
yawned
As this diagram shows, the word yawned is of category V, and the V node is associ-
ated with certain functional syntactic information: the f-structure corresponding
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
118 syntactic correspondences
to the V node has a pred with value ‘yawn!subj"’, and the attribute tense with
value pst.
Often, more than one c-structure node is related to the same f-structure.
For instance, a phrase and its head correspond to the same f-structure:
(2) VP
’
V
yawned
There are also f-structures that are not related to any c-structure node. In
Japanese (Kameyama 1985), a “pro-drop” language, the single word kowareta
‘broke’ can appear without an overt subject noun phrase. In such cases, the subj
f-structure is interpreted pronominally, and does not correspond to any node in the
c-structure:
(3) koware-ta
break-pst
‘[It/Something] broke.’
S
’
V
kowareta
4.2.1 Heads
The functional properties and requirements of the head of a phrase are inherited
by its phrasal projections and become the functional properties and requirements
of the phrases projected by the head. This means that a constituent structure head
and the phrases it projects are mapped onto the same f-structure, as shown in (2).
This condition was originally proposed by Bresnan (1982a: 296) and discussed by
Zaenen (1983), who calls it the Head Convention.
regularities in the c-structure/f-structure mapping 119
4.2.2 Complements
4.2.2.1 Complements of functional categories Recall that complement posi-
tions are sisters to the head of a phrase. Complements of functional categories
are f-structure co-heads (Kroeger 1993; King 1995; Bresnan et al. 2016), meaning
that a functional category shares its functional properties with its complement.
For example, the English IP shares the functional syntactic properties of its VP
complement. Thus, an f-structure can be associated with two different c-structure
heads, one a functional category and the other a lexical category, as shown in the
English example in (4) and the Russian example in (5) (from King 1995: 227):
N I VP
David is V
yawning
čitat’ N
knigu
book
David V
V NP
greeted N
Chris
Of course, more than one complement may appear; in example (7), the first
complement bears the obj role, and the second complement is objtheme :
NP I
’
N VP
David V
−
V NP NP
gave N D N
Chris a N
book
The complement of a lexical category can also correspond to the same f-structure
as the head of the phrase, as in the case of English nonfinite auxiliaries and their
complement VPs (Bresnan et al. 2016: Chapter 6):
NP I
’
N I VP
David has V
V VP
been V
yawning
Falk (1984, 2001b), Butt et al. (1996b), Sadler (1998), and Bresnan et al. (2016)
provide further discussion of the English auxiliary system.
regularities in the c-structure/f-structure mapping 121
As É. Kiss (2002) and others have observed, the different orders of postverbal
arguments in Hungarian are essentially equivalent in terms of interpretation.
Laczkó (2014a) proposes that the subject can appear within V$ in Hungarian,
as a complement of the lexical category V, in the same position as postverbal
complements bearing other governable grammatical functions.
Since LFG does not define subcategorization properties of predicates in c-
structure terms, the arguments of a predicate may but are not required to appear in
complement positions. They may, for example, appear in the specifier position of
some projection, and are linked to their within-clause function by means of purely
functional specifications. Resolution of long-distance dependencies is discussed
briefly in the next section, and in more detail in Chapter 17.
4.2.3 Specifiers
Recall that the specifier of a constituent structure phrase is the daughter of XP
and sister to X$ . We propose that specifier positions are filled by phrases that
are prominent either syntactically or in information-structural terms (compare
Bresnan et al.’s “Clause-prominence of DFs” principle: Bresnan et al. 2016: 209).
Syntactically prominent phrases that can appear in specifier positions in the clause
are those bearing either the function subj or the overlay function dis heading
a long-distance dependency (see §2.1.11). Information-structurally prominent
phrases can also appear in specifier position; if they are not syntactically promi-
nent, they may bear any grammatical function within the local clause. The table in
(10) shows the permitted grammatical and information structure roles of phrases
appearing in specifier positions (see Snijders 2015 for further discussion):
In this, our position differs from some other LFG work (Dalrymple 2001; Bresnan
et al. 2016), which assumes that specifier positions are filled by phrases bearing
122 syntactic correspondences
particular f-structure roles: subj, topic, or focus. Following Alsina (2008) and
Asudeh (2012), we assume that topic and focus are information structure roles,
and are not represented at f-structure. Rather, phrases that play the role of topic or
focus at information structure can appear in a specifier position with no necessary
restriction on the grammatical function that these phrases can bear, or in other
positions within the clause. See Chapter 10 for more discussion of information
structure in LFG’s projection architecture, including a historical overview of the
treatment of information structure in LFG.
SUBJ in specifier position: In English and many other languages, the specifier
position of IP is associated with the subj function:2
NP I
’
N VP
David V
yawned
King (1995) shows that only subjects can appear in the specifier position of VP
in Russian:
2 The tree in example (12) has an IP and an I$ node, but no I node; recall from our discussion in
§3.3.7 that no c-structure position is crosslinguistically obligatory, and heads are often optional.
regularities in the c-structure/f-structure mapping 123
I VP
’
posadil NP V
planted
N NP
ded N
old.man
repku
turnip
David V
eating
The value of dis in (14) is a set because more than one constituent can bear the
dis function in a single f-structure. This is the case, for instance, in a version of the
sentence in (14) in which David is left dislocated (David, what is he eating?), and
the f-structures for David and what are in the dis set. However, while this analysis
is appropriate for English, the same is not necessarily true for all languages. In this
book, we assume unless we are aware of evidence to the contrary that the value of
dis is a set. However, this must be established independently, on a language-by-
language basis.
124 syntactic correspondences
The dis phrase can also bear a grammatical function in an embedded clause
(for more discussion, see §6.1.2 and Chapter 17). In (15), dis in the main clause is
identified with the obj in the subordinate clause:
CP
NP C ’
N C IP
What does NP I ’
N VP
Chris V IP
think NP I
N VP
David V
ate
3 For more on the information structure functions of question words, see Mycock (2013) and
Chapter 17.
regularities in the c-structure/f-structure mapping 125
Example (17) shows that a phrase in specifier position can play a syntactic role in
a subordinate clause. Here, Ivan is in the specifier position of CP and kakvo is in
the specifier position of IP in the main clause; they bear the syntactic roles of subj
and obj in the subordinate clause:
Mari-t szereti
loves
126 syntactic correspondences
Laczkó (2014a,b) goes further, providing a detailed analysis of the sorts of phrases
that can fill the specifier of VP in Hungarian; he argues that besides hosting
phrases associated with the information structure role of focus (including question
phrases), the specifier of VP can also host verbal modifiers, a category including
verbal particles as well as parts of idiomatic expressions and certain object or
oblique arguments. He also discusses the consequences of the Hungarian data for
LFG assumptions regarding regularities in the c-structure/f-structure mapping.
NP I
Evgenija Onegina ’
I VP
Eugene Onegin
napisal NP
wrote
N
Puškin
Pushkin
nonprojecting words: that is, lexical and functional categories that are not the heads
of full phrases. Toivonen (2003: 68) proposes the following principle governing
the f-structural role of nonprojecting categories, according to which nonprojecting
categories cannot be adjuncts:
NP
D N
a N0
A N0
happy person
We have not found motivation for assuming minor categories in this work, and so
we will not adopt any particular view about the permitted f-structure roles of such
categories.
N I S
wawirri ka NP V
kangaroo
N panti-rni
ngarrka-ngku
“movement” and discontinuity 129
N I S
ngaju ka-rna V
parnka-mi
Examples such as these need no special analysis within LFG, and the verb is
not thought of as having “moved” to the position in which it appears. Rather,
the principles we have outlined so far predict that this configuration is possible
and wellformed. In Russian, all tensed verbs have the phrase structure category I;
the verb in (24) appears in I and not within VP. It is not possible for there to
be two lexical verbs in a single Russian sentence, one in the I position and one
within the VP, since this would produce an ill-formed f-structure; each lexical verb
contributes a pred value to its f-structure, and the Consistency Principle forbids
130 syntactic correspondences
an f-structure from containing a pred attribute with two different semantic forms
as its value. Conversely, a sentence with no verb is ill-formed, since in that case the
main f-structure would contain no pred, and the Completeness Principle would be
violated. Exactly one lexical verb must appear, and it must appear in the appropriate
position for its constituent structure category. Bresnan et al. (2016: Chapter 7)
provide more discussion of such cases, which they refer to as head mobility or
variable head positioning.
NP I
’
N I VP
Juan N I V
Juan
lo vió PP
saw
P
P NP
a N
Pedro
Pedro
“movement” and discontinuity 131
Thus, exactly one object phrase must appear, and it must appear in the constituent
structure position appropriate for its phrase structure category.
Interestingly, however, complementary distribution between a clitic and a full
noun phrase is not always found. Andrews (1990b) discusses dialects of Spanish
which allow clitic doubling, in which it is possible for both obj positions to be filled,
as in (27):
NP I
’
N I VP
Juan N I V
Juan
lo vió PP
saw
P
P NP
a N
Pedro
Pedro
Clitic doubling is possible in the River Plate and Peruvian dialects of Spanish as
described by Jaeggli (1980) (see also Grimshaw 1982b; Estigarribia 2005; Mayer
2006; Bresnan et al. 2016; and, on Limeño, Mayer 2008). In these dialects, the object
clitic pronoun lo is undergoing reanalysis as an agreement marker, and has begun
to lose its semantic status as a pronominal. Formally, this reanalysis is reflected
as optionality of the pred value of the clitic (see §5.2.4): lo need not contribute a
pred value to the f-structure. Since its other features (pers 3, num sg, case acc)
are compatible with the full phrasal obj, both the clitic and the full noun phrase
can appear in the same sentence.
IP
NP IP
D N NP I
that N N I VP
problem he didn’t V
V PP
think P
of
“movement” and discontinuity 133
In (30), there is no sense in which the phrase that problem has been moved
to clause-initial position from a position within the clause. Rather, the relation
between the initial NP and its within-clause function is specified in functional
rather than constituent structure terms, by functional uncertainty (see §6.1.2),
and phrase structure category is not a part of the specification. Exactly the same
constraints allow for a clause-initial CP to fill the role of object of the preposition
of, as shown in (31):4
IP
CP IP
C NP I
C IP N I VP
that NP I he didn’t V
N I VP V PP
he might V think P
V AP of
be A
wrong
4 The f-structure for the clause that he might be wrong has been simplified; the functional structure
of copular constructions is discussed in detail in §5.4.5.
134 syntactic correspondences
IP
NP I
N I S
kurdu-ngku ka V NP
wajilipi-nyi N
wita-ngku
These two phrases are functionally compatible, since both bear erg case: one
represents the head of the subject phrase, and the other represents a modifier. Here
too, we need say nothing special about examples involving constituent structural
5 There is no categorical distinction between adjectives and nouns in Warlpiri, and we follow
Simpson in categorizing both kurdu-ngku ‘child’ and wita-ngku ‘small’ as nouns. Simpson shows that
there are two readings for example (32); in the merged reading, the one which is of interest here,
wita-ngku ‘small’ is interpreted as a restrictive modifier of kurdu-ngku ‘child’. The other reading is the
nonrestrictive, “unmerged” interpretation, which Simpson paraphrases as ‘The child is chasing it and
she is small’; on Simpson’s analysis, this reading has a different syntactic structure.
the lexical integrity principle 135
discontinuity such as these; our theory predicts that they will occur, and we need
no special analysis to encompass them.
(33) Lexical Integrity Principle, Lapointe (1980: 8): No syntactic rule can refer to
elements of morphological structure.
(34) Lexical Integrity Principle, Bresnan et al. (2016: 92): Morphologically com-
plete words are leaves of the c-structure tree, and each leaf corresponds to
one and only one c-structure node.
Current work in LFG adheres to this principle: LFG does not assume the exis-
tence of processes that assemble word forms in constituent structure or reorder
subparts of word forms during syntactic composition. In this way, the LFG view
contrasts with approaches such as Distributed Morphology (Halle and Marantz
1993; Embick and Noyer 2007) and the Exo-Skeletal Model (Borer 2013), which
reject the Lexicalist Hypothesis. We will not address this controversy in detail
here; cogent defenses of the Lexicalist Hypothesis are provided by Bresnan and
Mchombo (1995), Spencer (2005b, 2016), Stewart and Stump (2007), and refer-
ences cited there. We provide a detailed discussion of the role of morphology in
the LFG architecture in Chapter 12.
Although words are not subject to processes of assembly in the constituent
structure, their syntactic contributions can be complex. A word that forms a unit
at the level of c-structure may introduce complex functional structure, structure
that in other languages might be associated with a phrase rather than a single
word. For example, Bresnan and Mchombo (1987) present evidence that subject
and object markers on the Chicheŵa verb can function as incorporated pronouns.
These affixes form a part of the verb at c-structure, meaning that the f-structure
corresponding to this verb is complex, similar to the f-structure for a sentence in
a language like English.7 In (35), the marker zi- represents a pronominal subj of
noun class 10, and the marker -wá- represents a pronominal obj of noun class 2:
(35) ’
V
zi-ná-wá-lúm-a
As this example shows, the notion of word is multifaceted and must be fully defined
at each syntactic level of representation; units at c-structure need not correspond
to simple units at other levels.
Much other work has been done within LFG in support of the strong lexicalist
hypothesis, demonstrating convincingly that a word that is atomic at the level of
c-structure can project the structure of a phrase at functional structure. Simpson
(1983, 1991), Ackerman (1987), O’Connor (1987), Mohanan (1994, 1995), Bres-
nan and Mchombo (1995), Nordlinger and Bresnan (1996), Matsumoto (1996),
Sadler (1997), Ackerman and LeSourd (1997), Börjars (1998), Alsina (1999),
Broadwell (2008), Asudeh et al. (2013), and Lowe (2016b) have all made valuable
contributions to our understanding of wordhood and lexical integrity.
7 See §5.4.3 and §14.1 for more discussion of the pronominal status of the subject and object markers
in Chicheŵa.
economy of expression 137
N$ N
N David
David
(38) Economy of Expression: All syntactic phrase structure nodes are optional
and are not used unless required by X$ -constraints or Completeness. (Toivo-
nen 2003: 200)
N$ N
N David
David
Both definitions of Economy proceed from the claim that every node in a
c-structure tree is optional. In fact, the issue of optionality of nodes as a general
principle of the grammar is separable from particular definitions of an Economy
principle. As discussed in §3.3.7, there are no phrase structure positions that are
obligatory in every phrase and in every language: a phrase XP need not dominate a
138 syntactic correspondences
head X in the tree, and the complements of a lexical predicate X need not appear as
sisters to X. However, Snijders (2012, 2015) shows that in some configurations in
some languages, c-structure nodes are obligatory and not optional: in particular,
Snijders shows that the NP complement of P in the Latin PP is obligatory. The
existence of obligatory nodes is not incompatible with the adoption of an Economy
principle: even if some nodes in some constructions are obligatory, an Economy
measure can still be used to choose among larger and smaller trees that otherwise
obey the rules of the grammar.
Dalrymple et al. (2015) discuss the various versions of Economy which have
been proposed, providing explicit formal definitions of three versions of the
Economy metric, and an overview of analyses which have appealed to an Economy
measure. They discuss whether Economy is best treated as a very general principle,
as on Toivonen’s and Bresnan et al.’s approaches, or if it is preferable to appeal
to more specific principles for particular subcases rather than a general Economy
principle.
Up to this point, our discussion has concentrated on the nature and represen-
tation of the two syntactic structures of LFG. We will now demonstrate how
to formulate descriptions of and constraints on c-structure, f-structure, and the
relation between them, and we will see how these constraints are important in the
statement of universal typological generalizations about linguistic structure. These
constraints are a part of the formal architecture of LFG theory.
The job of the designer of a formal linguistic framework such as LFG is to
provide a way of stating linguistic facts and generalizations clearly and precisely,
in a way that is conducive to a solid understanding of the linguistic structures that
are described and how they are related. Designing an appropriate representation
for a linguistic structure is very important in helping the linguist to understand
it clearly and in avoiding confusion about the properties of the structure and its
relation to other structures. A formal linguistic theory must provide efficient and
transparent ways of describing the facts of a language. Linguistically important
generalizations should be easy to express, and the linguistic import of a constraint
should be evident.
The job of a linguist working within a formal framework is to discover the facts
of a language or of languages and to express these facts by using the tools provided
by the designer of the framework so that the linguistic facts emerge precisely and
intuitively. Most importantly, the linguist must distinguish between constraints
that are needed and used in describing the facts of language, the linguistically
relevant constraints, and those that may be expressible within the overall formal
framework but do not play a role in linguistic description. A well-designed formal
framework aids the linguist in deciding which constraints are relevant and which
are not in the description of linguistic structures.
(1) S −→ NP VP
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
140 describing syntactic structures
This rule permits a node labeled S to dominate two nodes, an NP and a VP, with
the NP preceding the VP. In LFG, phrase structure rules are interpreted as node
admissibility conditions, as originally proposed by McCawley (1968): a phrase
structure tree is admitted by a set of phrase structure rules if the rules license
the tree. In other words, phrase structure rules are thought of as descriptions
of admissible trees, and the trees of the language must meet these descriptions.
McCawley’s groundbreaking work constituted an important alternative to the way
of thinking about phrase structure rules that was prevalent in the mid-1960s, which
viewed phrase structure rules as a procedural, derivational set of instructions to
perform a series of rewriting steps.
In many theories of phrase structure specification, all phrase structure rules
are of the type illustrated in (1): the right-hand side of the rule (here, NP VP)
specifies a particular unique admissible configuration. Constituent structure rules
in LFG are more expressive than this, in that the right-hand side of a phrase
structure rule consists of a regular expression (Kaplan and Bresnan 1982: 277; see
§6.1 for a full discussion and definition of regular expressions). The use of a regular
expression allows a sequence of category labels in which some categories may be
optional, some categories may be repeated any number of times, and disjunction
is permitted.
Since phrase structure rules represent constraints on phrase structure configura-
tions, it is reasonable and desirable to abbreviate a large or even an infinite number
of rules by means of a regular expression. A phrase structure rule is not thought
of as a large or infinite set of rules that can be applied individually in a series of
rewriting steps, but as a characterization of the daughters that nodes of a phrase
structure tree can dominate. This constraint-based view of linguistic descriptions
pervades the formal theory of LFG: a grammar of a language consists of a set of
constraints on linguistic structures, and these constraints admit only wellformed
analyses of utterances.
A simple LFG phrase structure rule can have the following form:
This rule indicates that either an NP or a PP can appear in the specifier position of
IP. The curly brackets mark a disjunction of phrase structure categories, with the
possibilities separated by a vertical bar |. This rule abbreviates the following two
rules, where the disjunction in (2) has been fully expanded into two separate rules:
(3) a. IP −→ NP I#
b. IP −→ PP I#
The abbreviation in (2) is not only more compact than the two-fold expansion
in (3), but more revealing of the linguistic facts: the specifier position of IP can
be filled either by NP or by PP without affecting any properties of the second
daughter of IP, namely I# . Stating this fact by means of two separate rules, as in
(3), makes it appear that the I# might have different properties in the two different
constituent structure rules 141
rules, since there is no apparent relation between the I# in the first rule and the I#
in the second rule.
In (4), the parentheses around the NP indicate optionality. The Kleene star
annotation ⃰ on the PP indicates that zero or more PPs may appear in the expansion
of the rule; more discussion of the meaning and use of the Kleene star is given
in §6.1:
Thus, this rule admits trees in which a V# node dominates a V, an optional NP, and
any number of PPs. The use of the Kleene star means that an infinite set of possible
rules—rules with any number of PPs—can be abbreviated with a single expression.
Thus, phrase structure generalizations can be stated once rather than separately
for each specific phrase structure configuration licensed by the rule.
We generally follow standard practice in displaying trees with node labels repre-
senting the nodes which they label, meaning that we represent the tree in (5) as in
(6). The node labeling function λ will occasionally be important in our subsequent
discussion, however.
(6) S
NP VP
(7) CP −→ XP C#
IP −→ XP I#
The use of XP in these rules indicates that a full phrase of any category (NP, PP,
and so on) can appear as the first daughter of CP and IP. XP is defined as follows:1
The use of abbreviations like this allows for the expression of general statements
about all phrases that appear in a particular phrase structure position.
In (8), the metacategory XP stands for any one of a number of phrasal categories.
In fact, a metacategory can abbreviate a longer sequence of categories, not just
a single category (Kaplan and Maxwell 1996). This is shown in the following
putative definition of the metacategory PREDICATE:
(9) PREDICATE ≡ V NP
(10) S −→ NP PREDICATE
Given this definition, the rule in (11) produces exactly the same result as the rule
in (10):
(11) S −→ NP V NP
The phrase structure rule in (10) and the definition of PREDICATE in (9) admit the
following tree:
(12) S
NP V NP
1 The symbol ≡ connects two expressions that are defined to be equivalent; the expression in (8) can
be read as: “XP is defined as the disjunction {NP | PP | VP | AP | AdvP}”; that is, XP is a phrase with
category NP, PP, VP, AP, or AdvP.
constituent structure rules 143
This pattern can be generalized to other values for finiteness by using parametrized
rules to allow information about these features to flow through the c-structure tree,
placing fine-grained constraints on how different subtypes of categories can be
realized. In Frank and Zaenen’s grammar, there are three possible instantiations
of the finiteness parameter: fin (finite), inf (infinitive), and part (participle). The
requirement for identical values of the finiteness parameter for the VP and the
initial V is captured by the use of variables over features. Notationally, parameter
variables are prefixed by an underscore: VP[_ftness] and V[main,_ftness] both con-
tain occurrences of the variable _ftness. Within the same configuration, identical
144 describing syntactic structures
variables must match. This means that the rule in (14a) is a succinct abbreviation
of the three rules in (14b):
(14) a. −→
b. −→
VP[inf] −→ V[main,inf]
VP[part] −→ V[main,part]
Thus, the rule in example (15) licenses the c-structure in example (13), as well
as any other c-structures in which the finiteness value of the initial V daughter
matches the finiteness value of the VP:
This analysis requires the French finite auxiliary est to have the parametrized
category V[aux,fin]:
(17) VP −→ V, NP
This rule states that a VP node dominates two other nodes in a tree, a V node and
a NP node, but does not specify the order of V and NP. Thus, it can be regarded as
an abbreviation for the two rules in (18), in which the order is fully specified:
(18) VP −→ V NP
VP −→ NP V
(19) VP −→ V, NP V<NP
The ID phrase structure rule in (17), combined with the linear precedence con-
straint V<NP stating that V must precede NP, is equivalent to the standard
ordered phrase structure rule VP −→ V NP. A more complicated example is
given in (20):
The ID phrase structure rule requires VP to dominate three nodes, V, NP, and PP.
The LP ordering constraints require V to precede both NP and PP, but do not place
any constraints on the relative order of NP and PP. Thus, the ID rule and constraint
in (20) are equivalent to the rule in (21), in which ordering is fully specified, and
NP appears either before or after PP:
ID/LP rules are used in King’s (1995) analysis of Russian. She proposes rules of the
following form:
(22) a. CP −→ XP, C#
b. C# −→ C, IP
c. IP −→ XP, I#
d. XP<Y# , Y<XP
(23)
+
The f-structure in (23) contains two attribute-value pairs: %pred, ‘man’& and
%def, +&. We can place various requirements on f-structures: they may be required
to contain certain attribute-value pairs or one of several possible attribute-value
pairs, or they may be required not to contain certain material. The following
sections explain how such constraints can be imposed.
(24) (m def) = +
This f-structure, which represents the expression this man, contains the attribute
def with value +; it also contains additional attributes and values.
2 As explained in Chapter 2, Footnote 21 (page 44), what is depicted in example (23) is an f-structure,
a formal object, not a constraint on f-structures.
functional constraints 147
The f-structure in (25) is special in that it is the smallest f-structure that satisfies
the constraint in (24). We call such an f-structure the minimal solution to the
f-description in (24): it satisfies all the constraints in the f-description, and it has
no additional structure that is not relevant in satisfying the constraints.
We require the f-structure for any utterance to be the smallest f-structure that
satisfies all of the constraints imposed by the f-description for the utterance, and
has no additional properties not mentioned in the f-description:
(27) The f-structure for an utterance is the minimal solution satisfying the con-
straints introduced by the words and phrase structure of the utterance.
For these cases, the following definition is relevant (Kaplan and Bresnan 1982;
Kaplan and Zaenen 1989b):4
This definition tells us that an expression like ( f subj def) denotes the same
f-structure as (( f subj) def): that is, the f-structure that appears at the end of the
path subj def in the f-structure f is the same f-structure that appears as the value
of the attribute def in the f-structure ( f subj). Longer expressions are treated
similarly. The second part of the definition tells us that the empty path 𝜖 can be
ignored: the expression ( f 𝜖) is the same as f .
3 The symbol ∈ is the set-membership symbol: the expression x ∈ Y (“x is in (the set) Y”) means
that x is a member of the set Y. Chapter 6, §3 provides a detailed discussion of predicates and relations
involving sets.
4 The epsilon symbol 𝜖, which represents the empty string or path, is not the same as the set
membership symbol ∈.
148 describing syntactic structures
A Hindi sentence like (31) (McGregor 1972) is associated with the f-description
given in (32) and (33) (some detail has been omitted):
The constraints in (32) come from the lexical entries for the proper noun Raam
and the verb calegaa ‘will go’; here, we consider only case and num requirements
imposed by the verb on its subject, ignoring tense, person, and gender features:
We also know that Raam is the subject of this sentence: that is, the f-structure r
for Raam is the subj of the f-structure c for the sentence. This is encoded in the
constraint in (33):
(33) (c subj) = r
Given this equality, we can substitute r for (c subj) in the constraints in (32). We are
then left with the set of constraints in (34), which are equivalent to the constraints
in (32) and (33):
Notice that some of the equations in the f-description constrain the same attribute-
value pairs. For example, the proper noun Raam requires its f-structure r to have
an index attribute specifying singular number (containing the attribute num with
value sg):
The verb calegaa ‘go’ places the same requirement on its subject:
(38) (c subj) = r
Thus, the two equations in (36) and (37) both require the f-structure labeled r,
which is also called (c subj), to contain the attribute num with value sg in the
index feature bundle. It is common for an attribute and its value to be multiply
constrained in this way; here, subject-verb agreement is enforced, since the num
requirements of both the subject and the verb are satisfied.
We will not discuss particular methods or algorithms for solving systems of
equations like the f-description in (34). Kaplan and Bresnan (1982) provide a
particularly clear presentation of one algorithm for solving such sets of equations,
and we briefly discuss some work on parsing and generation algorithms in §18.3.
There is a large body of work, summarized in detail by Rounds (1997), which
explores the logic of feature structures and their descriptions. Some work in LFG
and related frameworks discusses the operation of unification, first introduced in
linguistic work by Kay (1979) (see also Shieber 1986). Unification is an operation
(represented by () that combines compatible feature structures into a new feature
structure by taking the union of all the attribute-value pairs in the original
structures. For instance, the unification of the two feature structures f and g in
(39a) is the feature structure in (39b):
) * ) *
(39) a. a b c d
f g e [f g]
c d
a b
b. f ( g c d
e [f g]
This instantiated semantic form ‘David3 ’ is distinct from the semantic form
‘David7 ’ or ‘David25 ’ for different instances of the use of David.
Recall the discussion in §2.4 of f-structures in which two attributes have the
same value:
The verb seem requires its subj to be the same as its xcomp’s subj by means of an
equation like the following:
The value of the pred attribute of the subj ‘David’ is instantiated to a unique value
for this instance of its use. An equally correct but less succinct way of representing
the same f-structure is:
functional constraints 151
Crucially, the indices on the semantic forms of the subj and the xcomp subj
f-structures are the same. The f-structure in (44) is not the same as the one in (45),
in which two different semantic forms for David have distinct indices:
5 In our discussion of Warlpiri in Chapter 4, we discussed example (32) (page 134) in which two
separate phrases contributed to the subj function. In that example, unlike (46), the two phrases are
compatible and can both appear in the same utterance, since one phrase is interpreted as the head and
the other is interpreted as a modifier of the head.
152 describing syntactic structures
Bosnian/Croatian/Serbian also has full pronominal forms (here, the third singular
feminine pronoun nju), used for emphasis, which do not appear in the second-
position clitic cluster:
The f-structures for these examples are fundamentally similar; (50) gives an
abbreviated f-structure for both example (48) and example (49), with the obj
f-structure labeled f :
(50) ’
f
The lexical entries for the clitic and full pronoun contain the information in (51):
6 Following Franks and King, we translate the feminine full and clitic pronouns nju and ju as ‘it’
when they appear as the object of the verb čitao ‘read’.
functional constraints 153
Despite the fact that the pred value contributed by both pronominal forms is ‘pro’,
the sentence is ungrammatical; again, as above, a clash is produced by multiple
specification of semantic forms with different indices as the value of the obj pred:7
In general, following standard LFG practice, we will not display numeric indices on
semantic forms unless it is necessary for clarity, and two different semantic forms
are treated as distinct even if they look the same. If we want to indicate that the same
semantic form appears in two different places in the f-structure, as in example (42),
we will draw a line connecting the two occurrences.
In some cases, the value of an attribute other than pred might be required to be
uniquely contributed; for instance, the value of the tense attribute is contributed
by only a single form, and multiple contributions are disallowed:
An instantiated symbol can be used as the value of the tense attribute in such a
situation. Like a semantic form, an instantiated symbol takes on a unique value
on each occasion of its use. In general, any syntactic uniqueness requirement for
an attribute can be imposed by the use of an instantiated symbol as the value of
that attribute. Notationally, instantiated symbols are followed by an underscore;
for example, to indicate that the value for the attribute tense is the instantiated
symbol prs, we write:
5.2.2.2 Argument lists A semantic form, unlike other values, may contain an
argument list. In example (50) of this chapter, the pred value contributed by the
verb čitao ‘read’ is the complex semantic form ‘read%subj,obj&’. As discussed in
§2.4.6, this f-structure is complete and coherent because the requirements specified
by the semantic form for čitao ‘read’ are satisfied: the f-structure has a subj and an
obj, and there are no other governable grammatical functions in the f-structure
that are not mentioned in the argument list of čitao ‘read’. Additionally, since the
subj and obj arguments of čitao ‘read’ are semantic arguments appearing inside
the angled brackets of the argument list, each must contain a pred; nonsemantic
arguments mentioned outside the angled brackets in the argument list must also
be present in the f-structure, but are not required to contain a pred.
5.2.2.3 Reference to parts of the semantic form It is possible to use a special set
of attributes to refer to parts of a semantic form. For a simple semantic form like
‘David’, the attribute fn can be used to refer to the function component:
Numbered attributes such as arg1 and arg2 can be used to refer to the arguments
of the semantic form. For example, given the semantic form ‘read%subj,obj&’, the
following equations hold:
5.2.3 Disjunction
An f-description can also consist of a disjunction of two or more descriptions.
When this happens, one of the disjuncts must be satisfied for the f-description
to hold.
For instance, the form met of the English verb meet is either a past tense form
or a past participle:
This is reflected in the following disjunctive f-description in the lexical entry for
met, which says that the f-structure f for met must contain either the attribute-
value pair %tense, pst& or the attribute-value pair %vtype, pst.ptcp&:
(61) ’
f
b. ’
f
Each of these minimal solutions satisfies one of the disjuncts of the description.
Formally, a disjunction over descriptions is satisfied when one of the disjuncts
is satisfied:
(62) Disjunction:
A disjunction {d1 | . . . | dn } over f-descriptions d1 . . . dn holds of an
f-structure f if and only if there is some disjunct dk , 1 ≤ k ≤ n, that holds
of f .
156 describing syntactic structures
5.2.4 Optionality
An f-description can also be optional. When this happens, the f-description may
but need not be satisfied.
Bresnan and Mchombo (1987) show that verbs in Chicheŵa optionally carry
information about their subjects; in a Chicheŵa sentence, a subject noun phrase
may be either present (63a) or absent (63b):
Bresnan and Mchombo propose that the verb zi-ná-lúm-a ‘bit’ optionally con-
tributes an f-description constraining the value of the pred attribute of its subject.
This optional f-description is enclosed in parentheses:
Since the equation ( f subj pred) = ‘pro’ is optional, it may but need not contribute
to the minimal solution of the f-description for the sentence. If an overt subject
noun phrase does not contribute its own pred value, the f-structure for the sen-
tence is incomplete unless this equation is satisfied, and in this case the wellformed
f-structure for the subj contains the pair %pred,‘pro’&. If an overt subject noun
phrase appears, the equation cannot be satisfied, since the pred value of the overt
subject would produce a clash; instead, the pred value for the subj is the one
specified by the subject noun phrase:
b. zi-ná-lúm-a alenje
10subj-pst-bite-ind hunters
‘They bit the hunters.’
’
f
functional constraints 157
A similar analysis is appropriate for languages that allow clitic doubling. As dis-
cussed in §4.3, the River Plate and Peruvian dialects of Spanish allow either a clitic
or a full noun phrase object to appear:
Unlike many other dialects of Spanish, in these dialects the clitic pronoun can co-
occur with a full noun phrase object:
We account for these facts by assuming that in the River Plate and Peruvian
dialects, the pred value contributed by the clitic pronoun lo is optional:
When a full noun phrase object is present, the optional equation contributing the
pred value of the clitic pronoun is not satisfied; if two preds were present, the
example would not satisfy Consistency. When there is no full noun phrase, in order
to satisfy Completeness, the pred contributed by the clitic noun phrase appears.
The f-structure for example (66b) is given in (70):
(70) ’
f
(71) Optionality:
An f-description d optionally holds of an f-structure f if and only if the
disjunction {d | true} holds of f .
158 describing syntactic structures
5.2.5 Negation
An f-description can be negated; when this happens, the f-description must not be
satisfied. Notationally, negation of a single equation is indicated by a diagonal line
through the equals sign: f += g means that f does not equal g.
For example, as discussed in §2.5.3, Quirk et al. (1972: §15.6) claim that it is not
possible to use the complementizer if in the clausal complement of certain verbs,
including the verb justify:
We can analyze the verb justify as described by Quirk et al. differently from a
verb like know by prohibiting the value if for the attribute compform in its comp
argument:
However, the f-structure in (75) does not satisfy the constraint; the offending value
is circled:
This example shows that a single equation can be negated, requiring a particular
attribute-value pair not to appear. More generally, it is possible to negate not
just a single equation, but an entire f-description. The following expressions are
notationally equivalent:
(76) ( f a) += v ≡ ¬[( f a) = v]
(78) Negation:
A negated f-description ¬d holds of an f-structure f if and only if the
description d does not hold of f .
The f-structure for example (79a) is shown in (80). Note that the f-structure labeled
f contains the attribute tense with value pst:
160 describing syntactic structures
f ’
(81) ( f tense)
This constraint requires the f-structure f to contain the attribute tense, but it
does not constrain the value of the tense attribute; any value is acceptable. The
f-structure in (80) satisfies this constraint. However, an f-structure like the one in
(82) does not satisfy the constraint, since it has no tense attribute:
f ’
(85) ¬( f tense)
The constraint in (85) ensures that the f-structure f has no tense attribute. The
constraint is satisfied in (86):
We assume that the lexical entry for the complementizer that contributes the
following defining equation:
In (88c), the clausal argument that David yawned is the subj of the clause, and
contains the attribute-value pair %compform, that&:
The defining equation in (89) is satisfied, because the f-structure labeled f has
an attribute compform with value that. Hence, (90) is an acceptable minimal
solution to an f-description that includes the defining equation in (89).
In contrast, the pair %compform, that& is not required to belong to the
f-structure of a comp; example (88b) is grammatical, and its f-structure is
wellformed:
The equation in (89) is not a part of the f-description for example (88b), since the
word that does not appear; in the minimal solution to the f-description for this
sentence, f does not have an attribute compform with any value.
functional constraints 163
Our task, then, is to require a clausal subject such as the one in (90) to contain
the attribute compform with value that, and thus to ensure the presence of the
complementizer that. To impose this requirement, we can use the constraining
equation in (92):
(93) Introducing the semantic form ‘pro’ as the value of the obj’s pred:
( f obj pred) = ‘pro’
Kaplan and Bresnan (1982) provide an interesting discussion of the formal role of
constraining equations and existential constraints in LFG (see also Johnson 1999;
Saraswat 1999). As they note, constraining equations and existential constraints
are particularly useful when the presence of a feature is associated with a marked
value for the feature, with the absence of the feature indicating the unmarked value
(§2.5.2). For instance, all marked forms of a particular paradigm may be associated
with a certain feature: assume, for instance, that only passive verbs have a value pass
for the voice feature, and active verbs have no voice feature. In this situation, if a
particular voice of a verb is required, a constraining equation must be used to check
for the presence or absence of the voice feature; a defining equation mentioning
the voice feature would be compatible with passive verb forms as well as with
active forms unspecified for the voice feature, the wrong result.
In certain limited situations, the use of a defining equation does not produce a
different result from the use of a constraining equation. A constraining equation is
used to check whether an attribute and its value are present in the minimal solution
of an f-description. If we know that a particular attribute is always present in a
certain type of f-structure, we need not use a constraining equation to check its
value: in that situation, using a defining equation produces the same result. For
instance, suppose that all noun phrases in English are marked for number, so
that they are all either singular or plural. Suppose further that the value for the
number feature is not an instantiated symbol and can be specified by more than
one defining equation. Then, in specifying number agreement with a noun phrase,
it does not matter whether the specification involves a defining equation or a
constraining equation: we know that the minimal solution to the constraints always
contains a number attribute with some value, since all noun phrases are specified
for number. We can include an additional defining specification of the feature, or
we can use a constraining equation to ensure that the feature and its required value
are present in the minimal f-structure satisfying the defining equations.
(96) The
4 attribute case has the default value nom:
5
( f case) ( f case) = nom
( f case) += nom
8 Disjunction is discussed in §5.2.3, negation is discussed in §5.2.5, and existential constraints are
discussed in §5.2.6.
functional constraints 165
If this constraint is present, the f-structure f must have some value for the attribute
case. Some other component of the grammar may specify a non-default value
(a value other than nom) for case. If there is no other specification of a different
value for case elsewhere in the full f-description, then the value of the case
attribute for f is the default, nom.
Some works adopt a simpler treatment of defaults:
(97) The attribute case has the default value nom, alternative (simpler) specifi-
cation:
6 7
( f case) ( f case) = nom
• The existential constraint ( f case) holds: a value for case is specified by some
component of the grammar; or
• the value for case is the default value, nom.
This simpler statement has the same effect as the more complex one in (96), in
that it specifies a default value nom for the attribute case. However, it is less
constrained, in that it allows more than one equivalent way of satisfying a default
specification in some circumstances. If another component of the full f-description
specifies the default value nom for the case attribute, both alternatives in (97) are
satisfied: (i) there is a value for the case attribute, and (ii) it is the default value,
nom. This means that there may be more than one solution for an f-description
with this simpler form of default specification, giving rise to the appearance of
ambiguity where no ambiguity exists. Thus, we prefer the more complex statement
of defaults as in (96), since it does not permit both alternatives to be satisfied at the
same time.
The general form of default specifications is, then, as follows:
5.2.10 Implication
It is often convenient and intuitive to express a linguistic generalization as an
implication. For example, implicational constraints can be used to encode effects
of the person hierarchy in Lummi transitive sentences. Building on work by Jelinek
and Demers (1983), Bresnan et al. (2001) observe that Lummi requires the person
hierarchy in (99a) to be aligned with the grammatical function hierarchy, with the
result that a first or second person argument cannot be lower than a third person
argument on the grammatical function hierarchy:
This means that a third-person subject may not appear with a first or second person
object in an active Lummi sentence, and in a passive sentence, a third person
subject may not appear with a first or second person oblique argument.9 Jelinek
and Demers (1983) provide the data in (100–101) to illustrate these patterns.10 In
the active sentences in (100), a third person subject may not appear with a first or
second person object, as in (100c), but other combinations are allowed:
In the passive sentences in (101), a third person subject may not appear with a
first or second person oblique phrase, as in (101c), but other combinations are
allowed:
9 Constraints on voice and the person hierarchy in Lummi are more complicated than the data in
this section indicate; for a detailed discussion, see Jelinek and Demers (1983).
10 Jelinek and Demers do not provide Lummi equivalents for the ungrammatical patterns.
functional constraints 167
(102) In Lummi, if the subject is third person, the object or oblique phrase cannot
be first or second person.
An implication is written with the double right arrow “⇒.” For any clausal
f-structure f in Lummi, the following implication holds:11
(103) If f ’s subject is third person, its object or oblique is not first or second
person:
( f subj index pers) = 3 ⇒ ( f {obj|obl} index pers) += 1
( f {obj|obl} index pers) += 2
Since negation and disjunction are permitted in f-descriptions, we can define the
implication relation by using negation and disjunction. In the general case, the
implication in (104a) can be paraphrased as in (104b):12
11 For simplicity, we assume that the value of the pers feature is one of the atomic values 1, 2, or 3. The
complex person features discussed in §2.5.7.2 do not provide a means of referring to first and second
person as a natural class of feature values, to the exclusion of third person; further refinement to the
values proposed for the pers feature by Otoguro (2015) would allow us to simplify the rule in (103) by
referring to such a class.
12 Formally, this equivalence is motivated by the conditional law, which states that the implication
d1 → d2 holds if and only if either d1 fails to hold, or d2 holds:
[d1 → d2] ⇔ {¬d1 | d2}
Since either an f-description d1 or its negation ¬d1 is true of any f-structure, we can use the stronger
formulation of equality in (104).
168 describing syntactic structures
(105) Implication:
Given two f-descriptions d1 and d2, d1 ⇒ d2 (“d1 implies d2”) holds if
and only if {¬d1 | ¬¬d1, d2} (“either d1 does not hold, or it is not the case
that d1 does not hold, and d2 holds”).
(106) V
’
V
yawned
To make the following discussion simpler, we assume that the rule expanding V# is:
(107) V# −→ V
As discussed in §4.2.1, a phrase and its head correspond to the same f-structure.
Here, V is the head of V# , and so V# and V correspond to the same f-structure.
We would like a way to express this fact.
We accomplish this by annotating the V node with an expression requiring the
f-structure of the V to be the same as the f-structure for the V# . In general, any
daughter node—that is, any node on the right-hand side of a phrase structure
rule—may be annotated with constraints on the relation between its f-structure
and the f-structure of the mother node. If the daughter node is the head, the
f-structures must be the same. If the daughter node is a nonhead, its f-structure
bears some relation (say, the obj relation) in the mother’s f-structure.
In order to do this, we need a notation for the following concepts:
The symbol ⃰ stands for the node corresponding to the rule element on which the
constraint is written; note that this use of ⃰ is not related to the Kleene star notation
the c-structure/f-structure correspondence 169
(109) V# −→ V
𝜙(ˆ⃰) = 𝜙(⃰)
mother’s (V# ’s) f-structure = self ’s (V’s) f-structure
The intuition behind this notation comes from the way trees are usually repre-
sented: the up arrow ↑ points to the mother node, while ↓ points to the node
itself. Using these abbreviations, we can rewrite the rule in (109) in the following
more standard way:
(111) V# −→ V
↑=↓
mother’s f-structure = self ’s f-structure
(112) V# []
In some LFG literature, f-structure annotations are written above the node labels
of a constituent structure tree, making the intuition behind the ↑ and ↓ notation
clearer; written this way, the arrows point to the appropriate phrase structure
nodes:
(113) V#
↑=↓
V
170 describing syntactic structures
In the following, we will stick to the more common practice of writing functional
annotations beneath the node label in the phrase structure rule, as in (111).
Let us turn to a slightly more complicated rule, one that describes the following
c-structure and f-structure:
+ ,
(114) obj [ ]
V #
V NP
Here, the NP bears the obj function. Thus, the rule in (115) contains the additional
information that the f-structure for the NP daughter of V# is the value of the obj
attribute in the f-structure for the V# :
(115) V# −→ V NP
𝜙(ˆ⃰) = 𝜙(⃰) (𝜙(ˆ⃰) obj) = 𝜙(⃰)
mother’s f-structure’s obj = self ’s f-structure
(116) V# −→ V NP
↑=↓ (↑ obj) = ↓
Some LFG work follows another abbreviatory convention according to which the
annotation ↑ = ↓ is omitted when it is the only annotation on a node. According
to this convention, a rule like (116) can be written as:
(117) V# −→ V NP
(↑ obj) = ↓
In this book, we will try to be as explicit as possible, so we will not follow this
convention of omitting equations.
Restricting ourselves to rule annotations referring only to the f-structures corre-
sponding to a daughter node and the mother node in a phrase structure rule makes
a strong claim about the local applicability of syntactic constraints. For instance,
we cannot refer to the grandmother node in a tree, or to its f-structure. This means
that nonlocal syntactic relations are statable only in functional terms, not in terms
of constituent structure configuration. Within the tree, only local phrase structure
relations can be constrained by phrase structure rules.13
13 Some analyses in LFG involve reference to the f-structures of sister nodes in the phrase structure
rule: see §6.4.
the c-structure/f-structure correspondence 171
(118) V ’
yawned
The following lexical entry for yawned provides information about the f-structure
corresponding to V, the immediately dominating c-structure node:
This lexical entry asserts that the f-structure ↑ corresponding to the V node
immediately dominating yawned has an attribute pred whose value is the semantic
form ‘yawn%subj&’, and an attribute tense whose value is pst. The f-structure
displayed in (118) is the minimal solution to these constraints, the smallest
f-structure that satisfies the constraints.14
5.3.2.1 ↑ in lexical entries The use of ↑ in a lexical entry is exactly the same as
its use in a rule: ↑ refers to the c-structure node dominating the word itself. This
can be seen more easily if we recast the lexical entry in (119) in the equivalent
phrase structure rule format:
(120) V −→ yawned
(↑ pred) = ‘yawn%subj&’
(↑ tense) = pst
5.3.2.2 ↓ in lexical entries In most cases, lexical entries specify only information
about ↑, the f-structure of the immediately dominating c-structure node. Some
analyses also refer to properties of the f-structure ↓ corresponding to the word
itself and how the f-structure for the word relates to the f-structure for the
immediately dominating node. For example, Zaenen and Kaplan (1995) explore
an analysis of word order constraints in Dutch which assumes lexical entries such
as the partial entry in (121) for the verb wil ‘want’:
14 We rely here on a simplified view of the relation between elements of the string and nodes in the
c-structure. See §3.5 for a more complete view, and Chapter 11 for full details.
172 describing syntactic structures
In this lexical entry, ↑ refers to the f-structure of the c-structure node V, and ↓
refers to the f-structure of the terminal node, the word wil. The structure associated
with wil is the semantic form ‘want%subj,xcomp&’. According to this lexical entry,
then, the value of the pred attribute of wil is the structure associated with the
terminal node wil. The entire configuration is depicted in (122):
) *
(122) V pred ‘want%subj,xcomp&’
8 9: ;
wil
Bresnan et al. (2016) also use ↓ in lexical entries of words containing incorporated
pronouns to anchor these pronouns to particular positions in the tree (for a discus-
sion of incorporated pronouns, see §14.1). For instance, Bresnan et al. (2016: 160)
propose that the object marker (abbreviated as om) that appears with Chicheŵa
transitive verbs has the following lexical entry, where agr is an abbreviatory symbol
representing the attributes pers, num, and gend:
(124)
V
agr α
5.3.3 An example
We now have the notational equipment to do a complete analysis of a sentence
like David yawned. For clarity, in the following example the rules and lexical
entries have been considerably simplified; for example, the phrase structure rules
expanding VP and NP are clearly too impoverished to account for very many
constructions in English.
We assume the following annotated phrase structure rules:
(125) IP −→ NP I#
(↑ subj) = ↓ ↑=↓
I# −→ I VP
↑=↓ ↑=↓
the c-structure/f-structure correspondence 173
VP −→ V
↑=↓
NP −→ N
↑=↓
These rules and lexical entries admit the following tree, with as yet uninstantiated
variables ↑ and ↓ over f-structures:
NP I#
(↑ subj) = ↓ ↑=↓
N VP
↑=↓ ↑=↓
David V
(↑ pred) = ‘David’ ↑=↓
yawned
(↑ pred) = ‘yawn%subj&’
(↑ tense) = pst
The next task is to instantiate the ↑ and ↓ metavariables to the f-structures that
they stand for in this example. It will be useful to have names for the f-structures
corresponding to each node. Taking advantage of the coincidental fact that in this
tree, every node has a different label, we will give the name fv to the f-structure
corresponding to the node labeled V, fvp to the f-structure for the node labeled VP,
and so on.
We begin with the information contributed by the lexical entry for David. The
f-structure variable ↑ in the annotation (↑ pred) = ‘David’ for David refers to fn ,
174 describing syntactic structures
the f-structure of the N node immediately dominating the leaf node David, and so
we replace ↑ in that expression with fn :
(128)
IP
NP I#
(↑ subj) = ↓ ↑=↓
N VP
↑=↓ ↑=↓
David V
( fn pred) = ‘David’ ↑=↓
yawned
(↑ pred) = ‘yawn%subj&’
(↑ tense) = pst
Let us now consider the N node. Its annotation is ↑ = ↓, meaning that the f-
structure fnp corresponding to its mother node NP is the same as fn , the f-structure
for the N node:
(129)
IP
NP I#
(↑ subj) = ↓ ↑=↓
N VP
fnp = fn ↑=↓
David V
( fn pred) = ‘David’ ↑=↓
yawned
(↑ pred) = ‘yawn%subj&’
(↑ tense) = pst
In a similar way, we replace the ↑ and ↓ nodes in the rest of the tree with the names
of the f-structures they refer to:
the c-structure/f-structure correspondence 175
(130)
IP
NP I#
( fip subj) = fnp fip = fi#
N VP
fnp = fn fi# = fvp
David V
( fn pred) = ‘David’ fvp = fv
yawned
( fv pred) = ‘yawn%subj&’
( fv tense) = pst
According to these constraints, fip , fi# , fvp , and fv all name the same f-structure,
which has three attributes, pred, tense, and subj. The subj of this f-structure is fnp ,
which is also called fn . The f-structure for this sentence is the minimal solution to
these constraints, the f-structure that satisfies all of these constraints and contains
no extra structure not mentioned in the constraints:
(132)
’
fip , fi , fvp , fv
fnp , fn
We have now deduced that the sentence David yawned has the following anno-
tated c-structure and f-structure:
176 describing syntactic structures
(133)
’
IP
NP I
( fip fnp fip = fi
N VP
fnp = fn fi = fvp
David V
( fn fvp = fv
yawned
( fv ’
( fv
5.4.1 Balinese
In Balinese, as described by Arka (2003), the grammatical functions subj, obj, and
obj𝜃 are primarily encoded configurationally. This means that phrase structure
rules contain specifications of particular grammatical functions: just as in English
(§4.2), the specifier position of IP is filled by the subject, and the object appears
as the first nominal complement of V.15 These principles of mapping between
c-structure and f-structure configurations are reflected in the annotations on the
rules given in (135): heads of phrases bear the annotation ↑ = ↓, ensuring that a
phrase and its head correspond to the same f-structure; the specifier of the func-
tional category IP bears the annotation (↑ subj) = ↓, ensuring that it is associated
with the grammatical function subj; the S complement of the functional category I
is an f-structure co-head, bearing the annotation ↑ = ↓; and the complement of the
lexical category V bears the annotation (↑ obj) = ↓, ensuring that it is associated
with the grammatical function obj:
S −→ { VP | NP }⃰
↑=↓ (↑ ggf) = ↓
< # =
VP −→ V
↑=↓
< = < =
V# −→ V NP
↑=↓ (↑ obj) = ↓
Predicates specify a list of the governable grammatical functions that they require:
The c-structure and f-structure for the sentence sampi ngamah padang ‘a cow eats
grass’ are given in (137), with the relation between the clausal head c-structure
nodes and the main f-structure indicated by arrows:
NP I
(↑ ↓ ↑ =↓
N S
↑ =↓ ↑ =↓
’
sampi VP
(↑ ↑ =↓
V
↑ =↓
V NP
↑ =↓ (↑ ↓
ngamah N
(↑ ’ ↑ =↓
padang
(↑
variation in grammatical function encoding 179
The requirement for the presence of a subj and an obj is lexically specified by the
verb, and the grammatical function of each argument is determined by its phrase
structure position.
C# −→ C S
↑=↓ ↑=↓
S −→ XP XP V XP∗
(↑ ggf = ↓ (↑ ggf = ↓ (↑ = ↓) (↑ ggf = ↓
In Ancient Greek, the specifier position of CP is associated with a phrase that fills
the dis function at f-structure. This is indicated by the annotation (↑ dis) = ↓;
unbounded dependencies are discussed further in Chapter 17. The dis function
is an overlay function and must be syntactically or semantically linked to a
primary non-overlay function: in this CP rule, the annotation (↑ ggf) = ↓ requires
this phrase to bear not only the dis role, but also a grammatical function ggf
in the sentence.17 The XP daughters of S are also annotated with the equation
(↑ ggf) = ↓, indicating that a phrase with any grammatical function can appear
there. Complementizers and some conjunctions can be analyzed as appearing in C.
In contrast to Balinese, the Ancient Greek verb specifies a great deal of informa-
tion about its arguments. The case of each argument is specified, and additionally
an optional pred value for each argument is provided. As described in §5.2.4, this
allows for the absence of overt phrasal arguments (pro-drop): an Ancient Greek
sentence may consist simply of a verb, with no overt subject or object phrases
present at c-structure. In such a case, the pred values of the arguments of the verb
come from the verb’s specifications:
16 The analysis of Ancient Greek presented here relies on the work of Dik (1995, 2007) on the relation
between word order and information structure in Ancient Greek, and on the work of Haug (2008a,b,
2011, 2012) on Ancient Greek syntax. The rules provided are, as with Balinese, necessarily simplified
for the purposes of exposition.
17 This is a simplification, since phrases from embedded clauses can appear in the specifier of CP. The
annotation (↑ ggf) = ↓ on the specifier of CP can be read as an abbreviated equivalent (in a monoclausal
sentence) for a more complex annotation involving specification of a path through the f-structure, as
we will see in Chapter 17.
180 describing syntactic structures
NP
C
(↑ ↓
↑=↓
(↑ ↓
N C S
↑=↓ ↑=↓ ↑=↓
strouthòn
dè
NP V
but
(↑ (↑ ↓ ↑=↓
(↑
(↑
é-laben
(↑ ’
N
((↑
↑=↓
(↑
((↑
oudeìs
(↑
(↑
(↑
The verb requires its subject to be in the nominative case; phrase structural annota-
tions allow the phrase oudeìs ‘no one’ to bear any governable grammatical function
ggf, but only the subj grammatical function is compatible with the nominative
casemarking requirements imposed by the verb and captured by the equation
(↑ concord case) = nom for oudeìs. Similarly, its accusative casemarking requires
the phrase strouthòn ‘ostrich’ to bear the obj function.
The verb also provides optional ‘pro’ values, enclosed in parentheses, for the
pred of its subject and object. These values do not appear in the final f-structure,
since the overt subject and object noun phrases oudeìs ‘no one’ and strouthòn
‘ostrich’ are present and contribute their pred values to the final f-structure. If
these phrases did not appear, the ‘pro’ value optionally provided by the verb would
appear as the value of the pred of these arguments.
variation in grammatical function encoding 181
5.4.3 Chicheŵa
Chicheŵa is typologically different from both Balinese and Ancient Greek, and
illustrates an interesting combination of configurational and nonconfigurational
characteristics (Bresnan and Mchombo 1987; Bresnan et al. 2016). The relevant
phrase structure rules for Chicheŵa are:
(141) S −→ NP , NP , VP
(↑ subj = ↓ ↓ ∈ (↑ dis ↑=↓
VP −→ V#
↑=↓
V# −→ V NP
↑=↓ (↑ obj = ↓
These rules show that grammatical functions in Chicheŵa are specified configura-
tionally to some extent, though not in the same way as in Balinese. Like Balinese
(and Ancient Greek), Chicheŵa makes use of the exocentric category S; it differs
from Balinese, however, in that the daughters of S are the subject, a dis phrase
(which is the topic at the separate level of information structure; see Chapter 10),
and the VP, which may appear in any order (the first rule in (141) is an unordered
ID rule, as described in §5.1.5). There is a configurationally specified postverbal
object position: the complement of the lexical category V is obj, as indicated by
the equation (↑ obj) = ↓.
The lexical entry for the Chicheŵa transitive verb zi-ná-wá-lum-a ‘bit’ is given
in (142):
Unlike the Balinese verb, and like the Ancient Greek verb, this verb contains an
optional ‘pro’ value for the pred of its subject; this means that an overt subj phrase
may but need not appear. The verb also carries information about the noun class
of its arguments: Chicheŵa, like many Bantu languages, has a complex noun class
system, and the prefix zi- indicates that the subj belongs to noun class 10.
The obj is treated differently from the subj. As Bresnan and Mchombo (1987)
show, this verb contains an incorporated pronominal object wá. This means that
the equation specifying the obj pred is obligatory, not optional. The c-structure
and f-structure for the sentence njûchi zi-ná-wá-lum-a ‘The bees bit them’ are
displayed in (143):
182 describing syntactic structures
N
V
↑ =↓
↑ =↓
njûchi
bees
(↑ V
(↑ ↑ =↓
zi-ná-wá-lum-a
(↑ ’
((↑
(↑
(↑
(↑
When the incorporated object pronoun wá does not appear, the sentence is
incomplete unless an overt noun phrase is present. The lexical entry for the verb
zi-ná-lum-a ‘bit’, with no incorporated obj pronoun, is:
In the following Chicheŵa sentence, there is no overt subj noun phrase. This
means that the subject noun phrase is an incorporated pronoun, as optionally
specified by the verb. An overt obj noun phrase, alenje ‘hunters’, also appears;
if there were no overt obj noun phrase, the sentence would be incomplete and
therefore ungrammatical. The grammatical function of alenje is determined by
the phrase structure configuration in which it appears:
variation in grammatical function encoding 183
V
↑=↓
V NP
↑=↓ (↑ ↓
zi-ná-lum-a
N
(↑ ’
↑=↓
((↑
(↑
alenje
hunters
(↑
(↑
NP NP VP
(↑ ↓ ↓ ∈ (↑ ↑=↓
N N V
↑=↓ ↑=↓ ↑=↓
zi-ná-wá-lum-a
njûchi alenje
bees hunters (↑ ’
(↑ (↑ ((↑
(↑ (↑ (↑
(↑
(↑
5.4.4 Bulgarian
Bulgarian is unusual in combining relatively free word order with a lack of nominal
inflection (Rudin 1985): only pronominal forms show casemarking. In some cases,
the subject can be identified as the argument that agrees with the verb; additionally,
Bulgarian allows clitic doubling, so that the case, gender, and number of clitic-
variation in grammatical function encoding 185
doubled arguments are specified. In other cases, however, these clues do not
serve to disambiguate the grammatical functions of the argument phrases, and a
phrase may be associated with any of the grammatical functions selected by the
predicate. In such cases, only contextual information and world knowledge help in
determining the intended structure.
The relevant phrase structure rules for Bulgarian are:
(147) IP −→ NP I#
↓ ∈ (↑ dis) ↑=↓
(↑ ggf) = ↓
I# −→ I S
↑=↓ ↑=↓
I −→ N I
(↑ obj) = ↓ ↑=↓
∗
S −→ { NP | V }
(↑ ggf) = ↓ ↑=↓
At this rough level of detail, these rules are very similar to the rules for languages
such as Ancient Greek, which also allow relatively free word order. An important
difference is that the I rule allows adjunction of a nonprojecting DN object clitic to I.
The third person singular past tense verb vidja ‘saw’ has this lexical entry:
As Rudin (1985) shows, a Bulgarian clause can appear without an overt subject, as
in Ancient Greek and Chicheŵa: the verb contains an optional equation specifying
a pronominal value for the pred of its subject. If the Bulgarian verb is transitive,
either an overt object phrase or the object clitic pronoun must appear, since the
verb does not specify a pred value for its object.
Consider (149). The noun knigata ‘the book’ and the proper noun Georgi (a male
name) are unmarked for case; each of them is compatible with either nominative
or accusative case. The symbol ggf, which represents any governable grammatical
function, is arbitrarily instantiated to subj for Georgi and obj for knigata ‘the book’.
As Rudin (1985) notes, it is only world knowledge that enforces the interpretation
of knigata ‘the book’ as the object of the verb vidja ‘saw’, and Georgi as its subject.
Neither phrase structure position, casemarking, nor agreement requirements serve
to disambiguate the syntactic role of these arguments.18 However, in (149), the
18 But see King (1995) and Mahowald (2011) for discussion of ways in which word order constraints
may contribute to disambiguation in similar constructions.
186 describing syntactic structures
IP
NP
I
↓ ∈ (↑
↑ =↓
(↑ ↓
N I S
↑ =↓ ↑ =↓ ↑ =↓
vidja
Knigata
book (↑ ’
(↑ ((↑ NP
(↑ (↑ (↑ ↓
(↑ (↑
(↑ (↑ N
(↑ ↑ =↓
Georgi
Georgi
(↑
(↑
(↑
(↑
Example (150) differs from (149) in that the dis noun phrase decata ‘the children’
is plural. Thus, since the verb shows third person singular agreement with its
variation in grammatical function encoding 187
subj, the only available syntactic analysis is the one in which the third person
singular phrase Georgi is the subject. In this example, decata ‘the children’ appears
in the specifier of IP and can be interpreted as bearing the prominent information
structure role of focus:
IP
NP
↓ ∈ (↑ I
(↑ ↓
N I S
↑=↓ ↑=↓ ↑=↓
vidja
Decata
(↑ ’
children
((↑ NP
(↑
(↑ (↑ ↓
(↑
(↑
(↑ N
(↑
(↑ ↑=↓
Georgi
Georgi
(↑
(↑
(↑
(↑
The presence of a doubled clitic object pronoun can also help to disambiguate a
potentially ambiguous sentence; example (151) is unambiguous. Here, the subject
Georgi appears in the specifier of IP, and is likely to be interpreted as a topic:
188 describing syntactic structures
IP
NP
I
↓ ∈ (↑
↑=↓
(↑ ↓
N I S
↑=↓ ↑=↓ ↑=↓
Georgi
Georgi
(↑ N I NP
(↑ (↑ ↓ ↑=↓ (↑ ↓
(↑
(↑ gleda
ja
(↑ ’
((↑
((↑ N
(↑
(↑ ↑=↓
(↑
(↑
(↑ Marija
(↑
(↑ Marija
(↑
(↑
(↑
(↑
(↑
The lexical entry for the feminine singular accusative clitic pronoun ja is:
variation in grammatical function encoding 189
If there is no full object noun phrase, the pred value for the obj function is given
by the object clitic pronoun phrase. Since the pred of the clitic pronoun is optional
(as in Spanish; see §5.2.4) the presence of ja is also compatible with the obj being
filled by the feminine phrase Marija (but not the masculine phrase Georgi, since
that would produce a clash in gend values).
Contrast this with comparable examples from Maori (154a), Japanese (154b), and
Russian (154c), in which no verb appears in the sentence. The following examples
are from Rosén (1996) citing Biggs (1969: 24), Dalrymple et al. (2004a: 190), and
Attia (2008) citing Avgustinova and Uszkoreit (2003), respectively:
Which of these two types of copular construction is used to encode the relevant
relationship may vary according to a number of different factors including tense
(as in the Hebrew examples), aspect, polarity, syntactic category, main versus
subordinate clause status, and the precise nature of the semantic relation between
the elements in the sentence. Crosslinguistically, whether or not a verbless con-
struction is available does not invariably correlate with any particular one of these
factors. In other languages, the copula may be expressed by use of a particle which
does not exhibit all of the usual properties of a verb, or by inflectional marking on
a non-verbal element in the sentence similar to that more usually associated with
a verb: for details, see Curnow (1999) and Pustet (2003). Nordlinger and Sadler
(2003, 2007) propose that the clausal pred information which in other cases would
be contributed by a verb is lexically associated with these other expressions.
In terms of c-structure (and, by extension, the correspondence between c-
structure and f-structure), there is a fundamental difference between copular
constructions with a copular particle or verb and those without. Compare the c-
structures for the copular constructions in English (156) and in Russian (157). In
English, the copular verb appears in I: it precedes negation (He is not a student),
and it inverts in question formation (Is he a student?).19 There is no copular verb
in (157); IP is headless:
19 An alternative to the English c-structure in (156) is the c-structure in (a), which includes a VP
node. The choice between the trees in (156) and (a) depends on whether a single-tier analysis is assumed
(as shown in 161), for which the tree in (156) may be preferred in order to maintain the generalization
that the complement of I is always an f-structure co-head, or a double-tier analysis (as shown in 162),
for which the tree in (a) is appropriate.
(a)
IP
NP I#
I VP
He
is V#
NP
a student
variation in grammatical function encoding 191
(156) He is a student.
IP
NP I#
I NP
He
is
a student
(157) On student.
he student
‘He is a student.’
IP
NP I#
NP
On
he student
student
While we have seen that it is possible to have f-structures that are not related
to any c-structure node, as in cases of pro-drop (example (3), §4.1, page 118),
the examples that have been presented in this chapter and earlier chapters have
all included a verb which contributes the main pred of the clause. In a copular
construction including a verb such as (156), the verb is often assumed to introduce
the main clausal pred. In the case of verbless constructions such as (157), though,
it is not immediately clear what contributes the main pred that is required for the
f-structure to be coherent. In turn, this raises questions relating to the analysis of
copular constructions that do include a verb, and the broader issue of whether
a unified analysis of copular constructions is possible and desirable. We will see
that different analyses appear in the literature: for instance, it has been proposed
that the copular verb in a sentence such as (156) contributes only tense and
aspect information, with the main pred contributed by the copular complement.
Similarly, in the case of verbless constructions such as (157), different analyses are
possible as to what contributes the main pred at f-structure. A related issue, which
has also been the subject of debate in the literature, concerns which grammatical
functions the main pred subcategorizes for. Before considering the grammatical
functions that may be involved, we review two influential analyses of verbless
copular constructions. The first proposes that the main clausal pred is contributed
by the phrase structure configuration (§5.4.5.1); the second proposes that the main
clausal pred is contributed by the complement (§5.4.5.2).
192 describing syntactic structures
5.4.5.1 Main pred contributed by phrase structure Rosén (1996) proposes that
in the case of a verbless copular construction like the Maori example in (154a), the
phrase structure configuration rather than the lexical specifications of the words
in the construction must introduce the copular relation. Under a constructional
analysis such as this one, the pred value that licenses the grammatical functions
of non-verbal elements is introduced by the phrase structure of the copular
construction:20
D N D N
he te
N N
taariana hoiho
stallion horse
As shown in (158), Rosén (1996) associates the clausal pred with the first NP node
in the c-structure of Maori, which is interpreted as the predicative complement.21
This is one possible analysis of this construction.
20 Other analyses of verbless copular constructions have been formulated in terms of an empty node
rule (§6.10.4). On Russian, see Dalrymple et al. (2004a); on Arabic, see Attia (2008).
21 Rosén’s analysis makes use of the grammatical role ncomp, or nominal complement. This role
is no longer in general use in LFG analyses, since it violates modularity by encoding c-structure
distinctions (the phrase structure category of the complement) in the statement of f-structural subcat-
egorization requirements. The current equivalent would be predlink, discussed in §2.1.8 and below,
though the predlink role is not associated with any particular phrase structure category.
variation in grammatical function encoding 193
5.4.5.2 Main pred contributed by the complement When a verb is not present,
it can also be argued that it is the predicative complement that contributes the
main clausal pred, selecting for arguments just as a verb would and resulting
in a fundamentally different f-structure from that shown in (158). For example,
according to Dalrymple et al.’s (2004a) analysis of Japanese sentences such as
(154b), the lexical entry for akai ‘red’ includes (↑ pred) = ‘red%subj&’:
(159) C-structure and f-structure for hon wa akai according to Dalrymple et al.
(2004a):
hon wa akai
book topic red
‘The book is red.’
’
S
NP AP
(↑ ↓ ↑ =↓
A
hon wa
book akai
red
(↑ ’
subject, as in a sentence like A doctor came into the room, where a doctor is not
used predicatively:
(161) Single-tier analysis of Ona vrač ‘She is a doctor’:
’
22 A third kind of double-tier analysis with an oblique copular complement is proposed by Falk
(2004) for locative complement constructions in English and Hebrew, and by Laczkó (2012) for locative
and existential complement constructions in Hungarian (see also Bresnan et al. 2016: Chapter 12):
(a) ’
For other copular constructions in Hebrew, Falk argues for a predlink analysis; Laczkó (2012) argues
that a single-tier analysis is appropriate for Hungarian copular constructions involving attribution or
classification, while for constructions involving identity or possession, a double-tier predlink analysis
is motivated.
variation in grammatical function encoding 195
b. ’
Nordlinger and Sadler (2007: 142) point out that although most treatments of
copular constructions that include a verb assume a double-tier analysis (with
the copular verb contributing the main clausal pred), a single-tier analysis is
also possible. Under such an approach, the copular verb contributes information
about tense, aspect, mood, and other verbal features, while the non-verbal element
contributes the main clausal pred, meaning that the copular verb and the non-
verbal predicate are functional co-heads of the clause.
In summary, two main approaches to the analysis of copular constructions have
been put forward: the single-tier analysis, in which a non-verbal element supplies
the main clausal pred, and the double-tier analysis, in which the main clausal pred
is contributed by some other means (for example, by a copular verb or by the
phrase structure). Double-tier analyses can be subdivided according to whether
the complement’s grammatical function is taken to be open or closed.
regarded as the default for languages with only verbless copular constructions,
though they do not argue for a unified analysis of all copular constructions.
Other researchers have advocated an approach that allows more variation. Dal-
rymple et al. (2004a) observe that when the complement in a copular construction
has a subject that is not the same as the one in the matrix clause, a single-
tier analysis is not possible. This is because the subj of the copular complement
(David in 163) would clash with the subj of the copula (the problem in 163),
resulting in a violation of Consistency. If the main clausal pred selects for a
closed complement, however, the f-structure includes two distinct subjects and
Consistency is respected, as shown in (163):
(163) IP
NP I
I VP
The problem ’
is V
CP +
C ’
C IP
that NP I
VP
David
V
yawned
Compare (163) with (164), in which the (expletive) subject of the copula and its
complement clause are the same. Dalrymple et al. (2004a) argue that in such cases
an open complement xcomp analysis is required to account for the fact that a
standard control relation appears to exist between the two subjects (see Chapter 15
for a full discussion of open complements and control):
(164)
IP
NP I
I VP
It
is V
AP
A
A VP
likely
to rain
further reading and related issues 197
On the basis of examples such as (163) and (164), Dalrymple et al. (2004a) argue
that a single analysis cannot apply to all copular constructions: whether a closed
or open complement is involved must be determined on a case-by-case basis—
both crosslinguistically and within a single language—taking into account relevant
syntactic properties, including the presence or absence of a copular verb and
the presence or absence of agreement marking. On this view, the suitability of
an analysis is determined independently for each copular construction, and the
f-structures of all copular constructions need not be fundamentally the same, even
within a single language. This approach to the analysis of copular constructions
in LFG is also advocated by Falk (2004), Nordlinger and Sadler (2007), Laczkó
(2012), and Lowe (2013).
Precisely how c-structure and f-structure are related in the case of copular
constructions is the subject of ongoing debate in the LFG literature. Further
research which considers data from a range of languages is required to determine
if a unified analysis of some or all copular constructions is justified, either within
a particular language or crosslinguistically.
A class of languages with particularly simple formal properties is the class of regular
languages, which are languages that can be characterized by regular expressions. In
an LFG setting, regular expressions play an important formal role in the expression
of constraints on both the c-structure and the f-structure. In §5.1.1, we saw that
the right-hand side of an LFG phrase structure rule is a regular expression over
c-structure categories, allowing disjunction, optionality, and repetition:
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
200 syntactic relations and syntactic constraints
b. The disjunction of two regular expressions R1 and R2, written {R1 | R2},
is a regular expression. The regular language denoted by the disjunction
{R1 | R2} is the union of the language R1 and the language R2: that is, all
of the strings in R1 as well as all of the strings in R2.
c. Parentheses indicate optionality; if R is a regular expression, (R) denotes
the union of the strings in R with the empty string, which could also be
written as {R | 𝜖}.
d. The concatenation of two regular expressions R1 and R2, written R1 R2,
is a regular expression. The regular language denoted by R1 R2 is the con-
catenation of the strings in R1 with the strings in R2. (The concatenation
of a with ab is aab, and the concatenation of aa with bb is aabb.)
e. A regular expression annotated with the Kleene star operator ⃰ is a regular
expression. The regular expression R⃰ denotes the set of strings with 0 or
more repetitions of the strings in R.
f. A regular expression annotated with the Kleene plus operator + is a regular
expression. The regular expression R+ denotes the set of strings with 1 or
more repetitions of the strings in R.
The use of the Kleene star operator allows the description of an infinite language,
since an expression such as R⃰ denotes a language in which the strings in R are
repeated any number of times (including zero). The language characterized by the
regular expression in (5a) consists of strings which begin with a, then contain any
number of sequences of bc (including none), and end with d:
1 For definitions of union, intersection, and complementation, see (46) in this chapter (page 214). It
is not currently known whether arbitrary regular languages over annotated phrase structure categories,
regular expressions and regular languages 201
Care must be taken when defining the term complement of an annotated phrase
structure category: the term complement of an annotated category C with anno-
tation A denotes the set of all category/f-structure pairs that do not match the
annotated category. This is the disjunction of \C (that is, any category other than
C, with any annotation) and the category C with the annotation ¬A. For example,
the term complement of the category NP with annotation (↑ subj) = ↓ is written
as in (7a), and is the same as the disjunction in (7b):
! "
NP
(7) a. \
(↑ subj) = ↓
b. {\NP | NP }
(↑ subj) &= ↓
associated with f-structure constraints such as ↑ = ↓ or (↑ subj) = ↓, are closed under complementation
(Ron Kaplan, p.c.).
2 Given an alphabet Σ, the universe (the set of all possible strings in the language) is Σ ⃰, and the
complement of a set of strings S is Σ ⃰ − S (all strings in the universe except for those in S).
202 syntactic relations and syntactic constraints
two nodes, one labeled V and one labeled NP, but without specifying an order
between V and NP. The LP (linear precedence) constraint in (8b) specifies that
the V node precedes the NP node. This is a particular instance of the general rule
in (8c), stating that a lexical (X) node must precede a phrasal (YP) node:
(8) a. V! −→ V, NP
b. V < NP
c. X < YP
This is an example of the use of a formal device, the ID/LP rule format, to express
a generalization about word order across classes of phrase structure rules: for
example, whether the phrases in a language are head-initial or head-final, or
whether the specifier of a phrase precedes or follows the head.
More generally, it is possible to write rules that combine different constraints on
phrasal structure, allowing for the succinct expression of linguistic generalizations
about the structure of phrases. Such generalizations can be expressed in terms
of separate constraints that must be simultaneously satisfied. Formally, this is
possible because we are allowed to specify any regular language as the right-hand
side of a phrase structure rule; this means that concatenation, disjunction, Kleene
star, Kleene plus, intersection, and term complement can be used in specifying
sequences of categories.
We have already seen that disjunctions over various possibilities for phrase
structure expansion can be specified:
(9) X −→ { Y1 Y2 Y3 | Z1 Z2 Z3 }
This schematic rule indicates that a phrase with category X dominates either the
series of daughters Y1 Y2 Y3 or the series Z1 Z2 Z3. As we have seen, disjunction in
a phrase structure rule corresponds to the union of two regular languages, since
the union of two regular languages encompasses all of the alternatives in each
language.
Intersection of two regular languages corresponds to the combination of two
descriptions; each description must hold of the result:
(10) X −→ X1 & X2
This schematic rule indicates that X dominates a sequence of categories that must
satisfy the description represented by X1 as well as the description represented by
X2. Formally, this corresponds to characterizing a regular language by intersecting
two regular languages, represented by X1 and X2.
To take a concrete example, we can restate ID/LP rules in this way. Consider the
following simple rule:
(11) C! −→ C, IP C < IP
regular expressions and regular languages 203
(12) C! −→ {C IP | IP C}
This expression allows any sequence of categories, including C, IP, and other
categories, but with constraints on the relative order of C and IP. The expression
takes advantage of the term complement operator: for example, \IP is any category
except IP. Each of the members of the set of strings in the language of this
expression consists of three parts. First, there is an optional sequence beginning
with any number of categories not including IP and ending in C; next, there is
a sequence of categories that does not contain either IP or C; finally, there is an
optional sequence of categories that begins with IP and continues with a sequence
not including C. In other words, this expression disallows IPs before the initial C,
and also disallows Cs after the final IP.
Now that we have reformulated the right-hand side of the ID rule and the LP
constraint as regular expressions, we can state the combination of the two by
intersection. The ID/LP rule in (14a) is equivalent to the expression in (14b), where
the set of strings characterized by the regular expression on the right-hand side
of the rule in (11) is intersected with the set of strings characterized by the LP
constraint in (13):
(14) a. C! −→ C, IP C < IP
b. C! −→ {C IP | IP C} & ([\IP]⃰ C) \[C | IP]⃰ (IP [\C]⃰)
This example illustrates two points. First, we can state constraints on constituent
structure configurations by defining a sequence of daughter nodes in a phrase
structure rule in terms of a regular expression. This is a powerful and general tool
for the linguist to use in the description of phrase structure.
Second, this way of stating generalizations may not be the clearest or most
revealing. Although the encoding of linguistic facts in terms of operators such as
Kleene star, union, intersection, and term complement leads to a solid understand-
ing of their underlying formal and computational properties, it may be preferable
to devise a special notation for some common operations, since (as can be seen
above, from the restatement of ID/LP rules in terms of the intersection of two
regular languages) it may not be perspicuous or revealing to state these in terms of
the standard notation of regular expressions: the ID/LP notation in (14a) is simpler
204 syntactic relations and syntactic constraints
and easier to understand than the expression in (14b), where a complex regular
expression is defined.
Kaplan and Maxwell (1996) define and discuss some operators that allow for the
compact expression of particular kinds of regular languages:
The “Ignore” operator: Using the Ignore operator, written as a forward slash
/ (Kaplan and Kay 1994), we can write a rule that allows for a category or a sequence
of categories to be interspersed with the other categories:
This rule can be read as: “XP dominates X1 X2 X3, ignoring occurrences of Cat.”
In other words, XP must dominate X1 and X2 and X3 in that order, and may also
dominate any number of occurrences of Cat at any position. This rule is equivalent
to the following rule, containing a more complicated regular expression:
The Ignore operator rule can be used to describe the appearance of parenthetical
elements, elements that can be inserted before or after any phrasal constituent
(see McCawley 1982).
(18) a. X1 X2 X3 Y1 Y2 Y3
b. X1 X2 Y1 X3 Y2 Y3
c. X1 Y1 X2 Y2 X3 Y3
d. Y1 X1 X2 Y2 X3 Y3
Any ordering between the Xs and the Ys is allowed, as long as the order X1 X2 X3
and the order Y1 Y2 Y3 are preserved. The regular expression corresponding to this
rule is quite complex and will not be displayed.
The Shuffle operator is used to characterize constraints involving partial orders
holding among constituents in languages with otherwise fairly free word order. It
has been proposed within the Head-Driven Phrase Structure Grammar framework
regular expressions and regular languages 205
by Reape (1994) in his analysis of word order in German; however, the phe-
nomena Reape analyzes using Shuffle are best analyzed within LFG in terms of
functional syntactic relations, as shown by Kaplan and Zaenen (1989b).
Other operators based on regular predicates can be proposed if their use simpli-
fies linguistic description. The use of these operators assists the linguist in develop-
ing firm intuitions about linguistic phenomena; as long as the operators are defin-
able in terms of regular predicates, no new formal power is added to the theory.
IP
NP I
I S
Evgenija Onegina
Eugene Onegin napisal NP
wrote
N
Puškin
Pushkin
This definition uses the set membership symbol ∈ as an attribute to allow refer-
ence to some member of a set, as discussed in §6.3.1.
The abbreviatory symbol gf appears in the phrase structure rule for IP in
Russian (following King 1995: 204):
(21) IP −→ XP I!
(↑ gf) = ↓ ↑=↓
CP
NP C
’
N C IP
What do NP I
’
N VP
you V IP
think NP I
N VP
Chris V
bought
regular expressions and regular languages 207
CP ’
NP C
N C IP ’
What do NP I
N VP
you V
V IP
think NP I
N VP
Chris V
V IP
hoped NP I
N VP
David V
bought
A simplified version of the annotated phrase structure rule for English constituent
questions is:
(24) CP −→ XP C!
↓ ∈ (↑ dis) ↑=↓
∗
(↑ comp gf) = ↓
The first annotation on the XP daughter of CP, ↓ ∈ (↑ dis), requires the f-structure
corresponding to the XP to be a member of the set value of the dis attribute of
the CP mother node. The second annotation on the XP node, (↑ comp⃰ gf) = ↓,
contains a new sort of expression. As discussed in §6.1, the Kleene star operator ⃰
indicates that an expression may be repeated zero or more times. In particular,
comp⃰ represents paths containing any number of comps: the empty path, comp,
comp comp, and so on. Thus, the equation (↑ comp⃰ gf) = ↓ indicates that within
208 syntactic relations and syntactic constraints
the f-structure ↑, the dis f-structure also bears some grammatical function gf
which lies at the end of some path in the set of paths comp⃰: that is, some gf that
may be embedded inside any number of comps. The constraint holds if there is
some path in the set of paths comp⃰ gf for which the equation is true. In (22), the
path is comp obj. In (23), the path is comp comp obj. In some other example, a
different path might be chosen.
More complicated paths can also be characterized. A slightly more complete
version of the rule for question formation in English is:
(25) CP −→ XP C!
↓ ∈ (↑ dis) ↑=↓
∗
(↑ xcomp|comp gf) = ↓
The regular expression {xcomp | comp}⃰ denotes paths containing any number
of xcomps or comps in any order: for example, comp xcomp, comp comp, or
xcomp comp xcomp.4
Equations of this sort, involving abbreviatory symbols over grammatical
functions or more complex regular expressions denoting paths through an f-
structure, exemplify functional uncertainty. Functional uncertainty was first
introduced by Kaplan et al. (1987) and Kaplan and Zaenen (1989b) in the
treatment of long-distance dependencies such as topicalization, question formation,
and relative clause formation in English. The expression in (25) more adequately
captures constraints on question formation in English than the one in (24), but still
does not completely characterize the possible grammatical functions which can
be borne by the sentence-initial dis constituent in English questions; a detailed
discussion of the syntax and semantics of long-distance dependencies is provided
in Chapter 17.
Definition (30) in Chapter 5, repeated in (26), tells us how to interpret con-
straints involving a string as of length greater than one:
For instance, the f-structure ( f comp subj) is the f-structure at the end of the
path comp subj within the f-structure f ; it is the same as (( f comp) subj),
the subj of the f-structure ( f comp). In (26), the string s is the suffix of the
string as.
Building on this definition, Kaplan and Zaenen (1989b) provide the following
interpretation for constraints involving regular expressions over paths:
4 Note that this expression is not the same as the regular expression {xcomp⃰ | comp⃰} which denotes
paths containing either any number of xcomps or any number of comps: xcomp xcomp xcomp or
comp comp, but not xcomp comp.
regular expressions and regular languages 209
(f a suff(a, α)) = v
for some symbol a, where suff(a, α) is the set of suffix strings s such that
as∈α.
The definition in (27) is stated with reference to the predicate suff, which takes
two arguments: a string a and a regular language α, a set of strings composed of
grammatical functions. suff(a, α) is the set of strings which are possible continu-
ations of a in the regular language α. For example:
The noun janyi-ni ‘dog.m-erg’ bears an ergative marker -ni, signaling its gram-
matical function: nouns marked with ergative case are subjects. According to
Nordlinger’s theory of constructive case, casemarking specifies the syntactic envi-
ronment within which the casemarked phrase must appear. The f-structure corre-
sponding to the n node dominating janyi-ni is:
(30)
d
(31)
d
b
In (31), the f-structure for janyi-ni is the one labeled d; it is required to bear the
subj relation within its clause. Nordlinger enforces this requirement by means of
the following lexical entry for janyi-ni:
The first three equations state that the f-structure d for janyi-ni must have a pred
with value ‘dog’, the value m for its gend attribute, and the value erg for its case
attribute, as shown in (30). The expression (subj ↑) in the fourth line of this lexical
entry is different from the other expressions in (32), in that the attribute subj
appears to the left of the f-structure designator ↑. This inside-out expression refers
to an f-structure through which the path subj leads to the f-structure ↑; this is the
outermost f-structure, labeled b, in (31). In (32), this expression is an existential
constraint (see §5.2.6) requiring such an f-structure to exist.
Formally, (a f ) is the f-structure whose value for the attribute a is f :
In (32), the inside-out functional uncertainty path consists of just one attribute,
subj. Longer paths in an inside-out equation are interpreted incrementally, as with
outside-in expressions (see §5.2.1):
the pcase attribute 211
Notice that even when the inside-out path is fixed and the expression con-
taining it appears to be determinate, it may denote any of several f-structures.
Consider the f-structure in (37) for a verb like seem, whose subject is shared
with the subject of its open infinitival complement according to the lexical entry
in (36) (see Chapter 15 for a discussion of verbs taking an open complement
xcomp):
In (37), the f-structure for David, labeled d, is the subj of two f-structures: the
f-structure for seem, labeled s, and the f-structure for yawn, labeled y. In this case,
we can construct an inside-out expression rooted in the f-structure d, (subj d),
which refers to either of the two f-structures of which d is the subj: (subj d) is
either s or y.
Kaplan and Bresnan further propose that the value of the pcase attribute is also
the attribute whose value is the oblique phrase, so that oblgoal is an attribute
name as well as a value. The following annotation on the PP phrase structure node
accomplishes this:
(39) PP
(↑ (↓ pcase)) = ↓
This annotation appears as a part of the rule expanding V! . The full expansion of
the V! node is as follows:
V!
PP
V NP
(↑ (↓ pcase)) = ↓
P!
gave
the book ↑=↓
P
NP
↑=↓
to
N
(↑ pcase) = oblgoal
Chris
Using mnemonic names for f-structures, such as fpp for the f-structure correspond-
ing to the PP node in the c-structure tree, the equations in (40) for the PP and the
nodes it dominates are instantiated as in (41):
5 Kaplan and Bresnan’s analysis prefigures Nordlinger’s (1998) theory of constructive case, discussed
in the previous section, according to which the grammatical function of an argument is specified by
the case morpheme with which it appears.
talking about sets 213
(41) PP
(fv (fpp fpp
P
fpp = fp
P NP
fp = fp
to N
(fp
Chris
These equations tell us that the f-structure fpp corresponding to the PP node is the
same as the f-structures fp! and fp corresponding to the P! and P nodes, and that fp ’s
pcase is oblgoal . Thus, we have the following equivalences:
Substituting oblgoal for (fpp pcase) in the first equation in (42), we have:
(45)
’
fv
fpp , fp , fp
of conjuncts; or for the modifiers of a phrase, where any number of modifiers may
appear. In the following, we discuss ways of describing sets and constraining their
members. We first provide definitions for the standard symbols which are used in
talking about sets and relations between them:
(46) ∈ Set membership. a ∈ S means that a is a member of the set S.
&∈ Negated set membership. a &∈ S means that a is not a member of the set S.
⊆ Subset. A ⊆ B means that A is a subset of B: that is, all of the members of
A are also members of B. If A is a subset of B and vice versa (A ⊆ B and
B ⊆ A), then A = B.
⊂ Proper subset. A ⊂ B means that A is a proper subset of B: that is, all
of the members of A are also members of B, and B has some additional
members that are not in A.
∪ Union. A ∪ B is the set which is the union of A with B, which is the result
of combining the two sets. In other words, the union of set A with set B
contains all members of set A as well as all members of set B.
∩ Intersection. A ∩ B is the set which is the intersection of A with B, which
contains all of the members that A and B have in common.
∅ The empty set, sometimes written as { }. ∅ is a set that has no members.
A! Set complement. In defining the complement (A! ) of a set A, we must
make reference to the universe, which consists of all possible members
of the set; the complement of a set A is then the set containing all of the
elements in the universe that are not in A. For example, suppose that
the universe is all of the letters in the Roman alphabet, and A is a set
consisting of all strings containing any number of as, including none
({𝜖, a, aa, aaa, aaaa, . . . }):6 this is the set characterized by the regular
expression a⃰, as discussed in §6.1. The complement A! is, then, the set of
all strings that contain any letter other than a.
(47) V! −→ V NP PP∗
↑=↓ (↑ obj) = ↓ ↓ ∈ (↑ adj)
The expression PP⃰ represents a sequence of zero or more PPs. What about the
annotation ↓ ∈ (↑ adj)? This annotation means that each PP that appears is an
adjunct, and so its f-structure is a member (∈) of the adj set of the mother’s
f-structure ↑. That is, there may be zero or more occurrences of the following
annotated node:
6 Recall that 𝜖 is the empty string: see example (3) in this chapter.
talking about sets 215
(48) PP
↓ ∈ (↑ adj)
(49) (↑ adj ∈) = ↓
This expression uses the set membership symbol ∈ as an attribute and states that ↓
is a member of the set (↑ adj). Expressions such as these are sometimes useful in
stating constraints on set members, particularly in expressions involving inside-out
functional uncertainty, discussed in §6.1.3. The two expressions in (50) are equiva-
lent; each states that the f-structure ↓ is a member of the set of f-structures (↑ adj):
(50) ↓ ∈ (↑ adj)
(↑ adj ∈) = ↓
The c-structure and f-structure for a V! like yawn in class on Monday are:
(51) V
PP PP
V
↓∈ (↑ ↓ ∈ (↑
V P P
yawn P NP P NP
in N on N
class Monday
’
’
’
As the annotations on the rule require, the f-structure of each modifying adjunct
PP is a member of the adj set of the f-structure for the mother V! node.
Formally, an expression involving set membership is defined as we would
expect:
Although both David and Chris are singular phrases, the coordinate structure as a
whole is plural. The c-structure and f-structure for such an example are:
(56) NP −→ NP Cnj[main] NP
↓∈↑ ↑=↓ ↓∈↑
(↑ index num) = pl
f-structures for David and Chris as members, the set representing the coordinate
structure has the attribute index whose value contains the attribute-value pair
)num,pl*, signifying plural number. What does it mean to specify a property of
a set in this way?
Modifiers can also modify a coordinate structure as a whole, and so the adj feature
is also a nondistributive feature. For example, the adverb alternately in the sentence
David alternately sneezed and coughed modifies the conjoined phrase sneezed and
coughed: the sentence does not mean that David alternately sneezed and that he
also alternately coughed, but that he alternated in performing the two actions. We
assume the adjunction rule in (59), which adjoins an AdvP at the VP level, as well
as the VP coordination rule in (60):
(59) VP −→ AdvP VP
↓ ∈ (↑ adj) ↑=↓
(60) VP −→ VP Cnj[main] VP
↓∈↑ ↑=↓ ↓∈↑
218 syntactic relations and syntactic constraints
sneezed V
coughed
6.3.2.2 Distributive features In contrast, a distributive feature and its value are
a property of each member of the set, not of the set as a whole. We assume that all
features other than adj, index, conj, and preconj are distributive. Example (61)
shows that grammatical functions such as subj are distributive attributes, since
specifying the f-structure for David as the subj of the coordinate phrase amounts
to specifying that f-structure as the subj of each conjunct: hence the link between
the subjects in the f-structure in (61).
To take another example, suppose that concord is classified as a distributive
feature, and that we want to specify the concord case of a particular coordinate
phrase f as nom. In this situation, specifying the set f as having a particular
value for concord means that each member of the set f —here, d and c—must
be specified with that value for concord:7
7 Though concord features such as case are often analyzed as distributive features, Peterson
(2004a) and Hristov (2012: Chapter 5; 2013b) argue that case is best treated as a nondistributive feature
talking about sets 219
When they are not features of sets, the behavior of distributive and nondistributive
features is the same.
As features of sets, distributive and nondistributive features are formally treated
in the following way (Dalrymple and Kaplan 2000):
(f a suff(a, α)) = v
for some symbol a, where suff(a, α) is the set of suffix strings s such that
as ∈ α.
in light of acceptable examples involving case mismatch. If case is distributive, case mismatch is not
allowed, since any specification of the case value for the set as a whole applies to all of the conjuncts.
See §2.5.7.5 for further discussion.
220 syntactic relations and syntactic constraints
(66) Values for the pers feature according to Dalrymple and Kaplan (2000):
{s,h}: first person
{h}: second person
{ }: third person
(67) We yawned.
’
The closed set description characterizing the value of the pers attribute for the
pronoun we is:
This description differs from the open set description in (69), which is consistent
with the presence of other members of the set; the constraint in (68) is not:
talking about sets 221
As Sadler (2011) observes, any analysis that is expressed with set values is also
expressible using a feature-based representation, though the reverse does not hold.
In particular, it is possible to translate any set-based analysis of a complex feature
value into a feature-based analysis by representing the presence of a member of the
set with a positive (+) value for the corresponding attribute, and its absence by a
negative (−) value:
(71)
representation: representation:
+
+
−
+
−
Third person { } −
(72) Values for the pers feature according to Vincent and Börjars (2007):
{s,h}: first person inclusive
{s}: first person exclusive
{h}: second person
{ }: third person
(73)
representation: representation:
+
+
+
−
−
+
−
Third person { } −
8 Kaplan (2017) points out that subsumption (§6.9.2) can be used in a set-based analysis to capture
many of the effects of underspecification in a feature-based setting.
local names for f-structures 223
9 Nordlinger (1998) uses the left arrow ← to refer to the f-structure of the left sister, and the right
arrow → to refer to the f-structure of the right sister. We prefer the use of 𝜙(< ⃰ ) and 𝜙(⃰ >) for the
f-structures corresponding to the left and right sisters of a node, since this avoids confusion with the
use of the ← symbol in off-path constraints: see §6.6.
10 In §17.1.2, we augment the f-structure for relative clauses with an attribute relpro, whose
value is the f-structure of the relative pronoun within the clause-initial phrase. Although the pres-
ence of the relpro attribute would allow for a considerable simplification in the analysis of this
example, we omit this attribute here in order to provide a clear illustration of how a local name is
used.
224 syntactic relations and syntactic constraints
This rule states that a Russian noun phrase consists of a head noun N and a CP
relative clause. The first cp constraint, ↓ ∈ (↑ adj), requires the CP’s f-structure to
be a member of the set of modifiers of the NP (see §6.3.1 for more on specification
of set membership). This is true of the f-structure in (75).
The relative clause CP contains a relative phrase, the phrase okolo kotorogo ‘near
which’ in (74). The CP rule (not displayed here) ensures that the f-structure of this
relative phrase is a member of the dis set within the relative clause. According to
the rule in (76), this dis f-structure must contain a relative pronoun at some level
of embedding gf⃰ inside the dis. This f-structure is referred to by a name local to
this rule element as %relpro:
The final three annotations in (76) place further constraints on the f-structure
%relpro: it must have a prontype of rel (on prontype, see 2.5.7.7), and its num
and gend must match the num and gend of the mother NP:
These constraints are satisfied if %relpro names the f-structure labeled g in (75).
Using a local name like %relpro is essential in this instance: the use of a local name
ensures that all of the constraints in (78) refer to the same f-structure. In particular,
a set of expressions like the following are not equivalent to those in (78):
The equations in (79) require some f-structure in the dis set to contain an f-
structure with a prontype of rel, a possibly different f-structure to have the
same num as the full noun phrase, and a possibly different f-structure to have the
same gend as the full noun phrase; crucially, these constraints impose no require-
ment for the same f-structure to satisfy all of these constraints. It is the use of a
local name that enforces the proper requirement.
Verbs allowing extraction are often called bridge verbs, while those disallowing
extraction are called nonbridge verbs.
There is no reason to assume that the grammatical function of the clausal
complements of these two verbs differs. Other than this difference between them,
they behave the same syntactically; both bear the grammatical function comp.
A sentence with a bridge verb allowing extraction has an f-structure like the
following:
We propose that a nonbridge verb specifies that its complement bears a feature
that bridge verbs lack, which we will call ldd (for long-distance dependency),
with value −. The path in a long-distance dependency may not pass through an
f-structure with this feature:
’
−
off-path constraints 227
The attributes comp and obj do not reflect the prohibition against extraction.
Instead, this requirement must be stated “off the path” characterizing the depen-
dency, as an additional condition on the f-structures along the path. In this case,
we would like to express the following constraint:
(85) A comp in the extraction path must not contain the pair )ldd, −*.
In the expression in (87), the right arrow → stands for the value of the attribute
comp:
(87) comp
(→ ldd) &= −
In (88), the comp attribute is boxed, and the f-structure denoted by → is labeled s:
’
−
s
The f-structure s contains the attribute ldd with value −. This is forbidden by
the negative constraint (→ ldd) &= − in (86), accounting for the ungrammaticality
of (88).
Slightly more generally, we can use an expression like the following to constrain
long-distance dependencies in English:11
This expression indicates that any number of occurrences of the annotated comp
attribute comp displayed in (87) can occur in the long-distance path; in
(→ ldd) &= −
other words, the f-structure ↓ bears some grammatical function gf embedded
inside any number of comps, as long as none of the comp f-structures on the path
contain the pair )ldd, −*.
The occurrence of comp which is annotated with the off-path constraint is boxed
in (91), and the left arrow ← in the off-path constraint refers to the f-structure
in which the annotated attribute comp appears, the one labeled f . The off-path
constraint (← attr) = value requires f to have an attribute attr with value
value:
If we replace the left arrow by the right arrow, the result is as in (92): it is not f
which contains the attribute and value in question, but g, the value of the comp
attribute within f :
These requirements hold for coordinate structures as well, with the result that
the conjuncts in a coordinate structure may bear different cases. For example, a
nominative noun phrase may be coordinated with an accusative numeral phrase,
forming a coordinated subject in which each conjunct obeys the casemarking
rules in (93):
Przepiórkowski and Patejuk propose two analyses for this pattern; the second,
which they call the “conservative solution,” involves off-path constraints and the
features sc and acm. The sc feature marks f-structures which are assigned struc-
tural case, and the acm attribute with value rec marks f-structures for numerals
governing a genitive noun. These features are involved in these implicational con-
straints on casemarking (see §5.2.10 for definition and discussion of implicational
constraints):12
(f ) =c +
(95) ⇒ (f )=
(f )=
(f ) =c +
b. ⇒ (f )=
(f ) =c
The constraint in (95a) specifies that subjects that are assigned structural case
(these have the value + for the attribute sc) and are not numeral phrases (and
so do not have the value rec for the feature acm) are assigned nominative case.
The constraint in (95b) specifies that subjects that are assigned structural case and
are numeral phrases are assigned accusative case.
12 Przepiórkowski and Patejuk’s treatment of casemarking has been simplified for the purposes of
exposition; see their paper and §2.5.7.5 for more discussion and a formal analysis of the case attribute
and its value.
230 syntactic relations and syntactic constraints
The issue, as Przepiórkowski and Patejuk observe, is that this constraint cannot
be evaluated with respect to a coordinate structure as a whole, since, as shown
in (94), some conjuncts in a coordinate phrase might be assigned nominative
case according to constraint (95a), while others are assigned accusative case
according to constraint (95b). Their solution involves passing the constraint into
the coordinate structure via an off-path constraint on a distributive feature, so that
the constraint can be satisfied independently for each conjunct.
According to Przepiórkowski and Patejuk’s analysis, casemarking for Polish
subjects is controlled by the following constraint, involving a conjunction of the
implicational constraints in (95):
Using the f-structure variables ← and →, any kind of constraint can be written
as an off-path constraint: defining equations, constraining equations, existential
constraints, and other kinds of f-descriptions may be specified. We return to a
discussion of off-path constraints and long-distance dependencies in Chapter 17.
6.7 Templates
The previous examples used functional descriptions associated with lexical entries
and phrase structure rules in determining the f-structure for a particular annotated
templates 231
c-structure tree. The lexical entries and phrase structure rules of a language are
associated with f-descriptions that characterize and constrain the f-structures
(and, as we will see, other structures) of the utterances in which they appear. In
formulating a theory of the structure of the lexicon and permissible annotations
on phrase structure rules, it is important to be able to specify commonalities in
f-descriptions associated with lexical entries and annotated phrase structure rules.
Templates are used to express these commonalities.
This lexical entry contributes a pred value for the verb, and also contains informa-
tion about the form of the verb and its subject agreement requirements.
Some of the elements of this f-description are shared with other verbs:
for example, all present tense verbs contribute the equations (↑ vtype) = fin
and (↑ tense) = prs, and all third person singular verbs contribute (↑ subj index
pers) = 3 and (↑ subj index num) = sg. We can define the templates present
and 3sg to encode this information, associated with the appropriate pieces of the
functional description:
The template name present names the functional description consisting of the
two equations (↑ vtype) = fin and (↑ tense) = prs, and similarly for 3sg. In fact,
however, we can further subdivide the functional description present in terms of
the following template definitions:
nodes in the graph representing relations between templates. In the case at hand,
we have the template hierarchy fragment in (102), showing that the template
present is defined by reference to the template prestense and the template
finite:
(102) present
prestense finite
Note that this template hierarchy is unlike the type hierarchies of Head-Driven
Phrase Structure Grammar (Pollard and Sag 1987, 1994; Ginzburg and Sag 2000)
in that it encodes a relation between f-descriptions rather than a relation between
structures. The template hierarchy in (102) indicates merely that the present
template is defined by reference to the prestense and finite templates. Templates
stand for f-descriptions, and f-descriptions can be conjoined, disjoined, or negated.
Thus, the f-description contributed by the daughter nodes in this hierarchy may
be related in a complex way to the f-description of the mother node, including via
negation. This is not possible in a type hierarchy, where the hierarchy represents
inheritance in a semilattice. Asudeh et al. (2008, 2013) provide further discussion
of this point.
We can also subdivide the 3sg template as follows:
(103) ≡ (↑
≡ (↑
(105) 3sg
3PersonSubj SingSubj
Finally, we can define a template pres3sg for verb forms like yawns:
(107) Pres3sg
present 3sg
(109) @intransitive(yawn)
Given these template definitions, the lexical entry for yawns is:
(112) a. The package arrived unopened. (subj control: the package is unopened)
subj
b. John will serve the fish to you raw. (obj control: the fish is raw)
obj
c. I sent you the letter sealed. (obj𝜃 control: the letter is sealed)
obj𝜃
arrived unopened
In this configuration, the subject of the xadj unopened is required to be the same
as the matrix subj, obj, or obj𝜃 :
(114) VP −→ VP AP⃰
↑=↓ ↓ ∈ (↑ xadj)
(↑ {subj|obj|obj𝜃 }) = (↓ subj)
permit control only by the matrix subject, and not by the object or other
arguments:
(116) IP −→ AP⃰ IP
↓ ∈ (↑ xadj) ↑=↓
(↑ subj) = (↓ subj)
(117) IP
AP IP
NP I!
Sure of winning
N VP
Mary
entered the competition yesterday
The annotations on AP in the rules in (114) and (116) are similar but not the same,
and indeed, as we will see in §15.8.1, Icelandic makes use of a phrase structure rule
for depictive xadj that is identical to (114). By using a template, we can capture
the commonalities and differences in these rule annotations within and across
languages.
We can define the parametrized template xadj in the same way as for the
parametrized intransitive template in (108), with an argument for the gram-
matical function of the matrix controller:
Given this definition, we can rewrite the rules in (114) and (116) in a way
that illuminates the similarities and differences between them; the similarities
are embodied in the call to the template xadj, while the difference in control
possibilities between the two constructions is reflected in the different arguments
to each of the templates:
c-structure rule macros 237
(119) a. VP −→ VP AP⃰
↑=↓ @xadj({subj|obj|obj𝜃 })
b. IP −→ AP⃰ IP
@xadj(subj) ↑=↓
(121) I! −→ I VP
↑=↓ ↑=↓
C! −→ C IP
↑=↓ ↑=↓
We can encode the relevant generalization by means of the rule macro in (122),
which provides a definition of coheadedness, and captures the commonalities
between these two rules. The definition of a rule macro is similar to a template
definition; in particular, rule macros can be parametrized, taking arguments. The
macro definition in (122) takes two arguments, _h1 and _h2, representing the two
cohead categories in the rule:
238 syntactic relations and syntactic constraints
This macro definition is unlike a metacategory in that its categories are not
plain (unannotated) categories, but are annotated with functional descriptions;
it is unlike a template in that the definition includes one or more annotated c-
structure categories, not only an f-description as in the definition of a template.
Of course, metacategories and templates can be used in the definition of a rule
macro, in the same way as they are normally used on the right-hand side of a phrase
structure rule.
Using the macro in (122), we can restate the two rules in (121) as in (123), by
using the cohead rule macro. The rules in (123) are exactly equivalent to the rules
in (121):
(124) F-command:
f f-commands g if and only if f does not contain g, and all f-structures that
contain f also contain g.
In (125a) and (125b), the f-structure labeled f f-commands the f-structure labeled
g. In (125a), but not in (125b), g also f-commands f :
(125) f f-commands g:
f[ ] f[ ]
a. b.
g[ ] g[ ]
The definition of f-command given in (124) is correct for cases like (125). However,
as pointed out by Ron Kaplan (p.c.), this definition may not make the right
predictions in cases in which two attributes share the same value. Consider the
relations between f-structures 239
f-structure in (126), where the f-structure labeled f is the subj as well as the
xcomp subj:
(126) f[ ]
g[ ]
h
The f-structure labeled f in (126) does not f-command the f-structure labeled g
according to the definition in (124), because there is an f-structure (namely h) that
contains f but does not contain g. For the f-command relation to hold between f
and g, we can formulate a new definition of f-command using inside-out functional
uncertainty (defined in §6.1.3):
6.9.2 Subsumption
Subsumption is a relation that holds between two f-structures f and g if g is
compatible with but perhaps has more structure than f . In other words, f subsumes
g if f and g are the same, or if g is the same as f except that it contains some
additional structure that does not appear in f . For example, the f-structure labeled
f in (129) subsumes the f-structure labeled g:
(129) f subsumes g:
’
’
f g
(130) Subsumption:
An f-structure f subsumes an f-structure g (f 2 g) if and only if:
f = g; or
f and g are sets, and for each member f1 of f there is a member g1 of g such
that f1 2 g1 ; or
f and g are f-structures, and for each attribute-value pair )a, v1 * ∈ f , there
is a pair )a, v2 * ∈ g such that v1 2 v2 .
6.9.3 Generalization
Intuitively, the generalization of two f-structures is the structure that they have in
common. For example, in (131) the f-structure labeled f is the generalization of
the f-structures g and h:
relations between f-structures 241
(132) Generalization:
An f-structure f is the generalization f1 3 f2 of two f-structures f1 and f2 if
and only if:
f1 = f2 = f ; or
f1 and f2 are f-structures, and for each pair )a, v1 * ∈ f1 , if there is a pair
)a, v2 * ∈ f2 , then )a, v1 3 v2 * ∈ f ; or
f1 and f2 are sets, and each member of f is the generalization of some
member of f1 and some member of f2 .
According to the definition in (132), both of the following two sets constitute a
generalization of the sets in (133):
>? @? @A
(134) a. f1 v1 f4 v4
>? @? @A
b. f2 v2 f3 v3
6.9.4 Restriction
The restriction of an f-structure with respect to an attribute can be intuitively
defined as the f-structure that results from removing the attribute and its value
(Kaplan and Wedekind 1993). The f-structure labeled g |TENSE in (135) is the
restriction with respect to tense of the f-structure labeled g:
(135) g |TENSE is the restriction of g with respect to tense:
’
g
’
g |TENSE
More formally, Kaplan and Wedekind (1993) define the restriction of an f-structure
f with respect to an attribute a as follows:14
(136) Restriction:
f |a ≡ {)s, v* ∈ f : s &= a}
If a is a set-valued attribute, restriction removes the specified member of
the set:
B
f |a if (f a) = {v}
f |)av* ≡
f |a ∪ {)a, (f a) − {v}*} otherwise
Restriction is useful if an f-structure plays two syntactic roles, but with different
syntactic requirements associated with each role. For example, let us assume that
some f-structure f is shared as the subj value of two different f-structures g1 and
g2 , but that f must take on a different case value in each structure. The equation in
14 The definition in (136) allows us to characterize a set by describing its elements, according to the
following format:
Y = {X : ...constraints on X... }
According to this schematic definition, Y is a set containing all elements X that satisfy the constraints
on X that appear after the colon. For example, in the first part of the definition in (136), f |a is a set of
all pairs )s, v* which are in f and which satisfy the constraint s & = a. In other words, the first part of the
expression, )s, v* ∈ f , directs us to consider all of the attribute-value pairs )s, v* in the f-structure f , and
the constraint s & = a restricts attention to those attribute-value pairs in f in which the attribute is not a.
relations between f-structures 243
(137a) requires all of the attribute-value pairs of f other than case to be the same
as the attribute-value pairs of g1 ’s subj other than case, and the equation in (137b)
imposes the same requirement for g2 :
We can then specify different case values for the subjects of g1 and g2 ; the
constraints in (138) are consistent with the requirements in (137):
Kaplan and Wedekind (1993) use restriction in their analysis of the semantics of
modification. We do not adopt their analysis in this book; we discuss the syntax and
semantics of modification in Chapter 13. Asudeh (2011, 2012) uses the restriction
operator in his analysis of syntactically inactive resumptive pronouns (§17.2).
Intuitively, the priority union of two f-structures contains all the structure that each
f-structure has, with the f-structure on the left of the slash “winning” if there is a
conflict. For example, in (140) f /g is the priority union of f with g:
The priority union f /g has all the structure in f as well as all the structure in g that
does not conflict with f .
15 The forward slash notation is used both for priority union of f-structures (f /g) and as the “Ignore”
operator in c-structure rules (§6.1.1).
244 syntactic relations and syntactic constraints
An analysis of gapping that appeals to priority union might propose that the final
f-structure f for the second conjunct Matty Ken is obtained by taking the priority
union f /g: in effect, the f-structure for Matty would replace the f-structure for
David, and similarly for the f-structures for Chris and Ken:
’
f g
Priority union produces for the second conjunct an f-structure like the one that
would be associated with a sentence like Matty saw Ken.
Kaplan (1987) purposely formulated the definition of priority union to refer only
to the top-level attributes of f and g; the definition given in (139) is not recursive.
However, later work (for instance, Brun 1996b) assumes a recursive definition for
priority union like the following:
Future work will show which of these definitions is the most useful one.
c-structure/f-structure constraints 245
..
.
XP
A tree like this one is not permitted: an XP cannot dominate another XP without
also dominating some other material as well. If this were permitted, there could be
a chain of XPs of unbounded length dominating any XP, giving rise to an infinite
number of possible constituent structure trees for that XP:
(146) Disallowed:
XP
..
.
XP
..
.
XP
Of course, a phrase may dominate a phrase of the same type if other material is
present as well:
(147) Permitted:
VP
V!
V VP
want
to go
246 syntactic relations and syntactic constraints
This constraint is also discussed by Pereira and Warren (1983), who refer to the
constraint as “off-line parsability”; this is because their formulation depends on
the application of the Nonbranching Dominance Constraint “off-line,” after the
parsing algorithm has been applied to derive a set of trees for a string. Johnson
(1988) also provides a definition of off-line parsability that is very similar to the
definition in (144).
In subsequent work, Kaplan and Maxwell (1996) revise the Nonbranching
Dominance Constraint to allow nonbranching dominance chains with nodes of
the same category if the two nodes have different functional annotations. Under
this view, the following dominance chain is ill-formed:
..
.
XP
↑=↓
..
.
XP
(↑ gf) = ↓
Under these assumptions, the final definition of the Nonbranching Dominance
Constraint is:
(150) Nonbranching Dominance Constraint, final version:
A c-structure derivation is valid if and only if there are no categories
with the same functional annotation appearing twice in a nonbranching
dominance chain.
Kaplan and Bresnan (1982) show that the Nonbranching Dominance Constraint
is important in proving that the membership problem for lexical functional gram-
mars is decidable—that it is always possible to determine whether a string is
acceptable according to a given grammar:
c-structure/f-structure constraints 247
Sag et al. (1985) claim that the verb become requires either an AP or NP comple-
ment, and cannot appear with a PP or VP complement:
V AP
waxed A
poetical
N VP ’
p
Kim V
V AP
−1
waxed A
poetical
For each f-structure f , the inverse correspondence relation 𝜙−1 (f ) gives the
c-structure nodes that are associated with that f-structure; the relation between
f-structures and their corresponding c-structure nodes is therefore not a function,
because more than one c-structure node can correspond to a single f-structure.
For the f-structure labeled p in (155), 𝜙−1 (p) yields the nodes labeled AP and A.
Example (156) gives the c-structure and f-structure for the ill-formed sentence
Kim waxed a success:
waxed D N −1
a N
success
For this example, the f-structure labeled s is associated via the inverse 𝜙
correspondence with three nodes, those labeled NP, N! , and N.
We can now state the categorical requirement imposed on the complement of
the verb wax by using the predicate CAT (Kaplan and Maxwell 1996; Crouch et al.
2008), defined in terms of the inverse 𝜙 correspondence and the λ function from
c-structure nodes to category labels (see §5.1.2). The CAT predicate allows the
250 syntactic relations and syntactic constraints
Thus, we posit the following category requirement, lexically associated with the
verb wax:
This requirement ensures that one of the c-structure nodes associated with the
xcomp has the category AP, as required.
According to the generalization that Simpson formulates, there are two possibilities
for phrasal expansion of I! . In (161), with a “non-null” auxiliary, the first daughter
of I! is I:
16 This definition of CAT conforms to the definition given by Kaplan and Maxwell (1996) and
Crouch et al. (2008), but differs from the definition in Dalrymple (2001).
17 The symbol ∃ is an existential quantifier: an expression like ∃n.... asserts that some individual n
exists that meets the description represented by the following formula. The first part of the expression,
∃n ∈ 𝜙−1 (f ), directs us to consider the nodes n which are related to f via the inverse 𝜙 correspondence,
and λ(n) ∈ C requires the category label λ(n) of at least one of those nodes to be in the set of node
labels C.
c-structure/f-structure constraints 251
wawirri ka NP V
kangaroo
N panti-rni
ngarrka-ngku
IP
NP I
’
N S
Japanangka-rlu V NP
pantu-rnu N
marlu
kangaroo
(163) I! −→ { I | 𝜖 } S
↑=↓ (↑ aspect) = prf ↑=↓
In this rule, the symbol 𝜖 corresponds to the empty string and represents the
absence of a phrase structure constituent. Importantly, the rule does not license
252 syntactic relations and syntactic constraints
the presence of an empty category or node in the c-structure tree. Instead, it simply
constitutes an instruction to introduce some functional constraints in the absence
of some overt word or phrase. No empty node is introduced into the tree, as (162)
shows.
The rule in (163) contains a disjunction: in one case, an auxiliary appears in I,
while in the other case no auxiliary appears, and the sentence is interpreted as
having perfect aspect. In fact, the rule in (163) is exactly equivalent to:
(164) I! −→ { I S | S }
↑=↓ ↑=↓ ↑=↓
(↑ aspect) = prf
However, although the rule in (164) also represents the two possible phrase
structure expansions of I! , it is more cumbersome and fails to express Simp-
son’s generalization. By using the symbol 𝜖, the generalization can be expressed
concisely.
(167) Who did you say that, just a minute ago, sneezed?
The pattern is reversed for relative clauses, where the complementizer is required
in subject relativization in most varieties of contemporary English (Bresnan 1977):
18 As discussed in §6.1, the symbol & that is generally used for the intersection of two regular
languages is different from the symbol ∩ used for the intersection of two sets.
254 syntactic relations and syntactic constraints
Asudeh (2009) discusses these and other patterns involving subject and nonsubject
extraction in question formation and relative clause formation in detail; for the
full account, see Asudeh (2009). He proposes the following constraint, lexically
associated with the complementizer:
Asudeh also defines a function N (for Next), which relates a word in a string to
the next word in the string. N is defined for all words in the string except for
the last word, since there is no word immediately following the final word in the
string. The inverse of this function, N −1 , relates a word to the previous word in the
string.
(171) N(word1) = word2 if and only if word2 immediately follows word1 in the
string.
N −1 (word2) = word1 if and only if word2 immediately precedes word1 in
the string.
Given this definition of N for the next word in a sentence and the π function from
words to terminal nodes in the c-structure, we can use the familiar 𝜙 function
to refer to the f-structure corresponding to the c-structure node dominating the
next word in the string. We follow Asudeh in using the expression 5 for this
f-structure.19
19 Our definition differs slightly from the one provided by Asudeh (2009) because we assume that
words in the string are related to nodes in the c-structure via the function π; see §3.5 and §11.4.1 for
more discussion of our architectural assumptions.
linear relations and the string 255
Note that 5 is not defined for the last word in the string (which has no word
following it), and ≺ is not defined for the first word in the string (which has no
word preceding it).
Either 5 or ≺ may appear in a lexical entry. Since these expressions are defined
over words in a string, they may not appear in a c-structure annotation, since
their definitions rely on the relation N which holds between words in the string
rather than the c-structure precedence relation discussed in §6.11.3. Consider the
following configuration:
6 7
(174) g[ ]
f
h[ ]
C1
C2 C3 𝜙
word1
word2 word3
In (174), the terminal string is word1 word2 word3. According to the definition of
the N function, the following statements define the Next function for this string:
We can restate the second and third line of (176), using the N function to refer to
the next word, as:
(178) 𝜙(π(N(word1))) = g
𝜙(π(N(word2))) = h
Thus, for word1, 5 would refer to g, and for word2, 5 would refer to h.
Still following the analysis of Asudeh (2009), we can now state the CAE con-
straint in formal terms:
This constraint appears in the lexical entry of the complementizer that, and rules
out the possibility that the f-structure of the word following that has a subject
(5 subj) that (i) has a c-structure realization (realized(5 subj)) and (ii) is a mem-
ber of the dis set, establishing a long-distance dependency (dis ∈ (5 subj)).20 This
constraint is satisfied in one of three circumstances. First, if the f-structure of the
word following the complementizer that has no subj, then (5 subj) does not exist
and the CAE Constraint is satisfied. Second, if the f-structure of the following word
has a subject, but it is not realized, the statement realized(5 subj) is false and the
CAE Constraint is satisfied (this is the case for relative clauses with no relative
pronoun, such as This is the person that/⃰∅ sneezed). Third, if the f-structure of the
following word has a subject that is realized but does not appear in the dis set, the
f-structure (dis ∈ (5 subj)) does not exist, and the CAE Constraint is satisfied;
this covers the case of nonsubject relatives such as Who do you think [that Kim
saw]? as well as declarative complements with no dis, such as I think [that Kim saw
David]). The analysis makes crucial use of the N relation, defined over elements of
the string.
20 The expression (dis ∈ (5 subj)) is an existential constraint requiring that the f-structure (5 subj)
is a member (∈) of a dis set. In this expression, the set-membership symbol ∈ is used as an attribute
selecting a member of the dis set (§6.3.1).
linear relations and the string 257
In contrast, there are no restrictions on the relation between the null pronominal
and the definite antecedent. Whether imooto ‘sister’ precedes or follows Taroo, we
may interpret imooto with a null possessive pronoun anteceded by Taroo:
Simplifying Kameyama’s analysis somewhat, these and other examples show that
the antecedent of an overt pronoun must precede the pronoun, while this con-
straint does not hold for null pronouns. The facts about pronominal binding in
Japanese can be given a uniform explanation in terms of f-precedence:
This generalization about anaphoric binding in Japanese holds under the following
definition of f-precedence (Kaplan and Zaenen 1989b; Kameyama 1989):
This definition appeals to the inverse relation 𝜙−1 , defined in §6.10.3, which
associates f-structures with the c-structure nodes they correspond to. It also relies
on the definition of c-precedence given in (180) in §6.11.3.
This definition of f-precedence states that an f-structure f f-precedes an f-
structure g if and only if all of the nodes corresponding to f c-precede all of the
258 syntactic relations and syntactic constraints
NP NP V
D N N sewasiteiru
be.taking.care.of
kare-no imooto-o Taroo-ga
What is the situation with null pronominals, pronouns that do not appear at
c-structure? Consider the c-structure and f-structure for (186), which is like (185)
except for having a null pronoun (realized only at f-structure) in place of the overt
possessive pronoun kare-no:
NP NP V
N N sewasiteiru
be.taking.care.of
imooto-o Taroo-ga
linear relations and the string 259
Crucially, the f-structure of the null pronoun does not correspond to any
c-structure node. According to the definition of f-precedence given in (184),
null elements vacuously f-precede and are f-preceded by all other elements in
the sentence; in particular, null pronouns vacuously satisfy the constraint of
being preceded by their antecedents. Thus, a uniform statement of antecedent
requirements for Japanese null and overt pronominals, together with the definition
of f-precedence given in (184), makes the correct predictions.
The definition of f-precedence in (184) is the one originally proposed by Bresnan
(1984) and used by Kameyama (1989) and by Kaplan and Zaenen (1989b) in their
analysis of word order in Dutch and German. A different definition is proposed by
Bresnan et al. (2016) in their analysis of weak crossover:
This definition gives a different result for f-structures that do not correspond
to any c-structure nodes, since such f-structures do not have a corresponding
rightmost node. Under definition 2 in (187), f-structures that do not correspond
to any c-structure node do not f-precede any other f-structure, and they are not
f-preceded by any other f-structure.
With the formal tools that have now been introduced, we are ready to begin
a full-scale excursion into new linguistic realms. In the following chapters, we
will explore other levels of nonsyntactic linguistic structure and how they are
represented.
7
Beyond c-structure and
f-structure: Linguistic
representations and relations
Any postulated component should be of a kind that one might want in a full description of
the basic facts of language, regardless of the way in which the modules are related to one
another.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
264 beyond c-structure and f-structure
1 For more on the history of the treatment of long-distance dependencies in LFG, see Dalrymple
et al. (1995d); we discuss long-distance dependencies in detail in Chapter 17.
modularity 265
7.2 Modularity
Analyzing an utterance requires the linguist to analyze different types of lin-
guistic structure and information; a modular approach requires different aspects
of linguistic structure to be represented as distinct, independent components
of the grammar. These separate components or grammatical modules each have
their own primitives and organizing principles, and therefore their own internal
structure and formal representation. For instance, a concept such as “noun phrase”
is relevant at the level of c-structure, but not f-structure; similarly, syllables are
important in a phonological analysis, but a semantic analysis of an utterance is
effectively blind to them.
A modular approach ensures that different types of linguistic information
are not conflated. This enables generalizations, whether universals or language-
particular constraints, to be stated for and apply only to the level of representation
for which they are relevant. Moreover, separating parts of the grammar in this way
minimizes the possibility of adverse effects should it become necessary to modify
the analysis of one component of the grammar: advances in the analysis of one
aspect of linguistic structure can be incorporated into the model without the need
for a wholesale revision of the grammatical architecture. This ensures that LFG’s
approach to the architecture of the grammar is stable yet flexible.
While the different modules of the grammar are separate in LFG, they exist in
parallel. This does not entail a one-to-one correspondence between the subparts of
structures at any of the levels, nor does it mean that the information encapsulated
within one module (for example, the precedence relation between c-structure
266 beyond c-structure and f-structure
2 Halvorsen’s analysis has been considerably simplified for presentation here, leaving aside many
of the details of the meanings and rules that he proposes. Further, we have modified his analysis
by assuming that a word like every is treated as a generalized quantifier (Barwise and Cooper 1981)
specifying that the every relation holds between two properties R and S. We discuss quantification in
more detail in Chapter 8. The lambda operator λ is explained in §8.4.1.
268 beyond c-structure and f-structure
Halvorsen also presents a set of rules for determining the meaning of an f-structure
based on its semantic forms. The following is a simplified version of his Rule III,
spec-pred configuration, which defines the meaning Mk of an f-structure fk :
Let us see how this rule operates in the analysis of the f-structure proposed by
Halvorsen for the noun phrase every horse. For this simple example, we will leave
out most of the detail in the f-structure:
We apply the rule in (3) in the following way. First, we inspect the f-structure in
(4), noting that it is of the form required by the rule in (3). Therefore, we introduce
a constraint on the structure M1 , the semantic structure corresponding to f1 : it
must contain the attribute predicate with value λP.λQ.Q(P). We also add the
additional constraint that M1 must contain the attribute arg1, and we determine
the value of that attribute by consulting the table in (2); in a similar way, we add
a constraint on the value of arg2. This yields the following set of constraints
on M1 :
These equations describe the following semantic structure M1 for the f-structure
f1 :
(6)
predicate λP.λQ.Q(P)
λR.λS.every (R, S)
M1 arg1
arg2 λx.horse (x)
defining relations between structures 269
We also need a set of rules for interpreting this structure: Halvorsen (1983) presents
a set of rules that produce logical formulas from these semantic structures. Among
them are rules like the following:
Applying this rule to the semantic structure for every horse in (6) produces the
following result, which is just what is required for the phrase every horse:
7.4.2 Codescription
Grammatical structures can also be related by codescription, where multiple struc-
tures are simultaneously described; this is the most common way in which the
relations between grammatical modules are encoded and constrained. For exam-
ple, annotated c-structure rules simultaneously encode constraints on (at least) two
kinds of grammatical information: c-structure category and f-structure.
As discussed in §5.1.2, nodes of the c-structure tree are associated with two
different functions: the 𝜙 function relates c-structure nodes to their corresponding
f-structures, and the λ function relates nodes to their labels. We can represent
the configuration in (9a) in a more explicit but also more cumbersome way
as (9b):
(9) a. IP
NP I
b. IP •
λ
NP • •
I
The trees in (9) are licensed by the phrase structure rule in (10a), which is
equivalent to the rule in (10b) except that the λ labeling function is explicitly
shown, and 𝜙(ˆ⃰) and 𝜙(⃰) appear rather than the abbreviatory ↑ and ↓ arrows in
the former. The format in (10a) makes it clear that two different aspects of structure
are constrained by this rule. The λ labeling function constrains the phrase structure
270 beyond c-structure and f-structure
category of the mother and daughter nodes,3 and the 𝜙 function constrains the
f-structures that correspond to the mother and daughter nodes. In other words,
this rule simultaneously describes, or codescribes, two levels of structure: the 𝜙
projection from c-structure—that is, the f-structures 𝜙(ˆ⃰) and 𝜙(⃰), and the λ
projection from c-structure nodes defining the category of the node:
(10) a. • −→ • •
λ(⃰) = NP λ(ˆ⃰) = IP
(𝜙(ˆ⃰) subj) = 𝜙(⃰) λ(⃰) = I%
𝜙(ˆ⃰) = 𝜙(⃰)
b. IP −→ NP I%
(↑ subj) = ↓ ↑=↓
(11) ) *
pred ‘yawn&subj'’ 𝜎 s:[]
𝜙 f
V subj [] s1 : [ ]
yawn
3 The constraint λ(ˆ⃰) = IP defining the category of the mother node could have been written on
either daughter node: here we have made the arbitrary decision to associate it with the second daughter.
4 As mentioned in §7.3, an alternative subscript notation f𝜎 , exactly equivalent to 𝜎(f ), is often used
for semantic structures.
defining interstructural constraints 271
Given the functions 𝜙 and 𝜎, we can define a function between c-structure nodes
and semantic structures as the composition of the two correspondence functions,
𝜎 ◦ 𝜙.5 In the case at hand, we can apply the composition function 𝜎 ◦ 𝜙 to the
c-structure node V to get the semantic structure s:
(14) 𝜙
VP ' (
pred ‘yawn&subj'’
V
tense pst
yawned
At c-structure, the distinction between the nodes labeled VP and V is clearly repre-
sented: two different c-structure nodes are involved. At f-structure, the distinction
between the two nodes is collapsed, since the two nodes correspond to the same
f-structure. Thus, any level of linguistic representation that must make reference
to a distinction between the V and the VP node cannot be a projection of the
f-structure, since the distinction between VP and V is collapsed at that level and
cannot be reintroduced in projections from the f-structure.
(15)
IP
NP I
5 The composition of two functions is obtained by taking the result of applying one of the functions
to its argument and then applying the other function to that result (see Partee et al. 1993: 33). The
composition of two functions is also a function.
272 beyond c-structure and f-structure
(16) IP f s[ ]
NP I
In the configuration depicted in (16), the f-structure labeled f is related via the 𝜙−1
correspondence to two c-structure nodes, labeled IP and I% , and the subj f-structure
is related via the 𝜙−1 correspondence to the node labeled NP. The definition of the
CAT predicate (given in (157) of Chapter 6) is stated in terms of the inverse of the
𝜙 function, 𝜙−1 :
Thus, the projection architecture allows for the statement of complex relations
among linguistic structures, and for constraining these relations in appropriate
ways.
In the remainder of Part II of this book, we will present the different linguistic
modules which the LFG projection architecture comprises, exploring their repre-
sentation and how they are related to one another by correspondence functions.
6 Some approaches allow phrase structure trees that violate some of these formal conditions; see,
for example, McCawley (1988). McCawley’s trees are, then, a more abstract representation of syntactic
information than the constituent structures of LFG.
further reading and related issues 275
(19) walk
John
◦
1
d
’
Andrews and Manning (1999) also present a very different view of the relations
between linguistic structures within the LFG framework, an approach that for-
mally resembles the work of Fenstad et al. (1987) as well as work in HPSG.
Andrews (2008) proposes a view of semantic structure and the syntax-semantics
interface which, unlike the codescription approach to be presented in Chapter 8,
proposes that semantic structure and semantic constraints are established on the
basis of description by analysis of the f-structure. Andrews motivates this alterna-
tive analysis by providing a set of arguments against the standard codescription
approach and in favor of his description-by-analysis alternative. Though we do
not adopt his analysis in this book, it is interesting as an alternative view of how
different aspects of linguistic structure are related.
In our discussion of linguistic modules, we will not touch on proposals for
enriching the projection architecture to enable the analysis of multimodal com-
munication: see Giorgolo and Asudeh (2011b) for a very interesting proposal to
augment the projection architecture to account for the role that gestures (specifi-
cally, spontaneous hand gestures) may play in interpretation.
8
Meaning and semantic
composition
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
syntax and semantic interpretation 277
However, this version of the rule-to-rule hypothesis is actually just one way of
enforcing an orderly theory of semantic composition, one in which the intuition
that the meaning of a whole depends on the meanings of its parts is made explicit
by defining the relevant parts as phrase structure constituents. In fact, research on
the syntax-semantics interface and semantic composition in LFG has shown that
we can remain faithful to the Principle of Compositionality without assuming that
rules for putting meanings together must depend on phrasal primitives such as
linear order and phrasal dominance.
Since the inception of semantic research in LFG, convincing arguments have
been made that semantic composition should proceed mainly by reference to
functional structure rather than constituent structure organization. As argued by
Fenstad et al. (1987: Chapter 2), the units that are primarily relevant for semantic
composition are units at f-structure and not necessarily at c-structure. For example,
as we have seen, a semantic unit may correspond to discontinuous portions of
the c-structure tree. Example (32) in Chapter 4, repeated as (1) here, shows that
the Warlpiri analog of the English phrase small child need not form a c-structure
constituent; the noun kurdu-ngku ‘child’ appears at the beginning of the sentence,
while its modifier wita-ngku ‘small’ appears at the end.
N I S
kurdu-ngku ka V NP
wajilipi-nyi N
wita-ngku
However, rules for semantic composition in both Warlpiri and English treat the
subject of the Warlpiri sentence kurdu-ngku ka wajilipi-nyi wita-ngku and that of
the English sentence The small child is chasing it as an f-structure constituent and as
a semantic unit; in fact, the rules for semantic composition in the two languages are
remarkably similar, considering the great differences between the two languages
at the c-structure level. Guiding semantic composition by reference to f-structure
278 meaning and semantic composition
and not c-structure relations brings out and clarifies crosslinguistic commonalities
in principles of semantic composition, commonalities that would otherwise be
obscured by properties of the more variant c-structure.
Even given the centrality of f-structure in semantic composition, however, it
must be kept in mind that semantic composition does not depend upon f-structure
relations alone. For example, as pointed out by Halvorsen (1983) and Fenstad
et al. (1987), prosody has a strong effect in determining semantic interpretation.
Prosodic information is represented at prosodic structure, a structure that is
related to but separate from the c-structure; the modeling of prosodic influence
on semantic interpretation is explored in Chapter 11. Information structure,
described in Chapter 10, is also closely connected to semantic interpretation.
In the rest of this chapter, we will not examine constraints on meaning assembly
imposed at nonsyntactic levels of representation, leaving discussion of these issues
for Chapter 10 and Chapter 11.
This assumes a semantic form like the one in (3) for the verb give:
This semantic form specifies that the predicate give has three arguments with roles
agent, theme, and goal; that the agent is mapped to subj, the theme is mapped
to obj, and the goal is mapped to oblgoal ; that the f-structure for a sentence with
this verb must contain a subj, an obj, and an oblgoal in order for Completeness
and Coherence conditions to be met; and that this use of the verb give is distinct
from other uses of the same verb, since each use of a semantic form is uniquely
indexed (§5.2.2.1).
More elaborated theories of several of these aspects of semantic forms have
emerged in the years since Kaplan and Bresnan’s original work. Most obviously, the
semantic structure and meaning composition 279
mapping of grammatical functions to semantic roles has been the focus of much
theoretical attention; this is discussed in detail in Chapter 9. Further, the semantic
form is no longer assumed to represent semantic relations. Instead, the semantic
contribution of a verb like give is reflected in the semantic structure and its relation
to meaning (to be described in this chapter), as well as in argument structure
(Chapter 9). This separation leads to a more modular theory, since on this view
f-structure is a purely syntactic level of representation, not a mix of syntactic and
semantic information (on modularity, see §7.2). In addition, a more adequate view
of meaning and its relation to syntax is thereby available: the original view of the
semantic form was inadequate to represent anything but the most basic semantic
relations. Semantic forms could not represent many aspects of interpretation,
including scope of modifiers, quantification, and notions of coreference.
What, then, is the role of the semantic form value of the pred feature in the
current setting? First, the function of instantiation to indicate distinctness remains.
Many words make a syntactically unique contribution, and the fact that semantic
forms are instantiated uniquely for each instance of their use is a formal reflection
of this.
Second, semantic forms represent the array of syntactic arguments that a pred-
icate requires, making explicit the result of the application of mapping principles
to the argument structure of a predicate. As we discuss in Chapter 9, syntactic and
semantic argument structure are not the same; verbs like intransitive eat and rain
illustrate this point:
Although eat denotes a two-place relation between an eater and an eaten thing,
syntactically it has an intransitive use, as shown in (4a); conversely, rain does
not take a semantic argument, but is syntactically monovalent. Semantic forms
represent the syntactic grammatical functions required by a predicate, whether
or not they make a semantic contribution. Andrews (2008) provides further
discussion of the role of semantic forms in LFG.
(5) David
Formally, the expression in (6) indicates that the one-place function yawn is
applied to David —or, to say the same thing in a different way, the predicate yawn
holds of David . This expression means that David yawned, but does not represent
many details of the meaning of the sentence, including its tense. Again, since these
details are not immediately relevant to our discussion, we usually omit them. We
discuss how the semantics of tense and aspect can be modeled in §8.10.
Usually, the expression P contains at least one occurrence of the variable x, and we
say that these occurrences are bound by the λ lambda operator.
To avoid situations where the same variable name has accidentally been chosen
for two different variables, we might sometimes need to rename the variables that
are bound by a lambda operator. The expressions in (8a) are equivalent, even
though one contains the variable x and the other contains the variable y, and both
represent the function be a person. Similarly, the expressions in (8b) are equivalent,
representing the function admire oneself:
Besides the equivalences that come from variable renaming, there are many other
equivalent ways of writing a function. We generally try to represent a function
in the clearest way possible, which is usually the simplest and shortest way. For
example, in the case of a one-place function like person , the shortest and simplest
1 A more complete discussion of the lambda operator and the lambda calculus is provided by Gamut
(1991b: Chapter 4) and Partee et al. (1993: Chapter 13). Note that this use of λ has nothing to do with
the node labeling function λ, discussed in §5.1.2, nor with the projection function λ assumed in some
argument structure work and mentioned in §9.2.
expressing meanings 283
way is just to write the name of the function person . Alternatively, we can apply the
function person to an argument x that is bound by the λ lambda operator, and we
have constructed a one-place function λx.person (x) that is the same as the function
person . The following two expressions are equivalent:
At times it is clearer to write a function in this way; for example, writing the func-
tion as λx.person (x) shows explicitly that it is a function that takes one argument.
Another way of thinking of a function like λx.person (x) is that it picks out the
set of individuals that are people—that is, the set of individuals x for whom the
expression person (x) is true. The function λx.person (x) is called the characteristic
function of the set of people. We will sometimes refer to sets and their characteristic
functions in our discussions of meaning.
Square brackets around the function expression have been added to make the
groupings in this expression explicit. This expression is equivalent to the expres-
sion that results from replacing all occurrences of x in P with a. For example,
the expression [λx.yawn (x)](David ) is equivalent to the expression yawn (David ),
which is the expression that results from replacing all occurrences of x in yawn (x)
with David :
8.4.1.3 Types We assume that the expressions we are working with are typed. As
shown earlier, we propose the individual constant meaning David for the proper
name David; this meaning has type e (for entity), the type of individuals:
(13) David : e
The expression in (13) indicates that the constant David is of type e. We assume
that there are only two basic types: e is associated with individual-denoting expres-
sions, and t (for truth value) is associated with proposition-denoting expressions,
which have a truth value (that is, which are either true or false). The expression
yawn (David ) is of type t:
284 meaning and semantic composition
Types of other expressions are built up from these basic types. For example, the
type of a one-place relation like yawn is:
The function λx.yawn (x) is of type !e → t", a function from expressions of type e
(represented by x) to expressions of type t (represented by yawn (x)).
The type of a two-place relation like select is:
8.4.1.4 Quantification Since the work of Montague (1973) and Barwise and
Cooper (1981), there has been a great deal of interest in the properties of quan-
tificational words like every and most. Here we present a brief discussion of
quantification; details can be found in Gamut (1991b: Chapter 7), Partee et al.
(1993: Chapter 14), Peters and Westerståhl (2006), and Keenan and Westerståhl
(2011). In §8.8, we discuss how quantifiers are treated in the glue approach adopted
in this work.
The noun phrase everyone contains a quantificational meaning corresponding
to the determiner every. A sentence like Everyone yawned has a meaning that can
be represented in the following way:2
2 In our analysis of quantification, we use pair quantifiers—expressions like the one in (18)—
instead of standard generalized quantifiers (every (person, yawn)). There is a one-to-one correspondence
between the two types of quantifiers, as shown by Dalrymple et al. (1991).
expressing meanings 285
proposition yawn (x). The first proposition corresponds to what is often called the
restriction of the quantifier every , and the second proposition corresponds to the
scope. The type of a quantifier like every is:
This type associates an individual e with a pair of propositions !t, t" that involve
that individual. Different quantifiers place different requirements on this relation.
For example, for every (x, person (x), yawn (x)) to be true, any individual x that is a
person—for whom person (x) is true—must also yawn, satisfying yawn (x). In other
words, every individual that is a person must also be an individual that yawns.
The quantifier most requires that more than half of the individuals x satisfying the
proposition person (x) must also satisfy the proposition yawn (x).
The quantifier no requires that any individual x who satisfies the proposition
person (x) must not satisfy the proposition yawn (x)—that is, there should be no
individuals that are people and that also yawn.
The restriction of a quantifier—its first propositional argument—is fixed, spec-
ified by a noun like everyone as person (x) or by a noun like everything as thing (x),
or given by the meaning of the quantified common noun (people or person in
examples 20–21) and any modifiers it might have. In contrast, the scope of a
quantifier—its second propositional argument—is chosen more freely, and seman-
tic ambiguity can result from different scope choices. For example, although (22)
is syntactically unambiguous, with a single c-structure and f-structure, it has two
different meanings, depending on the order in which the quantifiers are applied.
That is, the sentence is ambiguous because there is more than one way of putting
its component meanings together:
In Reading 1, we say that everyone has wide scope, and something has narrow scope.
The scope relations are reversed in Reading 2, where something has wide scope and
everything has narrow scope.
286 meaning and semantic composition
According to Reading 1, the proposition a (x, person (x), yawn (x)) seems to hold,
but the existence of the person (and therefore the occurrence of the yawn) is not
certain—perhaps the only evidence was a noise. In contrast, Reading 2 claims that
there is some individual x that satisfies the proposition person (x) and that also
seemed to yawn, satisfying the proposition seem (yawn (x)).
Such examples show that it is not adequate to rely solely on the f-structure as a
representation of the meaning of a sentence; examples (22) and (23) have a single
f-structure corresponding to more than one meaning. Our theory of semantics and
the syntax-semantics interface allows us to deduce exactly the two meanings given
for (22) and (23), given an unambiguous syntactic structure.
This concludes our brief introduction to predicate logic. We have seen that pred-
icate logic provides a basic yet expressive way of representing linguistic meaning.
This is an advantage from our perspective, since much of our discussion focuses
on issues in meaning assembly, and our claims about the meanings of particular
constructions are fairly general.
that meaning composition need not proceed according to the rules of phrasal
composition. This represents a move away from the phrase-structure based rule-
to-rule hypothesis. In addition, the logic used to state constraints on meaning
combination is a resource logic, linear logic, which treats meaning contributions
as resources that are accounted for in the meaning deduction. Thus, the theory
conforms to the desiderata introduced at the end of §8.3. The theory is often
referred to as the glue approach because of the role of linear logic in stating how the
meanings of the parts of an utterance can be “glued together” to form the meaning
of the whole utterance.
We also assume the following simplified phrase structure rule for NP:
(25) NP −→ N
↑=↓
As discussed in Chapter 5, this lexical entry and phrase structure rule give rise to
the syntactic structures in (26):
(26) NP 𝜙
' (
N pred ‘David’
David
We now augment our theory with a semantic structure and its associated meaning.
As described in Chapter 7, a linguistic structure like the semantic structure
is related to other linguistic structures by means of correspondence functions.
Here, the function 𝜎 relates f-structures to semantic structures, and we say that
the semantic structure is a projection of the functional structure, related to the
functional structure via the 𝜎 (sigma) projection. This is shown pictorially in (27),
where d𝜎 is the semantic structure that is related to the f-structure labeled d by the
correspondence function 𝜎, represented as a dashed arrow:
' (
(27) NP d pred ‘David’ d𝜎 [ ]
N 𝜎
𝜙
David
meaning assembly and logical “glue” 289
As noted in §7.1.1, there are two common and equivalent notations for the
correspondence function:
(28) d𝜎 ≡ 𝜎(d)
In the following, we use the more common subscript notation, rather than the
parenthesis notation: that is, we write d𝜎 rather than 𝜎(d) for the semantic structure
corresponding to d via the correspondence function 𝜎. Nothing of substance
depends on this notational choice; using the parenthesis notation would be equally
correct.
We propose the augmented lexical entry in (29) for the proper name David. This
lexical entry differs from the one in (24) in that the expression David : ↑𝜎 has been
added. No additions or changes to the phrase structure rule in (25) are necessary:
When the type of a semantic structure is clear from the context, we often omit it
to reduce notational clutter.
For brevity, we can use a label like [David] to refer to the meaning constructor
in (30). In (31), [David] is defined as a label representing the typed meaning
constructor David : d𝜎 !e" , in which d𝜎 is a semantic structure of type e and David is
an individual constant representing the individual named David:
290 meaning and semantic composition
Using names or labels for meaning constructors is useful for presenting deductions
in a more compact form.
8.5.2.1 Example one: Intransitive verbs The syntactic structures for the sen-
tence David yawned, together with the semantic result we desire, are:
The semantic structure for the sentence is related to the f-structure for the sentence
by the correspondence function 𝜎, represented as a dashed arrow. We are not
concerned with the internal makeup of the semantic structure here, and so we have
represented it without internal attributes or values, as the structure [ ]. Below, we
will see cases in which the semantic structure has attributes and values which play
a crucial role in meaning deduction.
Let us see how the meaning yawn (David ) for the sentence David yawned is
obtained. We propose the following simplified lexical entry for the verb yawned:
The meaning constructor for yawned pairs the meaning for yawned, the one-place
predicate λx.yawn (x), with the linear logic formula (↑ subj)𝜎 ⊸ ↑𝜎 . This formula
meaning assembly and logical “glue” 291
NP I(
(↑ subj) = ↓ ↑=↓
N VP
↑=↓ ↑=↓
David
V
(↑ pred) = ‘David’
↑=↓
David : ↑𝜎
yawned
(↑ pred) = ‘yawn!subj"’
λx.yawn (x) : (↑ subj)𝜎 ⊸ ↑𝜎
292 meaning and semantic composition
NP
I
(y d
N VP ’
y
d
David
(d V
David : d
yawned
) ⊸y
(y ’
λx.yawn (x) : (y
The f-structure for yawn in (35) is labeled y, and the f-structure d for David is
y’s subj. Since (y subj) = d, we can replace the expression (y subj)𝜎 by d𝜎 in the
meaning constructors in (35), yielding the instantiated meaning constructors for
David (labeled [David]) and yawned (labeled [yawn]):
The meaning (left-hand) sides of the meaning constructors in (36) are familiar
from our discussion of predicate logic formulas in §8.4.1. The meaning side of
the meaning constructor labeled [David] is the proper noun meaning David ,
and the meaning side of the meaning constructor labeled [yawn] is the meaning
of the intransitive verb yawned, the one-place predicate λx.yawn (x).
The glue (right-hand) sides of these meaning constructors indicate how these
meanings are associated with the different parts of this sentence. The constant
David is associated with the semantic structure d𝜎 . The glue side of the meaning
constructor labeled [yawn] is more complex: as explained earlier, the connective
⊸ is the linear implication symbol of linear logic. In (36), we can think of this
expression as encoding a meaning like if d𝜎 , then y𝜎 . In other words, the glue side
of the meaning constructor labeled [yawn] in (36) states that if we consume a
resource associated with the semantic structure d𝜎 , then we can produce a resource
associated with the semantic structure y𝜎 .
We must also provide rules for how the glue side of each of the meaning
constructors in (36) relates to the meaning side in a meaning deduction. For
simple, nonimplicational meaning constructors like [David] in (36), the mean-
ing on the left-hand side (David ) is the meaning of the semantic structure on
the right-hand side (d𝜎 ). For implicational meaning constructors like [yawn],
which contain the linear implication operator ⊸, performing an inferential step
meaning assembly and logical “glue” 293
(37) x : f𝜎
P : f𝜎 ⊸ g𝜎
P(x) : g𝜎
By using the rule in (37) and the meaning constructors [David] and [yawned], we
have deduced the meaning yawn (David ) for the sentence David yawned, as desired.
8.5.2.2 Example two: Transitive verbs Our next example of meaning deduction
involves a transitive verb; the example differs from the one just presented only in
that the verb takes two arguments instead of one. The c-structure and f-structure
for the sentence David selected Chris are:
3 This is the standard correspondence as defined by the Curry-Howard Isomorphism relating propo-
sitions like d𝜎 ⊸ y𝜎 to terms like λx.yawn(x); see Crouch and van Genabith (2000) for more discussion.
294 meaning and semantic composition
NP I
N VP ’
David V
V NP
selected N
Chris
In the meaning constructor for the transitive verb selected, two arguments are
required: a resource for the subj, (↑ subj)𝜎 , and a resource for the obj, (↑ obj)𝜎 .
The square brackets in this expression are there to make the groupings in the
expression clear: selected requires a meaning for its subj, then a meaning for
its obj, to form a meaning for the sentence. In other words, this formula can
be paraphrased as: “If we find a resource for the subject and a resource for the
object, we can produce a resource for the entire sentence.” The meaning side is a
function that requires two arguments and is applied to those arguments to produce
a meaning for the sentence.
In (40), the glue side of the meaning constructor requires the verb to combine
with its arguments in a particular order—the subj first, then the obj—since this
order must respect the order of combination of meanings specified in the lambda
expression on the meaning side. Importantly, the meaning constructor shown in
(41) is exactly equivalent to the one in (40) except that the order of argument
combination on both the meaning and glue sides is reversed, so that the verb
combines with its obj first and then its subj:
In formal terms, the glue side of this meaning constructor is logically equivalent to
the glue side of the meaning constructor in (40). In principle, we can choose any
order of combination of premises, with no theoretical significance attached to the
choice we make.
The lexical entry for Chris is analogous to the one for David, providing a
meaning constructor with the meaning Chris :
meaning assembly and logical “glue” 295
With these lexical entries for the words in the sentence, we have the following
structures:
NP I
(↑ ↓ ↑=↓
N VP
↑=↓ ↑=↓
David
V
(↑
↑=↓
David : ↑
V NP
↑=↓ (↑ ↓
selected
N
) ⊸ [(↑ ) ⊸↑ ]
(↑ ’
↑=↓
λx.λy.select (x, y) : (↑
Chris
(↑
Chris : ↑
As desired, we have concluded that the meaning for the sentence David selected
Chris is select (David , Chris ).
In this book, we will present glue derivations in this informal way. Much glue
literature uses a more standard proof format, as in (46), which represents the same
proof as was presented informally in (45):
See Asudeh (2012: Chapter 4) for an accessible overview discussion of glue proofs,
and the papers in Dalrymple (1999) for additional discussion.
We sometimes take advantage of the possibility of referring to a meaning
constructor by a label to present an abbreviated representation of a derivation from
a set of premises. For example, we can abbreviate the derivation outlined in (46) in
the following way:
The final line in (47) represents the derivation of the meaning select (David , Chris )
for the semantic structure s𝜎 from the premises labeled [David], [Chris], and
[select]. It contains a new expression ), sometimes called the turnstile, which
indicates that the conclusion on the right is derivable from the premises on the
left. Thus, the final line in (47) means that the conclusion select (David , Chris ) : s𝜎
is derivable from the premises labeled [David], [Chris], and [select].
In sum, we have used linear logic as a glue language to provide instructions on
how to glue together or assemble meanings. The use of this logical language lets
us express constraints on meaning combination in a formally coherent and flexible
way, taking advantage of the syntactic relations imposed by the f-structure.
In this example, the phrase I met is a relative clause modifier of man. This
information is not associated with either the word I or the word met. Instead,
the interpretation of I met as a relative clause is due to the phrasal configuration
in which it appears. In Chapter 17, we will propose an analysis of the semantics
of relative clauses, and we will see that the phrase structure rule associated with
relative clause formation in English can be annotated with a meaning constructor,
and thus can make a contribution to meaning.
The view that meanings can be attached either to lexical items or to c-structure
configurations accords with the views of Kaplan and Bresnan (1982: fn. 2), but not
with some other proposals. In particular, Halvorsen (1983) proposes that semantic
content is introduced only in the lexicon, by words and not phrase structure rules
(see also Bresnan 1982a). In a very interesting discussion of verbless sentences,
including the topic-comment construction in Vietnamese and nominal sentences
with no copula in Maori (discussed in §5.4.5), Rosén (1996) shows that attempts to
restrict semantic content to appearing only in the lexicon are inadvisable. Phrase
structure configurations can be associated with meaning constructors, and these
constructors can make an essential contribution to meaning deduction.
298 meaning and semantic composition
In classical logic, if a conclusion can be deduced from a set of premises, the same
conclusion can still be deduced if additional premises are added:
In contrast, linear logic does not necessarily allow the same conclusion to be
deduced when additional premises are introduced. Instead, propositions in linear
logic can be thought of as resources, and an economic metaphor is sometimes used.
For instance, we can use the symbol $1 for the proposition that you have a dollar,
$1 ⊸ apple for the linear logic proposition that if you have $1, you can have an
apple, and apple for the proposition that you can have an apple. The following is
valid in linear logic:
4 In this section, we describe only the properties of the linear implication operator ⊸. Proof rules
for our fragment of linear logic are given in the appendix to the chapter (page 321).
the “glue” language: linear logic 299
Just as in the real world, it is not possible to get two apples with $1, or to still have
$1 as well as the apple:
(54) INCORRECT: A ) A, A
We cannot deduce A, A from A.
A resource cannot be duplicated.
INCORRECT: A, B ) A
We cannot deduce A from A, B.
A resource cannot be discarded.
INCORRECT: A, [A ⊸ B] ) A, B
The resource A is consumed by A ⊸ B to conclude B.
A resource is consumed by an implication.
INCORRECT: A, [A ⊸ B] ) A ⊸ B, B
Both A and A⊸B are consumed in concluding B.
A linear implication is also a resource and is consumed in the deduction.
CORRECT: A, [A ⊸ B] ) B
Semantic completeness and coherence are related in a clear way to the syntactic
Completeness and Coherence conditions on f-structures discussed in §2.4.6. This
is as expected, since most syntactic arguments also make a semantic contribu-
tion and thus must be accounted for in a meaning derivation; indeed, our logi-
cally defined semantic completeness and coherence conditions subsume syntactic
Completeness and Coherence in all cases except for pleonastic or semantically
empty (expletive) arguments, which make no semantic contribution and are not
accounted for in a semantic derivation. The following sentence is syntactically and
semantically incomplete:
(55) ⃰Yawned.
The sentence is syntactically incomplete because the verb yawned requires a subj,
and no subject is present; the sentence is semantically incomplete because the
meaning constructor for yawned requires a semantic resource corresponding to its
subject, but none can be found. Example (56) is both syntactically and semantically
incoherent:
(57) ⃰Rained.
The verb rained requires a subj, but there is no subj in (57); therefore, the
sentence is syntactically incomplete. The semantic completeness condition is not
violated, however, because the subj of rained is not required to make a semantic
contribution—an expletive subject is syntactically required, but not semantically
necessary.5
Another difference between syntactic and semantic coherence involves mod-
ifying adjuncts: a semantic deduction in which the meaning contribution of a
modifier is not incorporated is semantically incoherent, since all meanings must
be taken into account. That is, the semantic coherence condition prevents us from
assigning an unmodified meaning to a sentence with a modifier:
5 This is indicated by the appearance of the subj argument outside the angled brackets in the
semantic form for rain, as shown in (102) of §2.4.6.1.
the “glue” language: linear logic 301
The modifier quickly is not constrained by the syntactic Completeness and Coher-
ence conditions, which apply only to governable grammatical functions. Seman-
tically, however, its meaning contribution must be taken into account, and the
deduction is semantically incoherent if the modifier meaning does not appear.
8.8 Quantification
Here we briefly outline our theory of quantification and the treatment of gener-
alized quantifiers, since an explicit theory of the syntax and semantics of noun
phrases will be important in subsequent discussion, particularly in relation to
adjectival modification and relative clauses. For a full explication of the theory of
quantification presented in this section, see Dalrymple et al. (1997).
This entry has a number of new features, which we will now explain.
8.8.1.1 Quantifier scope and meaning assembly The glue side of the meaning
constructor in the second line of the lexical entry in (61) has several new aspects,
different from the meaning constructors for proper names and verbs that we have
examined thus far:
(62) ∀H.[↑𝜎 ⊸ H] ⊸ H
First, a universal quantifier ∀ binds the variable H, which ranges over semantic
structures that correspond to possible scopes of the quantifier. The universal
quantifier ∀ means something close to the English word all or every, and it binds
the variable that follows it; see Partee et al. (1993: Chapter 7) for a full explanation.
In (62), the expression [↑𝜎 ⊸ H] ⊸ H is asserted to be valid for any H: if
an implicational resource ↑𝜎 ⊸ H is provided for any H, we can consume the
implicational resource and obtain the resource H.
The second new aspect of the meaning constructor in (61) is that it contains
an embedded implication: the implication ↑𝜎 ⊸ H appears on the left side of
the main linear implication operator. We can think of the expression ↑𝜎 ⊸ H as the
argument required by the meaning constructor for everyone. As we have seen, the
arguments required by a meaning constructor appear on the left side of the main
implication operator. An intransitive verb like yawned requires as its argument the
meaning of its subject, (↑ subj)𝜎 :
8.8.1.2 Quantifier scope meaning The meaning (left-hand) side of the lexical
entry for everyone in (61) is:
In this expression, S(x) represents possible meanings for the scope of the quantifier.
To take a concrete example, we present the c-structure, f-structure, and
meaning constructors for the sentence Everyone yawned in (65); we have also
filled in the target meaning for the sentence, associated with the semantic
structure y𝜎 :
The right-hand side of the meaning constructor labeled [everyone] requires as its
argument a meaning constructor of the form:
(66) e𝜎 ⊸ H
The glue side of the meaning constructor labeled [yawn] is of exactly this form,
and the derivation is successful if the variable H for the scope semantic structure
304 meaning and semantic composition
To determine the meaning that results from combining the meaning constructors
labeled [everyone] and [yawn] according to the glue deduction in (67), we follow
the function application rule presented in (37) on page 293, applying the mean-
ing of the quantificational pronoun λS.every (x, person (x), S(x)) to its argument
λz.yawn (z). The resulting meaning expression is:
or, equivalently:
In sum, assuming the meaning constructors shown in (65) for everyone and
yawned, we can perform the following full glue deduction:
We conclude that the sentence has the meaning every (x, person (x), yawn (x)), as
desired.
quantification 305
(72) every (x, a (y, student (y), relative.of (x, y)), attend (x))
8.8.1.4 Abstraction The rule for function application given in (37) has been
important in the examples we have seen so far. In order to handle sentences with
multiple quantifiers, we also require a rule of abstraction that allows a hypothetical
resource to be proposed in order to create a function. The rule in (74) allows us
to temporarily posit an additional premise in the deduction, a semantic resource
f𝜎 associated with the meaning x. A semantic resource hypothesized in this way
is notationally distinguished from other premises in that it is enclosed in square
brackets: [f𝜎 ]. If we can successfully perform a deduction (represented by elliptical
.
dots .. ) from this and other meaning constructor premises, producing a semantic
resource g𝜎 with meaning P(x) as in (74), we can discharge the assumption x : [f𝜎 ],
and we are left with the meaning constructor λx.P(x) : f𝜎 ⊸ g𝜎 :
(74) x : [f𝜎 ]
..
.
P(x) : g𝜎
λx.P(x) : f𝜎 ⊸ g𝜎
NP I
N VP ’
Everyone V h e
V NP s
heard N
something
[everyone] λS.every (x, person (x), S(x)) : ∀H.[e ⊸ H] ⊸ H
[something] λS.a (y, thing (y), S(y)) : ∀H.[s ⊸ H] ⊸ H
[hear] λz.λw.hear (w, z) : s ⊸ [e ⊸ h ]
quantification 307
The meaning constructors for the quantifiers [everyone] and [something] each
require an implicational meaning constructor of the general form F𝜎 ⊸ H. They
differ in that the meaning constructor [everyone] refers to the subject semantic
structure, e𝜎 , while the meaning constructor [something] refers to the object
semantic structure, s𝜎 .
The meaning constructor [hear] is not exactly of this form, since it requires two
resources to produce a meaning for the clause: a resource corresponding to the
subject meaning e𝜎 , and a resource corresponding to the object meaning s𝜎 . In
fact, we can obtain a meaning constructor of the form that [everyone] requires as
its argument by using the abstraction rule given in (74), supplying a hypothetical
resource v : [s𝜎 ] to [hear]. We can then combine [hear] with [everyone] under
the hypothesized resource, discharge the v : [s𝜎 ] assumption, and combine the
resulting meaning constructor with [something]. The deduction is shown in (76):
Thus, we begin the deduction of the narrow-scope reading of something with the
logically equivalent version of the meaning constructor [hear] in (77b):
This illustrates the possibility of ambiguity resulting from two quantifiers which
take the same semantic structure as their scope, but apply in a different order.
Quantifier scope ambiguity can also arise when a quantifier can choose from
multiple possible scope points, as in (23). In a glue setting, these possibilities
arise naturally, given the different proofs available from the same set of meaning
constructor premises.
IP
NP I
(↑ ↓ ↑ =↓
D N VP
(↑ ↓ ↑ =↓ ↑ =↓ every (x, student (x), yawn (x)) : y [ ]
N V
Every
↑ =↓ ↑ =↓
student yawned
The meaning constructor for every uses inside-out functional uncertainty (§6.1.3)
to refer to the f-structure for the noun phrase that contains it. The expression
(spec ↑) in this entry refers to the f-structure in which the f-structure for every
appears as the spec value, which is the f-structure labeled s in (79).
The lexical entry for the common noun student is:
According to the lexical entries in (80) and (81), the semantic structure for every
student is complex and has internal structure; it contains two attributes, var and
restr, with semantic structures as their values. The attribute var represents a
variable of type e, and the attribute restr represents a restriction of type t on
that variable—in this case, that the variable must range over individuals that are
students.
These lexical entries, together with the standard English phrase structure rules,
give rise to the structures shown in (82); to save space, only the glue sides of the
meaning constructors for every and student are displayed, and the meaning sides
are omitted:
quantification 311
NP
D N
(↑ ↓ ↑ =↓
every
) ⊸ ((
(↑ N
⊸ [∀H.[( ↑) ⊸ H] ⊸ H]
[(( ↑) ↑) )] ↑ =↓
student
) ⊸ (↑
(↑
(↑ )
Instantiating the ↑ and ↓ variables and using the labels v and r for the semantic
structures (s𝜎 var) and (s𝜎 restr), as shown in (82), we have the meaning
constructors in (83), labeled [every] and [student]:
The meaning constructor for [every] requires two arguments: just as a transitive
verb needs two semantic contributions, one from its subject and one from its
object, a quantifier like every needs a semantic contribution from its restriction
(the meaning of the common noun and any arguments or modifiers it might have)
and its scope.
The first requirement is for a meaning for the restriction of the quantifier:
(84) v ⊸ r
This requirement exactly matches the contribution of the common noun student,
and the meaning of student becomes the restriction of the quantifier every .
The second requirement for the quantifier every is a meaning for its scope:
(85) s𝜎 ⊸ H
We can now deduce the meaning constructor for every student from the meaning
constructors for every and student:
The resulting meaning constructor for every student is in crucial respects funda-
mentally the same as the meaning for everyone. This is desirable because in terms
of meaning construction, they behave alike; only the meanings associated with the
semantic structures differ.
Completing the deduction, we have the meaning every (x, student (x), yawn (x))
for this sentence, which is the desired result:
David . The name David, used in a particular context with a particular referent,
may refer to an entity that is animate, for example a human or a dog. In the former
case, the referent has the property of humanness; in the latter case, it does not. Most
of the time, the precise referent of a name like David, and the particular semantic
features associated with that referent, are not important for our purposes. In some
situations, however, we need to be able to refer to a particular semantic feature
associated with a meaning (and/or its referent). Rather than simply referring to
↑𝜎 , for example, we may need to refer to the fact that the value of the attribute
animate associated with ↑𝜎 is positive or negative. Example (88) shows the f-
structure and meaning constructor for David in a context in which it is animate and
human:
There may be many other semantic and pragmatic features associated with a name
like David in different contexts (Liao 2010, as cited by Dalrymple and Nikolaeva
2011: 78–9). As with functional structures, we usually do not explicitly represent
every attribute in the semantic structure, but only those that are relevant to the
current discussion. Semantic features may be contributed by, for example, syntactic
structure, prosody, or discourse context. Semantic features will prove to be of
particular importance to our discussion of information structure (Chapter 10).
Another semantic feature which will be relevant in Chapter 9 is rel. This feature
is often used as a label representing the meaning of the eventuality denoted by a
verb (i.e. the meaning of a verb minus any tense, aspect, or other grammatically
contributed meaning). For example:
Note that the value of rel is merely a label that stands for the inherent meaning
of the word concerned, and does not directly correpond either to the f-structure
semantic form or to the meaning side of a meaning constructor. This accounts for
the lack of subcategorization information in the semantic structure in (89).
delve more deeply into the semantics of verbs, to show how it is possible to make
distinctions relating to tense and aspect.
Early work on tense and aspect in LFG represented such information purely in
the f-structure; see, for example, Butt (2001b) and Glasbey (2001). Fry (1999b) was
one of the first to propose a model of event semantics for LFG. Haug (2008b) and
Bary and Haug (2011) present a detailed and relatively comprehensive model of the
semantics of tense and aspect in LFG in the context of participial modification (see
also Lowe 2012). The approach presented here is largely based on the proposals of
Haug (2008b), Bary and Haug (2011), and Lowe (2012), though departing from
their assumptions in certain minor ways.
Expressions such as yawn (David ) and λx.λy.select (x, y) treat the meaning of a
verb as a function that takes as arguments the meanings of the syntactic arguments
of the verb. That is, just as David is a syntactic argument of yawn in the sentence
David yawns, so the meaning David is a semantic argument of the meaning yawn
in yawn (David ). This is a convenient simplification, but it does not enable us to
explicitly distinguish the different semantic roles that syntactic arguments may
have. There is clearly a semantic difference between the roles played by the subject
arguments of the verbs yawn and select, for example: the former is a theme,
while the latter is an agent. Some models of argument structure, discussed in
Chapter 9, assume that semantic roles are also relevant for syntactic subcatego-
rization requirements and alternations. Such models represent semantic roles at
argument structure, but these representations do not imply that such relations are
not part of the semantics; rather, they tend to leave the semantic representation to
one side. Other approaches to argument structure seek to separate semantic roles
from argument structure representations as far as possible. We maintain that it is
desirable to represent the semantic relations between verbs and their arguments as
a part of the meaning representation.
In order to capture the semantic aspects of the relationship between a verb
and its syntactic arguments, we can adopt a “neo-Davidsonian” approach to the
representation of syntactic arguments, treating semantic roles as predicates of
events (Davidson 1967; Dowty 1989; Parsons 1990; Schein 1994). Thus far, our
representation has not distinguished event types from individual events. We now
use yawn to refer to an event type, that is, the set of events that can be characterized
as consisting primarily of yawning. This predicate holds of individual events.
Events are a special kind of individual, and we therefore extend our understanding
of the type e (§8.4.1.3) so that e ranges over both events and entities. The meaning
constructor in (90) represents a function from events of unspecified nature to the
proposition that the event is a yawning-type event:
This meaning constructor is identical in form to that given in (15) on page 284.
It is now used, however, in an entirely different way. The variable x no longer
represents the subject of yawn, as in previous sections, but represents an event that
is characterized as an event of yawning. The simplified expression yawn (David )
can now be expressed in a rather different form as:
tense and aspect 315
This meaning constructor encodes the information that David is the theme of
an event x, which is an event of yawning. This expression therefore represents a
function from events in general to propositions in which David is the theme of
an event of yawning. In the glue expression, (↑𝜎 ev) refers to a semantic structure
labeled ev, which represents the event variable of the proposition:
(92) ’ yawn :
In the same way, we previously used the expression select (David , Chris ) to represent
the meaning of the sentence David selected Chris. Now, the semantic relation
between the verb select and its arguments can be more accurately represented as
in (93), where David is specified as the agent of the selecting event, and Chris as its
theme:
(93) λx.select (x) ∧ agent (x, David ) ∧ theme (x, Chris ) : (↑𝜎 ev) ⊸ ↑𝜎
This meaning expresses that David is the agent, and Chris the theme, of an event
x, which is an event of selecting.
Now that we are able to refer to a particular event (or set of events), we are able
to specify the temporal and aspectual characteristics of that event. In order to do
this, we need a theory of the representation of the semantics of tense and aspect.
Reichenbach (1947) was one of the first to develop a formal model of tense and
aspect relations by reference to a limited number of temporal variables. This model
has been further developed by Kamp and Reyle (1993), and is highly influential;
a very similar formal approach is that of Klein (1992, 1994, 1995). As discussed
by Kamp and Reyle (1993), tense and aspect can be modeled with reference to the
following four points in time relative to the discourse:6
(94) E – event time, i.e. the time during which or at which an event occurs.
R – reference time, i.e. the time referred to by the utterance.
P – perspective time, i.e. the “now” point of temporal deixis.
S – speech time, i.e. the moment of utterance.
In most simple sentences, S and P are equivalent, and even in complex sentences,
the relation of P to S can be determined purely by reference to E, R, and P; hence
S is generally omitted from the formalization of tense and aspect, and we will
do so here. Temporal and aspectual relations are indicated using logical symbols
indicating precedence, inclusion, and overlap. In line with Klein (1992), tense
is formalized as the relation between R and P, while aspect is formalized as the
6 These points may be true points, that is instants with no measurable temporal extent, or they may
be periods extending over an interval of time.
316 meaning and semantic composition
Tense Aspect
Name Present Past Future Imperfective Perfective Anterior
Repr. P⊆R R≺P P≺R R⊆E E⊆R E≺R
relation between E and R. Table 8.1 illustrates the major temporal and aspectual
relations, and their logical representation.7
The symbols ≺ and . indicate a precedence relation; thus, R ≺ P (or, equiva-
lently, P . R) indicates that the time referred to by the utterance entirely precedes
the perspective time of the utterance. This is the basic meaning of relative past time.
The symbols ⊆ and ⊇ indicate an inclusion relation: E ⊆ R or R ⊇ E indicates that
E is included within or coextensive with R. This is a possible definition of perfective
aspect within this framework.8
In order to specify the tense and aspect properties of an event, we therefore need
to refer to the event time E of that event. We use the function τ (tau) to do this:
τ(x) is the temporal interval (which may or may not be a point) during which an
event x occurs (Krifka 1989). Aspect describes the relation between τ(x) and the
reference time R of an utterance. Let us take as a concrete example the sentence
David yawned. We assume for the sake of argument that the English simple past
expresses perfective aspect and past tense. The meaning of perfective aspect is
encoded in the meaning constructor in (95), labeled [perf]:
The term (↑𝜎 rt) in the glue expression refers to the value of the rt (Reference
Time) attribute in the semantic structure ↑𝜎 for the verb, representing its reference
time. On the meaning side, an existential quantifier is introduced to bind the event
variable x, and an inclusion relation (⊆) is specified between τ(x), the time of the
event x, and the variable r, which represents the reference time.
The meaning of the verb yawn in combination with its subject argument
David is given by the meaning constructor in (91), repeated as (96) and labeled
[David-yawned]:
7 The formalization of tense and aspect presented here may be too simplistic to account adequately
for the wide range of temporal and aspectual features expressed in different ways crosslinguistically. The
variety of use of the English perfect, for example, cannot be accounted for under this system without
assuming that it has two recursively applied aspect properties. Nevertheless, the system presented here
has relatively wide currency, and is adequate for the demonstration of tense-aspect semantics provided
here.
8 The alternative is proper inclusion: E ⊂ R, meaning that E is included within R but R is not included
within E. Following Kiparsky (1998) and Haug (2008b: 294, fn. 1), among others, we assume improper
inclusion here.
tense and aspect 317
(97) [David-yawned-perf]
λr.∃x.yawn (x) ∧ theme (x, David ) ∧ τ(x) ⊆ r : (↑𝜎 rt) ⊸ ↑𝜎
The meaning of tense is formally similar to the meaning of aspect. Tense quantifies
the reference time R of an event (rt) in relation to a perspective time P (pt). The
meaning of past tense, for example, can be represented as in (98); we give this
meaning constructor the label [past]:
We can combine the meaning constructor [past] with the meaning construc-
tor [David-yawned-perf] to produce the meaning constructor in (99), labeled
[David-yawned-perf-past]:
(99) [David-yawned-perf-past]
λp.∃r.∃x.yawn (x) ∧ theme (x, David ) ∧ τ(x) ⊆ r ∧ r ≺ p : (↑𝜎 pt) ⊸ ↑𝜎
Let us now summarize the derivation of the more complex, but more accurate,
meaning which we have obtained for the sentence David yawned:
In this way, then, it is possible to represent the semantics of tense and aspect using
the glue framework which we have presented in this chapter. The representation
is highly complex, however, even for a sentence as simple as David yawned. In
subsequent chapters, we will ignore the semantic contributions of tense and aspect
by finite verbs and other words, so as not to obscure the points made.
(Asudeh et al. 2008, 2013), nonrestrictive relative clauses (Arnold and Sadler
2010, 2011), adjectival constructions (Al Sharifi and Sadler 2009), the semantics of
concomitance (Haug 2009), affected experiencer constructions (Arnold and Sadler
2012), optional or ambiguous arguments (Asudeh and Giorgolo 2012; Giorgolo
and Asudeh 2012b), and distance distributivity (Przepiórkowski 2014a,b, 2015).
The logic of pronominal resumption is treated in detail by Asudeh (2012), and will
be discussed further in §17.6.4.
Emendations or extensions to the glue system which we have presented here
have been proposed by some authors, but we omit discussion of these issues.
Kokkonidis (2007) discusses the development of glue theory and proposes a “first-
order” glue system that differs somewhat from the “second-order” system used in
this book, as well as much other glue work; Bary and Haug (2011) adopt and extend
Kokkonidis’ proposals. Andrews (2007b, 2008, 2011) proposes a purely propo-
sitional glue system, and discusses the formal integration of semantic structure
into the grammatical architecture, proposing an alternative view of the relation
between semantic structure and f-structure. For Andrews, meaning constructors
are not directly associated with lexical entries in the way that we have assumed
in this chapter; rather, they are associated with special lexical entries that link
meanings to f-structures. Lowe (2014b) discusses the role and function of semantic
structures within the LFG architecture, and proposes a reformulation of meaning
constructors so that each lexical meaning is associated, via its glue term, with a
single semantic structure.
Giorgolo and Asudeh (2011a) propose a new formal treatment of the meaning
contribution of expressions such as expressives, epithets, appositives, and non-
restrictive relative clauses: such expressions are often taken to motivate a multi-
dimensional semantics, encompassing an “at-issue” dimension and a “side-issue”
dimension (Potts 2005). Giorgolo and Asudeh’s treatment of multi-dimensional
meaning extends the formal glue model to incorporate monads, a construction
from category theory; the model is explored further in their work on the argument-
adjunct distinction and argument linking (Giorgolo and Asudeh 2012b), on con-
ventional implicatures (Giorgolo and Asudeh 2012a), and on opaque contexts
(Giorgolo and Asudeh 2014). Arnold and Sadler (2010, 2011) provide an alterna-
tive treatment of multidimensional meaning which does not incorporate monads,
exploring the possibility of extending the projection architecture to separate at-
issue from side-issue meaning, and also exploring an alternative treatment which
assumes a single semantic projection associated with a meaning pair.
We will not discuss proof methods or algorithms for deduction in linear logic,
since this material is widely available for consultation by those interested in formal
and computational aspects of glue theory. Girard (1987) introduced the notion of
proof nets for proofs in the fragment of linear logic we use; for a lucid description of
the use of proof nets for deduction in the glue approach, see Fry (1999a). Efficient
proof techniques for glue semantic deductions are also explored by Gupta and
Lamping (1998).
further reading and related issues 321
[A]i
.. A A⊸B
⊸ .
B B
A⊸B
[A]i [B]j
A B ..
⊗ .
A⊗B A⊗B C
C
A ∀x.A
∀
∀x.A[x/a] A[a/x]
In this chapter, we explore argument structure and its relation to syntax, particularly
concentrating on its role in determining the grammatical functions of the semantic
arguments of a predicate. We will examine different views of the representation
and content of argument structure, and outline the theory of the relation between
thematic roles and grammatical functions. The first five sections explore issues
relating to the theory of argument structure. Sections 9.6 to 9.9 focus on the
analysis of some important phenomena, and further issues relating to grammatical
functions and argument structure are considered in §9.10.
The verb eat can participate in out-prefixation, indicating that it has an intransitive
use:
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
syntax, semantics, and argument structure 323
Thus, the verb eat can be syntactically monovalent, requiring only a subj argu-
ment, whereas it is semantically bivalent, denoting a relation between two entities.
Conversely, a verb like rain in (1b) requires a syntactic subj argument, but
semantically denotes a predicate that does not take an argument; it does not make
sense to ask ⃰Who/What rained?, or to replace the subj argument of rain by any
other argument:
Here too, syntactic and semantic valence are different: rain requires a syntactic
argument that does not play a semantic role. These simple examples make it
clear that syntactic and semantic argument structures are different and must be
represented separately.
The influence of semantics on syntactic form is shown by the predictable
nature of the syntactic realization of arguments of newly coined verbs. Alsina
(1996: Chapter 1) illustrates this point by considering the nonce word obliquate,
which he defines as meaning “to build or place in an oblique orientation.” As he
notes, the possibilities for syntactic expression of the arguments of this verb are
limited:
on the theory of argument structure and its role in mapping between argument
roles and grammatical functions.1
This expression indicates that give has three argument “slots,” which we refer to
as argument positions. These argument positions are associated with the roles of
agent, theme, and goal; such roles are often referred to as thematic roles. The
three argument positions are also associated with syntactic functions: the agent
with subj, the theme with obj, and the goal with oblgoal . As Kaplan and Bresnan
(1982) note, the angled brackets are supposed to remind us of the parentheses
commonly used in logical expressions, so that the semantic form is thought of as
a kind of logical formula encoding some aspects of the meaning of the sentence as
well as the relation between thematic roles and their grammatical functions.
Since these first steps, researchers in LFG have sought to provide an explicit
view of the content and representation of argument structure and its place in the
overall grammar. Proposals for representing argument structure are often based
on one of two rather different sources: Zaenen (1993), Alsina (1996), Rákosi
(2008), Ackerman and Moore (2013), and others adopt and modify the Proto-
Role argument classification proposals of Dowty (1991), while Butt (1995, 1998),
Broadwell (1998), and others present analyses based on the Conceptual Semantics
framework of Jackendoff (1983, 1990).
1 In this book, we will not provide much discussion of the interaction of argument structure with
other grammatical processes. See Manning (1996a,b) for a very interesting theory of the syntactic
role of argument structure and its interactions with functional structure. Manning proposes that what
he calls construal processes such as binding, determination of imperative addressee, and control in
adverbial and complement clauses are sensitive to relations at argument structure rather than a syntactic
representation like functional structure. For a critical evaluation of Manning’s proposals, see Falk
(2006), particularly §7.4.
content and representation of argument structure 325
Alsina (1993, 1996) assumes that argument structure is represented via a refine-
ment of the traditional notion of semantic form, appearing in the f-structure as
the value of the pred feature. In line with much research on argument structure
and thematic roles, Alsina relies on an ordering of the arguments of a predicate
according to a semantically motivated hierarchy of thematic roles (to be described
in §9.4.2): for example, arguments bearing the agent role are higher on the
hierarchy than patients. Alsina claims, however, that the differences among
particular thematic roles like agent and patient are best represented semantically
rather than at argument structure; in his theory, such distinctions do not play a
direct role in determining the syntactic functions of arguments.
Besides the hierarchical ordering imposed on the arguments of a predicate,
Alsina builds on work by Dowty (1991) on Proto-Role argument classification,
distinguishing between proto-agent and proto-patient properties of arguments.
Unlike Dowty, however, Alsina assumes a set of criterial definitions for proto-agent
and proto-patient arguments; for example, he proposes that a causer argument with
volitional involvement is necessarily classified as a proto-agent, while an “incre-
mental theme” argument is a proto-patient. Arguments that do not meet these
criteria are not assigned proto-role status. Alsina (1996) provides the following
argument structure representations for the verbs come and give:
(7) Argument
! structures for come
" and give (Alsina 1996):
pred ‘come![P-P] [ ]"’
! "
pred ‘give![P-A] [(P-P)] [P-P]"’
According to these representations, the verb come has two arguments: the individ-
ual who comes is the proto-patient (P-P), and the destination argument bears no
proto-role status and so is not labeled as either P-A or P-P. The verb give has three
arguments: a proto-agent (P-A) or giver; a recipient argument that is optionally
classified as a proto-patient (indicated by parentheses around the P-P label),
depending on whether or not it is causally affected; and a second proto-patient,
the entity that is given. On Alsina’s view of argument structure, no other semantic
information appears in argument structure besides the proto-role status of each
argument and the ordering of arguments according to the thematic hierarchy.
Zaenen (1993) also builds on Dowty’s proposals to define argument structure in
terms of semantically based properties of a predicate. Unlike Dowty, Zaenen does
not assume that these properties relate to semantic entailments of particular uses
of the predicate. Instead, Zaenen proposes that predicates have lexically specified,
semantically definable characteristics that she terms dimensions; for instance, a
predicate has a volitional dimension if it can be used with an adverb like on purpose.
The existence of a volitional dimension in the argument structure of a verb does
not entail that every use of the verb denotes a volitional act; rather, the verb denotes
an act that can be volitional. Based on these semantic dimensions, Zaenen (1993)
proposes a set of role-defining properties that the arguments of a verb can bear,
326 argument structure and mapping theory
This argument structure representation is tripartite. The first line is the Thematic
Tier, representing causation, motion, and location: here, that the “cause” relation
CS holds between an actor α and an event GOPoss in which some entity comes
to be possessed by a beneficiary β. The second line is the Action Tier, which
encodes information about agency and affectedness. The function aff (affect)
in the Action Tier indicates actor/patient/beneficiary relations; aff+ is a relation
between an actor and a beneficiary (aff− would indicate that the arguments are
an actor and a patient). In this example, the second line indicates that α is the first
argument of the aff+ predicate and is therefore the agent. As Butt points out,
the Action Tier represents roughly the same subset of information that is taken
to be relevant for the Proto-Role theory of Dowty (1991). An innovation in Butt’s
approach is the postulation of an Aspectual Tier, represented on the third line,
which indicates whether the beginning, middle, and end of the event are lexically
specified; in the case of the verb give, neither the inception, duration, nor end point
of the event is intrinsically specified. Compare this with the Hindi-Urdu verb bh uul
‘forget’ shown in (9), which is negatively specified (with the integer 0) for inception
(Butt 1995: 151). Butt shows that aspectual information is important in character-
izing the mapping relation for complex predicates in Urdu:
Butt et al. (1997) assume a more traditional view of argument structure in which
the arguments of a predicate bear particular thematic roles like agent, goal, and
theme. Unlike many other approaches, they provide an explicit characterization
of the relation between argument structure and other grammatical structures:
content and representation of argument structure 327
(10) C-structure, argument structure, and f-structure for cut (Butt et al. 1997):
λ
α
V
cut
=λ◦α
2 Note that this use of λ is different from the λ projection from c-structure nodes to node labels,
discussed in §5.1.2. It is also unrelated to the λ operator used in semantic representation and introduced
in §8.4.1.1.
328 argument structure and mapping theory
(13) selected V (↑
(↑
(↑
(↑ = (↑ 1)
(↑ = (↑ 2)
2 ) ⊸(↑ 1 ) ⊸(↑ ⊸↑
λy.λx.λe.select (e) ∧ agent (e) = x ∧ theme (e) = y :
(↑
The fourth and fifth lines of the lexical entry in (13) state that the s-structure
of the subj is the value of the s-structure attribute arg1 , and the s-structure of
the obj is the value of the s-structure attribute arg2 . The final two lines contain
the meaning constructor encoding the relevant semantic information: Asudeh and
Giorgolo adopt a neo-Davidsonian event semantics for their meaning language
(§8.10) encoding information about thematic roles, treating them as functions
from events to individuals. At semantic structure, Asudeh and Giorgolo distin-
guish the arguments of a predicate by means of attributes like arg1 and arg2
rather than the names of specific thematic roles such as agent and patient for
the relevant s-structure attributes, since the meaning constructor in the final line
includes an encoding of these roles in the meaning language.
Given the lexical entry in (13), Asudeh and Giorgolo’s f-structure and s-
structure for David selected Chris is:
(14)
1 []
2 []
grammatical function alternations 329
According to Asudeh and Giorgolo’s proposal, then, the relation between gram-
matical functions and semantic arguments is specified directly in the lexical
entries of predicates, and there is no separate level of argument structure. Using
this model, Asudeh and Giorgolo (2012) develop a detailed analysis of optional
arguments, which we describe in §9.10.5.
A number of other theories of the nature and content of argument structure have
also been proposed; Alsina (1996: Chapter 1) and Butt (2006: Chapter 5) provide
a useful overview and summary. Spencer (2013: Chapter 7) provides a detailed
discussion of argument structure as a level of representation distinct from syntax
and semantics, containing only those aspects of the semantic representation that
are relevant to the morphosyntax, and only those aspects of the functional syntax
that are relevant to the semantics.
3 Complex predicates raise a potential challenge for the Principle of Direct Syntactic Encoding;
see §9.9.
330 argument structure and mapping theory
or in abbreviated form:
This lexical rule was thought of as producing a lexical entry for the passive form of
a predicate (with the agent argument realized as oblagent and the theme as subj)
on the basis of the lexical entry for the corresponding active form (with the agent
argument realized as subj and the theme as obj). The fact that relation-changing
lexical rules were formulated as operations on one lexical entry to produce another
is telling of the transformationally oriented way in which these rules tended to
be viewed in the early days of LFG: though the rules were formulated to apply to
structures pairing thematic roles with grammatical functions, thematic roles did
not usually play much of a role in regulating or constraining the rules. Instead,
research focused primarily on the morphological and syntactic characteristics of
grammatical function alternations rather than on a theory of the alignment of
thematic roles with grammatical functions.
Zaenen et al. (1985) provide further discussion of association principles for Ice-
landic, and also give a similar set of association principles for German; see also
Rappaport (1983).
From this beginning, a vibrant body of research has developed. We only have
room here to scratch the surface of the complex issues that are involved.
(20) −r: semantically unrestricted functions subj and obj; arguments with any
thematic role (or with no thematic role) can fill these functions (see §2.1.4).
+r: semantically restricted functions obj𝜃 and obl𝜃 : only arguments bear-
ing particular thematic roles can fill these functions. For example, only an
argument with the role of agent can appear as an oblagent .
−o: nonobjective (non-object-like) functions subj and obl𝜃 .
+o: objective functions obj and obj𝜃 .
(21)
−r +r
−o subj obl𝜃
+o obj obj𝜃
[−o]
This rule states that the agent argument must be a −o function: that is, the agent
may not be associated with an objective or object-like function. Therefore, the
agent role must be associated either with the subj or the oblagent function (one
of the obl𝜃 family of functions).
4 Some authors, such as Alsina (1996), have developed approaches which diverge in more significant
ways from Bresnan and Kanerva’s original formulation; see §9.11.
argument classification 333
Turning next to the theme or patient thematic role, Bresnan and Kanerva
propose the following classification, reminiscent of the theme-linking rule in (19):
[−r]
[−o]
The thematic role of a predicate that is highest on the thematic hierarchy, also
ˆ By default, it is classified
called the logical subject, is notationally represented as 𝜃.
as unrestricted:
[−r]
[+r]
These default classifications apply in every instance except when a conflict with
intrinsic specifications would result.
This principle rules out a situation where the same grammatical function is
associated with more than one thematic role, or where the same role is associated
with more than one grammatical function.5
5 The Function-Argument Biuniqueness principle has been challenged by Alsina (1995, 1996) in his
analysis of reflexive and reciprocal constructions, where a single grammatical argument appears to fill
two distinct thematic roles in relation to the predicate. Alsina’s proposals are adopted by Miliĉević
(2009). Reinhart and Siloni (2005) and Rákosi (2008) present alternative analyses of reflexive and
reciprocal constructions which preserve the principle of Function-Argument Biuniqueness (see also
Aranovich 2013). See Asudeh (2012: Chapter 5) for a discussion of Function-Argument Biuniqueness
and its relation to the Theta Criterion (Chomsky 1981) and other theoretical principles governing the
relation between thematic roles and grammatical functions. Argument Fusion in complex predicate
formation also raises potential issues for Function-Argument Biuniqueness; see §9.9.
argument classification 335
[−r]
secondary patient-like roles: 𝜃
[+o]
other thematic roles: 𝜃
[−o]
Secondary patient-like roles include, for example, the second object argument in a
ditransitive construction, though Bresnan et al. observe that the inventory of roles
belonging to this category seems to be subject to crosslinguistic variation.
Most theories assuming intrinsic argument classifications analyze the agent
argument as filling a nonobjective −o role, as shown in (22). However, Kroeger
(1993) argues that in Tagalog, the agent can be syntactically realized as a non-
subject term, or object. Manning (1996b) makes a similar argument for Green-
landic and other ergative languages. Arka (1998, 2003) and Arka and Simpson
6 For an approach which makes a finer distinction between different sorts of patient-like roles, see
Butt (1998).
336 argument structure and mapping theory
(2008) present analyses of the “objective voice” in Balinese which involve agent
objects. Lødrup (1999b) discusses the Norwegian presentational focus construc-
tion, which also allows agent objects. He gives an Optimality-Theoretic analysis
of Norwegian argument mapping that accounts for the differences between Nor-
wegian and languages like English, which do not allow this possibility. Lødrup
(2000) also analyzes Norwegian existential constructions as permitting agent
objects.
subj is the least marked grammatical function, since it has negative values for
both r and o, while obj𝜃 is the most marked. Notably, this markedness hierarchy
generalizes a segment of LFG’s grammatical function hierarchy, given in (5) of
Chapter 2. The recognition of distinct hierarchies for thematic roles and gram-
matical functions is important, as Falk (2006: 39–44) points out: while the same
argument can be the most prominent with respect to both hierarchies, it is also
possible for each hierarchy to have a different most prominent argument, resulting
in a mismatch. Such a mismatch is exactly what is found, for example, in the
passive construction (to be discussed in §9.6), where the argument ranked highest
on the thematic hierarchy does not map to the most prominent argument on the
grammatical function hierarchy.
The Markedness Hierarchy in (30) motivates a set of Mapping Principles for
associating thematic roles with grammatical functions. The Mapping Principles in
(31), from Bresnan et al. (2016: 334), were originally proposed by Bresnan (2001c):
According to rule (31a.i), if the highest thematic role 𝜃ˆ is associated with the feature
[−o], it is mapped onto subj;7 if this rule does not apply, rule (31a.ii) states that any
argument associated with the feature [−r] is mapped to subj. Rule (31b) defines
mapping rules for the remaining thematic roles to grammatical functions other
than subj.
Both the Markedness Hierarchy and the Mapping Principles have been subject
to revision as mapping theory continues to develop. Kibort (2001, 2004, 2006,
2007) argues for a simplified reformulation of the Mapping Principles in (31) into
a single Mapping Principle. (What we refer to as argument positions are termed
ordered arguments by Kibort.)
Kibort argues that this reformulation achieves the correct mappings for various
classes of predicates discussed in the literature, but avoids redundancy by depend-
ing upon the partial ordering of the Markedness Hierarchy. According to this
formulation (and in contrast to Bresnan et al.’s formulation), the Subject Condition
(§2.1.5) is not necessary: this is because the requirement for most predicates to
have a subject is ensured by the Mapping Principle, while impersonal predicates in
many languages do not require subjects.
Her (2013) proposes a revision of both the Markedness Hierarchy and the
Mapping Principles (see also Her 2009). Her proposes that [−r] is less marked
than [−o], with the result that the Markedness Hierarchy is fully ordered:
Her (2013) also proposes a “Unified Mapping Principle” (see also Her 1998, 2003,
2010; Her and Deng 2012) which is very similar to the Mapping Principle of Kibort:
7 Following Zaenen and Engdahl (1994), Bresnan et al. (2016) assume that semantically empty
arguments of verbs like rain or seem also appear in the argument structure and are eligible for linking
to a grammatical function. Semantically empty arguments can appear to the left of the highest thematic
ˆ this accounts for the additional restriction in (31a.i), which applies only when the highest
role 𝜃;
thematic role 𝜃ˆ is initial in the a-structure. If a semantically empty argument appears to the left of
ˆ the linking rule in (31a.i) does not apply.
𝜃,
338 argument structure and mapping theory
On this view, the a-structure for a verb like English select is represented as in (35).
The verb subcategorizes for two arguments, labeled according to their thematic
role. These roles each have an intrinsic classification, as discussed in §9.4.2, and
these classifications, together with default and construction-specific classifications,
constrain the grammatical functions of these roles at f-structure:
In this model, the arguments of a predicate are defined by thematic role, and
argument position plays no role. However, a number of authors including Joshi
(1993), Mohanan (1994: Chapter 3), Alsina (1996: 37), and Ackerman and Moore
(2001: 40–59) have argued for the earlier view, that it is necessary to separate
argument positions from thematic roles.
Kibort (2007) observes that some predicates show variation in the assignment
of thematic roles to grammatical functions, and argues that these patterns are
best analyzed by appeal to argument position rather than thematic role. Verbs
exhibiting the spray/load alternation are examples of such predicates: in terms of
argument positions, both variants are three-place predicates selecting for a subj,
obj, and obl𝜃 , but the two non-agent thematic roles can associate with either the
second or third argument position.8
8 The spray/load alternation is discussed by Ackerman (1990, 1992), Levin (1993: Chapter 2), and
Ackerman and Moore (2001). Similar alternations including verbs such as ‘present’ in Hungarian are
analyzed in mapping theory terms by Ackerman (1992) and Laczkó (2013).
selection and classification: syntactic or semantic? 339
Kibort also observes that morphosyntactic operations can make reference to argu-
ment structure positions independent of thematic roles. For example, passivization
and locative inversion can be described as lexical operations affecting the first
argument position of a verb, but there is no way to generalize over the particular
thematic role which that first argument may have. On the basis of such evidence,
Kibort (2007) concludes that argument positions must be separated from thematic
roles, and represents the former using labels such as arg1 , arg2 , and so on.9
Kibort proposes that a single “valency template” is available to non-derived
predicates in all languages. Following Zaenen (1993), Ackerman and Moore
(2001), and others, Kibort proposes that a fixed total ordering exists within this
valency template:10
Every predicate selects for one or more arguments from this template. The slots
which a particular predicate selects are lexically specified, and it is possible for
one or more of the argument positions to be skipped. Notice that while this
template broadly corresponds to the hierarchy of syntactic functions (subj >
obj > obj𝜃 > obl𝜃 ), it is defined in terms of the features [±o] and [±r]. The
thematic roles entailed by a particular predicate are associated with argument
positions partly on the basis of intrinsic associations between roles and [±o]/[±r]
classifications (such as the association of agentive roles with [−o] in many lan-
guages), but there is no sense in which a particular thematic role is inherently
associated with a particular argument position. Kibort (2007) also recognizes the
influence of a hierarchical ordering of thematic roles but, following Levin and
Rappaport Hovav (2005: Chapter 6), points out that no thematic hierarchy can be
defined that captures all generalizations over the association of argument positions
with thematic roles. Thematic roles may therefore be understood to be ordered,
but this ordering is only default, and different actual orderings are possible: under
certain conditions roles may map to slots in different ways, as for example in
the spray/load alternation shown in (36). We agree with Kibort that a separation
should be maintained between thematic roles and argument positions, and adopt
her approach to argument structure in the following four sections. We examine
some of the central phenomena which have been the focus of attention in mapping
theory, providing an argument structure analysis based on Kibort’s proposals.
9 Note that Kibort’s use of labels like arg1 , arg2 , etc. is not the same as in Asudeh and Giorgolo’s
model, described in §9.2, where arg1 and arg2 represent s-structure attributes whose values are
s-structures. We distinguish between the two typographically by using small capital letters for Asudeh
and Giorgolo’s s-structure attributes, and lower case letters for Kibort’s argument structure labels.
10 Kibort’s analysis of arg1 as either −o or −r is designed to account for the distinction between
unergative and unaccusative verbs; see §9.10.1.
340 argument structure and mapping theory
The intrinsic meaning of the verb select involves a selector and a selectee, or
an agent and theme argument. As a canonical transitive verb in English, its
subcategorization frame includes arg1 and arg2 , the first two argument positions
in the universal valency template presented in (37); the agent is associated with
arg1 , and the theme is associated with arg2 .
These argument positions are intrinsically associated with mapping feature speci-
fications:
We now apply the Mapping Principle (32). The highest, least marked function
on the Markedness Hierarchy is subj. Since arg1 is specified as [−o], assignment
to subj is compatible with its specifications, and it is mapped to subj. Of the
remaining core functions, only obj is compatible with the [−r] specification of
arg2 , so arg2 is associated with obj. We therefore have the following associations,
where M.P. stands for Mapping Principle:
Thus, the theory correctly captures the fact that in the active voice, the agent
argument of select is the subj, and the patient is the obj:
We now examine the passive counterpart of this active sentence. Most work
within LFG, following Bresnan and Moshi (1990), assumes an account of the
passive in which the thematically highest argument is suppressed and, therefore,
unavailable for linking. For example, under this approach, the passive of select has
this argument structure:
Under these assumptions, only the second argument slot is available for linking to
a grammatical function. Since [−r] is compatible with the highest function on the
Markedness Hierarchy, arg2 maps to subj. This can be thought of as a promotion
of arg2 from obj, as in (41), to subj:
∅
M.P. subj
By contrast, Kibort (2007) argues for a rather different account of the passive.
In her view, morphosyntactic operations such as voice alternations do not affect
the selection of argument positions by a predicate, nor the alignment of argu-
ment positions with thematic roles; passivization and similar morphosyntactic
operations affect only the mapping between grammatical functions and argument
positions. Given the Markedness Hierarchy and the Mapping Principle, there is
only one way in which this mapping can be changed: by means of a “mechanism
of increasing markedness,” augmenting (and thus further restricting) the classi-
fication of argument positions prior to the application of the Mapping Principle.
Only the addition of positive specifications, i.e. [+r] or [+o], counts as increasing
markedness. There are a limited number of logical possibilities for the addition of
further specifications to argument positions: a [−r] position can be additionally
specified as [+o], a [−o] position can be additionally specified as [+r], or a [+o]
position can be additionally specified as [+r]. It is not possible for a [−r] position
to be specified as [+r], or for a [−o] position to be specified as [+o], since an
argument position may have only one value for each of these features; in addition,
there is no argument position prespecified as [+r].
342 argument structure and mapping theory
Under Kibort’s analysis, then, passive voice is the result of adding a [+r]
specification to the first argument position of a predicate pre-specified as [−o].
This means that the first argument position of a passive predicate (arg1 ) can
associate only with the oblique grammatical function obl𝜃 , while the second
argument position (arg2 ), if present, is necessarily associated with the subj func-
tion. We therefore have the following argument structure for select in the passive
voice:
This produces the correct result, namely that the theme argument of passive select
is the subj, as in this f-structure:
11 Within the context of their own approach to argument structure (as described in §9.5 and
discussed further in §9.10.5), Asudeh and Giorgolo (2012) treat the oblique agent as an obl𝜃 , but state
that their model could easily be reformulated to treat it as an adj.
impersonal predication 343
Kibort argues that voice alternations are meaning preserving, while the anti-
causative is not meaning preserving. In the anticausative, an event affecting a
patient/theme is expressed without the cause of the event being specified. Mor-
phosyntactic operations are necessarily meaning preserving; therefore, the anti-
causative alternation cannot be morphosyntactic, but must be accounted for in
lexical terms. Kibort (2012) consequently argues that two distinct lexical entries
must be assumed for verbs that display the anticausative alternation:
In both cases, the verb entails a spiller agent and a spillee patient/theme, but
in the anticausative variant there is no argument position with which the agent
can be associated. If it is expressed, it must be expressed as an adjunct. Thus,
even though the anticausative alternation may initially appear to be very similar
to the passive, Kibort (2012) analyzes the two as being quite different in argument
structure terms.
Active forms of this verb have the following argument structure, with the agent
of the verb realized as its subj:
(52) agent
In the passive, Kibort analyzes the first argument position as being additionally
classified [+r]; it must therefore map to obl𝜃 . Since there is only one argument
position, and this position maps to obl𝜃 , the passive of an intransitive verb does
not have a subj argument, in violation of the Subject Condition:12
(53) agent
This analysis can be contrasted with the analysis of inherently impersonal predi-
cates like English rain under this model:
(55) It rained.
The difference between an inherently impersonal predicate like rain and other
intransitive verbs, such as the one shown in (52), is that rain does not entail any
participants, as (55) shows, i.e. it is not associated with any thematic role.13 Thus,
12 Impersonal constructions are not necessarily uniform in this respect; see, for example, German
impersonal passives as analyzed by Berman (2003), outlined in §2.1.5.
13 For more on intransitive verbs and argument structure, see §9.10.1.
locative inversion 345
arg1 appears in the subcategorization frame for rain, and maps to a grammatical
function, but there is no thematic role with which it is associated.
Other work on impersonal constructions and expletives within the LFG frame-
work includes Berman (2003: Chapters 4 and 8) on expletive subjects in Ger-
man; Maling and Sigurjónsdóttir (2013) on impersonal passive constructions
in Icelandic; Kelling (2006) on personal and impersonal passive constructions
in Spanish; Rubio Vallejo (2011), also on impersonal constructions in Spanish;
Lødrup (2000) on existential constructions in Norwegian; and Kibort and Maling
(2015) on the distinction between active and passive impersonal constructions.
Example (57), with a preposed locative and a focused theme argument, illus-
trates the locative inversion construction:
Constructions involving subject raising in Chicheŵa also provide evidence that the
locative phrase is the subject and the theme phrase is the object in examples like
(57); see Bresnan and Kanerva (1989) for discussion.
Under Kibort’s (2007) analysis, uninverted examples such as (56) can be ana-
lyzed as involving verbs which select for arg1 and arg4 , that is:
15 This explanation for argument linking in locative inversion is distinct from the earlier proposal
of Bresnan and Kanerva (1989), although the intrinsic classifications are the same: on Bresnan and
Kanerva’s analysis, the locative argument is further classified [−r], and so must map to subj, leaving
the theme argument to map to obj.
complex predicates and argument linking 347
The permissive construction consists of two verbs, the main verb (here likhne
‘write’) and the light, auxiliary-like verb de (here dii) ‘let’. Butt shows that at the
level of c-structure, the permissive construction involves either a V* constituent
containing the main and light verbs (62a), or a VP constituent containing the main
verb and the obj argument (62b,c):
Although the main verb/light verb combination need not form a c-structure con-
stituent, and in fact the verbs need not appear adjacent to each other, Butt (1995)
shows that the permissive construction is monoclausal at f-structure, involving
a single complex predicate constructed from the two verbal forms. Evidence for
this comes from verb agreement patterns in Urdu. The Urdu verb agrees with
the nominative argument that is highest on the grammatical function hierarchy;
if none of its arguments are nominative, the verb shows default (third person
masculine singular) agreement. In (61), we see that the light verb dii ‘let’ agrees
with the nominative feminine singular noun cit..thii ‘note’. This shows that cit..thi
‘note’ is a syntactic argument of the main predicate—one which must include
dii—since agreement is possible only between a predicate and one of its syntactic
arguments. Butt provides further evidence from constructions involving control
and anaphora that points to the same conclusion: the verbs likhne ‘write’ and dii
‘let’ combine to form a single syntactic predicate at f-structure, taking a single array
of syntactic arguments.
As Butt’s work makes clear, complex predicate formation—and, therefore, argu-
ment linking—cannot be defined by reference to the argument roles of a single
word or even a single phrase structure constituent; data discussed by Alsina (1996)
point to a similar conclusion. As mentioned above, the theory of argument linking
developed by Bresnan and Kanerva (1989) was originally referred to as lexical
mapping theory, but the work of Butt and Alsina, together with other research,
led to the relabeling of the theory as functional mapping theory or simply mapping
348 argument structure and mapping theory
theory, in acknowledgement of the fact that the theory cannot apply exclusively to
individual words.
In the ‘allow-to-do’ permissive in (63), argument merger between the two parts
of the complex predicate involves what Butt (2014a) calls “Argument Fusion”: an
argument of the lexical predicate likh ‘write’ (here the participial form likhne) is
coindexed with an argument of the light verb predicate de. In this construction, the
highest argument of the lexical predicate is coindexed with the lowest argument of
the light verb predicate.
The argument structure for the verb likh ‘write’ is shown in (65); (66) illustrates
the use of this verb on its own, as the only verb in the clause:
16 In addition to the works cited, complex predicates in Urdu have been discussed by Butt and Geuder
(2001), Butt and Ramchand (2005), Ahmed and Butt (2011), Raza (2011), Ahmed et al. (2012), Butt
et al. (2012), Sulger (2012), and Lowe (2015a).
complex predicates and argument linking 349
The subcategorization frame for the ‘allow-to-do’ permissive verb de ‘let’ (dii in 63)
can be represented as:
Here ‘%pred’ in the argument structure for de indicates that this verb, in its
use as a permissive light verb, is incomplete in terms of its subcategorization
frame: it requires another predicate’s argument structure with which to merge. The
argument structure merger of these predicates involves the coindexation of the
lowest argument of the matrix predicate de ‘let’, in this case arg3 , with the highest
argument of the embedded predicate likh ‘write’, in this case (and usually) arg1 .
This results in the following merged argument structure:
The coindexed arguments are treated as a single argument for the purposes of
mapping theory: they must be linked to a single grammatical function. Mapping
proceeds entirely in accordance with the Mapping Principle, but it must proceed
according to the hierarchy of arguments as denoted by their indices, and not simply
left-to-right. In this example, the arg2 of the merged predicate is ‘to the right’ of the
arg3 , but according to the hierarchy represented by their indices, it is arg2 , and so
must be mapped first. If arg3 were mapped before arg2 , it would be mapped to
the highest available [+o] argument, which would be obj. But with subj already
linked to arg1 , there would then be no [−r] argument available for arg2 to map
to. If arg2 is mapped first, it is mapped to obj, and subsequently arg3 is mapped to
obj𝜃 . Note that the properties of the fused argument (arg3 of ‘let’ fused with arg1
of ‘write’) derive entirely from the arg3 of the matrix predicate, and not from the
arg1 of the embedded predicate: the fused argument maps to obj𝜃 , even though
this is not compatible with the [−o] classification of the embedded arg1 (hence it
is presented in parentheses).
Butt (2014a) contrasts Argument Fusion with the distinct Argument Raising
construction, which is found with the ‘allow-to-happen’ permissive in Urdu. Butt
argues that the ‘allow-to-happen’ construction, shown in (64), involves a different
subcategorization frame for the light verb de ‘let’:
The two distinct subcategorization frames of de ‘let’ are associated with different
kinds of argument merger: the subcategorization frame in (67) is associated
350 argument structure and mapping theory
In this type of argument merger, the argument structure for the embedded predi-
cate, here pak ‘ripen’, is simply inserted into the slot of the light verb’s argument
structure, with no coindexation between (or fusion of) the arguments of the
two predicates. Mapping then proceeds entirely in accordance with the Mapping
Principle; since both arguments have the same index (i.e. both are arg1 ), the order
of mapping proceeds by left-to-right precedence. That is, the leftmost arg1 must
map first, followed by the embedded arg1 .
As shown in detail by Butt et al. (2010), complex predicates in Urdu can be
recursively embedded, and when this happens we find recursive argument merger.
There are many different light verbs in Urdu besides de ‘let’ which form complex
predicates. One of these is le ‘take’; this is an aspectual light verb, which adds
perfectivity, or a sense of completion, to the event described by the predicate:17
The subcategorization frame for the light verb le ‘take’ is shown in (72); the
argument structure for the complex predicate likh lii ‘wrote (completely)’, which
involves Argument Fusion, is shown in (73):
17 The precise semantic contribution of the light verb and how this might be accounted for is
discussed in Butt and Tantos (2004). Butt and Tantos consider the use of Petri Nets (Peterson 1981)
for the representation of lexical meaning in LFG with particular reference to complex predicates.
complex predicates and argument linking 351
As with complex predicates that show only one level of embedding, this predicate is
a true complex predicate: it is monoclausal at f-structure. The argument structure
for this complex predicate, which involves multiple instances of Argument Fusion,
combining (65) with (67) and (72), is:
(75) ‘let’ ! arg1 arg3 ‘take’ ! arg1 ‘write’ ! arg1 arg2 """
[−o] [+o] ([−o]) ([−o]) [−r]
subj obj𝜃 obj
As one would expect, this complex predicate can be further embedded under
other light verbs; Butt et al. (2010) provide detailed discussion of this recursive
embedding.
18 The approach of Andrews and Manning (1999) is LFG-based but architecturally quite different,
involving structures resembling those used in Head-Driven Phrase Structure Grammar much more
closely.
grammatical functions 353
9.10.1 SUBJ
The grammatical function subj is classified as [−o,−r], making it nonobjective
and semantically unrestricted. It is interesting to consider intransitive verbs and
some of the ways in which their subjects may differ, specifically in relation to the
unergative/unaccusative distinction (§9.4.4): the subject of an unergative intran-
sitive verb is agent-like, while the subject of an unaccusative intransitive verb is
patient-like.
Falk (2001b: 111) observes that if a separate level of a-structure is assumed, it
is possible to provide a straightforward account of the unergative/unaccusative
distinction without the need for additional assumptions. Falk illustrates the differ-
ence between unergative and unaccusative verbs in English using the resultative
construction.19 As discussed in §9.4.4, when the verb in a resultative construction
is transitive, the resultative predicate applies to the more patient-like argument and
not the more agent-like argument. This is true in both the active and the passive
voice, whether the more patient-like argument is the obj or the subj:
Kibort (2007) proposes the disjunctive classification [−o] or [−r] for arg1 in her
version of mapping theory in order to account for the unergative/unaccusative
distinction (§9.5). An unaccusative verb requires a [−r] argument, and shares this
classification with the obj of a transitive verb like wipe, whereas an unergative verb
requires a [−o] argument, and shares this classification with the subj of a transitive
verb like wipe. In this way, the grammaticality facts exemplified in (76) and (77) are
captured. Bresnan et al. (2016: Chapter 14) provide further discussion of argument
linking and unaccusativity.
19 On resultative constructions in Norwegian, see Lødrup (2000). Markantonatou and Sadler (1995)
provide an analysis of English resultatives which departs in significant respects from standard assump-
tions of LFG and lexical mapping theory. Bresnan et al. (2016: Chapter 14) provide a brief discussion
of resultatives, unergative intransitive verbs, and “fake” reflexives.
354 argument structure and mapping theory
9.10.2 OBJ
Börjars and Vincent (2008) discuss the grammatical function obj in detail, and
its relation to the more general concept of “object” and the thematic role theme.
They identify many problems in the definition and distinction of obj as a distinct
grammatical category, and argue that current versions of mapping theory which
make use of binary feature decompositions of grammatical functions are inade-
quate when it comes to the specific properties of obj.
Within the LFG framework, the dative alternation is discussed by Allen (2001)
and Kibort (2008); Her (1999) examines the dative alternation in Chinese in the
context of a new set of mapping proposals. Under the approach that we have
adopted, following Kibort (2008), the dative alternation involves a pre-syntactic
realignment of two of the participants in the event. In the construction illustrated
in (78a), including the PP oblique argument which we identify as oblgoal , the
argument position associated with the benefactive/recipient argument is arg4 ;
this is based on the set of semantic entailments invoked by the predicate in relation
to this participant in the event. According to Kibort’s theory, arg4 is intrinsically
associated with [−o]. Kibort proposes that the dative alternation involves the
“remapping” of the benefactive/recipient to primary object argument position
(arg2 ) plus the “downgrading” of the theme to secondary object argument position
(arg3 ), as shown in examples (79) and (80). In the following examples, b represents
the participant with the most proto-beneficiary/recipient properties:
(79) Ditransitive verb with oblique argument benefactive: John gave flowers to
Mary.
x y b
(80) Ditransitive verb with shifted dative benefactive: John gave Mary flowers.
x b y
20 Zaenen and Engdahl (1994) also discuss argument mapping and nonthematic or semantically
empty arguments, which they claim are present at argument structure with the intrinsic feature [−r];
see Bresnan et al. (2016: 332) for additional discussion.
grammatical functions 357
9.10.5.1 The core-oblique distinction Arka (2005) discusses the nature of what
he labels the “core-oblique” distinction in Austronesian languages spoken in
Indonesia, and particularly the obj/obl𝜃 distinction. Arka argues in detail that
there is no absolute distinction between obj and obl𝜃 in some languages. He
proposes a “core index,” based on a set of language-specific syntactic properties
characteristic of core arguments. The index ranges from 1.00 (“highly core”) to
0.00 (“highly oblique”). Arka shows that in Indonesian and Balinese the core index
of core arguments can vary, but it is always over 0.60, while the core index of
typical obliques is between 0.00 and 0.10. However, certain arguments of verbs
in Indonesian have a core index around 0.50. This is the case, for example, with
the stimulus NP argument of Indonesian suka ‘like’, which has a core index of
0.54. The puzzle of how to classify this stimulus argument is also discussed by
Musgrave (2001, 2008), who argues that this argument cannot be classified as
either subj or obj. The best analysis within the inventory of LFG’s grammatical
functions is, Musgrave argues, obj𝜃 . Arka (2005) rejects this proposal on the
basis that the stimulus argument does not show sufficient core properties. Both
Arka (2005) and Musgrave (2008) consider the possibility of adding to LFG’s
inventory of grammatical functions a ‘semi-obj’ function, but they both dismiss
this as a viable option given that such a class would be only negatively defined
and not necessarily coherent. Arka (2005) argues that the best solution may be to
reanalyze the distinctions between grammatical functions so that only subject and
complement are differentiated.
The version of (84b) which lacks the PP contains the verb in (85a), while the version
of (84b) which includes the PP contains the related but distinct verb in (85b). The
optionality of such dependents is therefore not due to any essential difference in
status between them and the subj or obj arguments of the same predicates: when
the PP is present it is as much an argument, and no less obligatory, than the subject
and object arguments. The PP’s optionality is only apparent, due to the existence
of two distinct lexical entries for the governing verb, one of which selects for an
obligatory instrument argument, the other of which does not.
However, Rákosi (2006a,b) shows that a more fine-grained approach to optional
dependents is required in order to capture the full range of variation in the status
of such dependents. In particular, some such optional dependents are, in semantic
terms, entailed by the meaning of the predicate, and so must form part of the lexical
conceptual structure of the verb. Rákosi (2006b) contrasts English assassinate with
a verb like drill.21 Superficially, the instrumental dependent of drill appears to be
just like the instrumental dependent of assassinate, since it is optional:
However, there is a difference between the two: assassinate does not necessarily
entail the existence of an instrument, whereas drill does. Rákosi (2006b) claims
that “one can in principle assassinate the president simply by jumping on him/her,”
while it is not possible to drill something without an instrument. This difference
is reflected grammatically since, as noted by Reinhart (2002), there exists a variant
of verbs like drill in which the instrument functions as the subject, whereas there
exists no such variant for verbs like assassinate:
On the basis of such observations, Rákosi argues that the optional dependents of
verbs like assassinate are intermediate, in descriptive terms, between arguments
21 In fact, Rákosi (2006b) bases his discussion on the verb peel, following Reinhart (2002) in
classifying this verb as entailing the existence of an instrument. However, it is possible to peel an orange
or tangerine without the aid of a peeling instrument; in fact, peel patterns with verbs like eat in allowing
but not requiring the presence of an instrument.
grammatical functions 359
proper and adjuncts proper. In formal terms, however, Rákosi seeks to maintain
the absolute distinction between arguments and adjuncts which is fundamental to
the treatment of grammatical functions in LFG. In order to resolve this paradox,
Rákosi argues for a class of thematic adjuncts (adj𝜃 ), which are adjuncts in syntactic
terms but which bear a thematic relation to the predicate, unlike “pure” adjuncts.22
A particular subset of the constituents which Rákosi (2006a,b) treats as thematic
adjuncts, namely optional to-PPs with verbs like seem, receives a different analysis
from Asudeh and Toivonen (2007, 2012). Asudeh and Toivonen refer to these PP
experiencers as instantiating a semantic role of pgoal (goal of perception). They
argue that such dependents should be treated as “pure” adjuncts, but ones which
bear a semantic relation to the predicate without bearing a relation in syntactic or
argument structure terms.
As mentioned above, the standard LFG analysis of the type of optional depen-
dents that Rákosi (2006a,b) analyzes as thematic adjuncts is, following Bresnan
(1982d), to treat them as arguments. This approach is explored further by Need-
ham and Toivonen (2011), who discuss a variety of optional dependents whose
status is, they argue, intermediate between true argument and true adjunct. They
present a number of argumenthood tests, showing that several types of optional
dependents display inconsistent properties. These include the constituents in
parentheses in the following examples: the demoted subject in a passive (88a),
possessive phrases with event nominals (88b), benefactives (88c), “displaced”
themes (88d), instruments (88e), experiencers (88f), and directionals (88g):
22 For a similar treatment motivated primarily by implementational concerns, see Zaenen and
Crouch (2009).
360 argument structure and mapping theory
relevant to the majority of verbs and so do not generally distinguish between verb
classes.
With this definition of argument in place, Needham and Toivonen (2011) pursue
a mapping-theoretic analysis of optional dependents which is similar in spirit to
the proposals of Bresnan (1982d) in treating them as derived arguments. In the
case of instrumentals, for example, they provide a lexical rule which is roughly
equivalent to Bresnan’s Instrumentalization, exemplified in (85). Reformulated in
terms of the approach that we have been using, following Kibort’s work, the rule
would look like this:
(89) Optionally add the following argument to verbs whose first argument is an
agent:
y
arg4
[−o]
M.P. oblinstr
Similar optional lexical rules are proposed for the other optional dependents
exemplified in (88).
The argument-adjunct distinction continues to be a focus of research within
LFG; see, for example, Rákosi (2013), Toivonen (2013), and Przepiórkowski
(2017). To date, such research has overwhelmingly focused on data from English.
As Needham and Toivonen (2011) observe, a fuller understanding of this issue is
likely to be gained in light of data from a wider range of languages.
Building on the derived argument approach of Needham and Toivonen (2011),
Asudeh and Giorgolo (2012) propose an integrated syntactic and semantic analysis
of the argument-adjunct distinction.23 They argue that the standard LFG approach,
involving distinct lexical entries for verbs that have optional dependents, fails to
capture the intuition that fundamentally the same verb is involved in both of the
following sentences, for example:
Asudeh and Giorgolo (2012) propose to capture such patterns by using optional
f-descriptions within a single lexical entry. In presenting their analysis, Asudeh
and Giorgolo contrast their model specifically with that of Butt et al. (1997),
presented in §9.2, according to which argument structure is interpolated between
c-structure and f-structure. Asudeh and Giorgolo argue that such an approach
could deal with optional dependents only if a further correspondence function
were assumed between a-structure and s-structure. That is because the patient
argument of a verb like eat, for example, must be present at argument structure and
23 For an alternative implementation of the same proposal using monads, building on Giorgolo and
Asudeh (2011a), see Giorgolo and Asudeh (2012b).
grammatical functions 361
semantic structure even if it is not realized in the c-structure and does not appear
in the f-structure. Therefore, alongside the decomposition of the 𝜙 function into
a function α (from c-structure to a-structure) and a function λ (from a-structure
to f-structure), it would also be necessary to assume a function, which Asudeh
and Giorgolo (2012) label 𝜃, from a-structure to s-structure. On this view, 𝜃 is the
composition of λ with 𝜎, as shown in example (91):
(↑ subj) = ↓ ↑=↓
NP I*
↑ =↓
Kim VP
↑ =↓
V*
α ↑ =↓ ↓ ∈ (↑ adj)
V* PP
α ↑ =↓
V at noon
ate 𝜎 ! "
rel eat pred ‘eat’
subj ) * e event ev[ ]
pred ‘Kim’ 𝜎
agent [ ] λ k[ ]
patient [ ] adj { [ “at noon” ] }
tense pst p[ ]
𝜃
𝜃
As discussed in §9.5, Asudeh and Giorgolo (2012) propose an alternative, archi-
tecturally simpler model in which no distinct a-structure and no additional corre-
spondence functions are required. Under this approach, grammatical functions at
f-structure are directly associated with s-structure attributes in the lexical entries
of predicates. Lexical entries also include an obligatory meaning constructor that
represents the essential meaning of a predicate; additional optional meaning con-
structors can introduce or existentially quantify optional arguments. For example,
Asudeh and Giorgolo (2012) provide the following lexical entry for English ate,
modified in accordance with the version of mapping theory proposed by Findlay
(2014, 2016) (see also Asudeh et al. 2014):
362 argument structure and mapping theory
Line 3 of the lexical entry in (92) associates the s-structure of the subj with the
s-structure attribute arg1 . Line 4 ensures that arg2 is mapped appropriately unless
it is unrealized. The rest of the lexical entry comprises meaning constructors.
The first meaning constructor introduces the meaning eat of the verb and its
agent and patient, and is obligatory. The second meaning constructor (enclosed
in parentheses) is optional, and existentially quantifies over the patient argument,
allowing the verb to appear with only the agent expressed. These f-descriptions
license the analyses shown in (93) and (94), which capture the two different uses
of the optionally transitive verb eat:
(93) F-structure and s-structure for David ate cheese, Asudeh and Giorgolo
(2012):
1 []
2 []
(94) F-structure and s-structure for David ate, Asudeh and Giorgolo (2012):
1 []
2 []
Under these proposals, the verb eat in English has two arguments at s-structure,
regardless of whether the object argument is realized at c-structure and f-structure;
it is realized syntactically in (93), but not in (94). Importantly, we are dealing with
only one lexical entry—only one verb eat—whether or not the optional object is
present.
Asudeh and Giorgolo (2012) also provide an analysis for instrumental with-
phrases, based on the derived argument approach of Needham and Toivonen
(2011). They propose that argument addition is best captured by associating the
further reading and related issues 363
relevant lexical information with the entry of the instrumental preposition with,
as in (95), rather than the verb:
This lexical entry not only includes the information that the derived argument is an
instrument, but also requires the subj to be animate, capturing the generalization
that instrumental with-phrases are only possible with “agent verbs” (Reinhart
2002). Example (96) shows the f-structure and s-structure for the sentence David
peeled the apple with the knife (example 86), based on Asudeh and Giorgolo’s (2012)
proposal:
(96)
+
+
Bresnan et al. (2016: Chapter 14) also provide an overview discussion of argument
structure and mapping theory, discussing a range of alternative approaches and
providing pointers to related literature.
The concept of argument structure, and the specific proposals of mapping
theory, have been subject to criticism, for example by Davis and Koenig (2000) and
Levin and Rappaport Hovav (2005); their arguments are addressed and countered
by, among others, Ackerman and Moore (2013) and Spencer (2013: Chapter 7).
As mentioned earlier, a number of alternative proposals have been made regard-
ing the theory of argument mapping and argument classification. Alsina (1996)
argues for a very different feature decomposition of grammatical functions, mak-
ing use of the features +subj, indicating whether or not the argument is “subject-
like,” and +obl, indicating whether the argument is a direct (−obl) or an oblique
argument. Alsina also eliminates the Function-Argument Biuniqueness condition,
arguing that its elimination allows the correct treatment of, among other construc-
tions, reflexive clitics in some Romance languages.
Rákosi (2006a,b) proposes a different feature decomposition of grammatical
functions, based on Reinhart’s (2002) Theta System decomposition of thematic
roles. Rákosi makes use of Reinhart’s binary features +c(ause) and +m(entally
involved). In this system, absence of a feature differs from a minus value for that
feature. For example, the [+c,+m] role corresponds to the thematic role of agent,
the [+c,−m] role corresponds to the thematic role of instrument, and the [+c]
role corresponds to a thematic role cause.
Linking in nominal predicates has been a subject of interest since relatively early
in the theory’s development, and continues to be a focus of research. Iida (1987)
and Saiki (1987) discuss deverbal nominals in Japanese, the realization of gram-
matical functions in nominals, and the role of argument structure; Markantonatou
(1992, 1995) presents an analysis of linking in deverbal nominals in Modern Greek.
Laczkó (1995, 2000) presents a detailed analysis of the syntax of Hungarian noun
phrases, including a theory of linking of nominal arguments. An alternative theory
of linking for nominal arguments is provided by Kelling (2003).
Another proposal for linking nominal arguments is made by Chisarik and Payne
(2003). They argue for a third binary feature [±d] “discourse-related,” alongside
[±r] and [±o]. This feature serves to distinguish two [−o], [−r] functions: subj
and “adnom.” Both of these functions can be governed by nominals, but adnom
cannot be governed by verbs.
A topic of debate in the literature concerns whether nominals can select for
unrestricted [−r] functions, or only restricted [+r] functions. Rappaport (1983)
and Kelling (2003) argue that nominals can govern only restricted functions.
Markantonatou (1995) assumes that nominals can select for both restricted and
unrestricted functions. According to Laczkó (2000), deverbal nominals can select
for both restricted and unrestricted functions, with poss being an unrestricted
function. Laczkó’s research also includes work on event nominalizations (Laczkó
2003, 2013). The selectional possibilities for adjectives are discussed by Vincent
and Börjars (2010a), while subcategorization by both nouns and adjectives is
discussed by Lowe (2013, 2017b). Spencer (2013: Chapter 7) develops a the-
ory of argument structure representation which is primarily designed to model
further reading and related issues 365
1 There are different uses of the term “information structure” in the literature. O’Connor (2006)
uses i(nformation)-structure as a term for relations between multiple levels, including prosody and
semantics, and refers to the level at which information is organized within an utterance as d(iscourse)-
structure. Here, we use the more standard term “information structure” (or i-structure) for this level,
and the term “discourse structure” rather for the relations between successive utterances in a discourse,
following King and Zaenen (2004). We will not have anything detailed to say on discourse structure in
this book.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
structuring information 367
Russian has a high degree of word order flexibility, meaning that constituents
can appear in different orders for the purpose of encoding information structure
distinctions. Another possibility is for a language to use specific constructions
which can be used to “package” information. For instance, in English a cleft
construction can be used as an answer to a constituent question when the clefted
element is the “answer” constituent; in (3) A is a discourse felicitous response, and
A! is infelicitous in the context provided.2 (A! would, on the other hand, be an
acceptable response to the question Who hates beans? while A would not be.)
Of course, syntax is not the only means by which information structure can be
encoded crosslinguistically. Morphology and prosody may also be used to package
2 The symbol # indicates semantic or pragmatic unacceptability, in the same way as the asterisk ⃰
indicates syntactic ill-formedness.
368 information structure
the Prague School tradition (for example Firbas 1964; Sgall 1967; Daneš 1974;
Mathesius 1983; Hajičová et al. 1998) and beyond (for example, Bolinger 1965;
Halliday 1967; Dahl 1969; Kuno 1972; Gundel 1974).3 To give an example, the
answer to the question translated as Who did John call? in the Hungarian dialogue
in (6a), repeated here as (7), can be divided into Topic and Comment as shown:
The categories of topic and focus have been of particular importance in the analysis
of information structure in the LFG literature. Based on much previous research,
including Gundel (1974), Reinhart (1981), and Lambrecht (1994), we understand
topic to be the entity or entities that the proposition is about. We understand
focus to be the informationally unpredictable part of the proposition, that which
is informative or contrary to expectation (Vallduví 1992; Vallduví and Engdahl
1996), or in Lambrechtian terms, the semantic component of a pragmatically
structured proposition whereby the assertion differs from the presupposition
(Lambrecht 1994). For more on the definitions of topic and focus in an LFG setting,
see Dalrymple and Nikolaeva (2011).
One approach to information structure and its primitives which has been
particularly influential within LFG is that of Vallduví (1992). Vallduví proposes
a trinomial articulation, integrating key insights relating to the two bifurcations
Working within the LFG framework, Choi (1999) proposes an analysis of the
relationship between syntax and information structure in German and Korean
which takes Vallduví’s approach as its starting point.4 Choi acknowledges the
importance of the categories which Vallduví proposes, but also points out that
even a trinomial system is not sufficiently fine-grained to capture the data whose
analysis is her main concern. She therefore proposes a feature-based approach to
information structure, in which a term like “topic” is a label for certain values of
a pair of binary-valued features. The advantage of such a system is that it allows
reference to linguistically significant natural classes of roles sharing a common
feature. Choi proposes two such features in her system to encode information
structural distinctions: +new and +prom(inent). Based on these features, Choi
(1999: 92) defines the four discourse functions shown in Table 10.1: topic, tail,
contrastive focus, and completive focus.
+prom −prom
+new contrastive focus completive focus
−new topic tail
4 In fact, Choi (1999) proposes an Optimality Theory (OT)-LFG analysis. For a brief overview of
OT-LFG, see §18.1.3.
372 information structure
sentence (Choi 2001: 21). Under Choi’s classification, topic and contrastive focus
are prominent (+prom), while tail and completive focus are non-prominent
(−prom). The analysis in (11) of the answer sentence in (10) exemplifies Choi’s
feature system:
Choi’s (1999) approach has the significant advantage of capturing, for example,
what topic and tail have in common: that is, what it means for information to be
classed as ground. Under this approach, ground can be defined as that information
which has a negative value for the feature new. Similarly, topic and contrastive
focus form a natural class because they have in common the specification +prom.
This is a crucial advantage over Vallduví’s system for Choi because in the German
and Korean data which she analyzes, topic and contrastive focus behave alike.
Under Vallduví’s “atomic” approach to information structure categories, by con-
trast, the fact that they pattern together like this is surprising: there is no reason to
expect that they should have more in common than any two of the other categories.
At the same time, under Choi’s feature system these two categories are not expected
to be identical because they have different values for the other feature new: topic
is −new, while any kind of focus is +new.
The binary features originally proposed by Choi in her thesis (published as Choi
1999) were adopted by Butt and King (1996b), who also use them to distinguish
four information structure roles, though not the same ones as Choi; compare Table
10.1 and Table 10.2. Butt and King (2000b) provide the following descriptions of
the four discourse functions which they identify as being associated with different
phrase structure positions in Hindi-Urdu:
+prom −prom
+new focus completive information
−new topic background information
the categories of information structure 373
In Butt and King’s system, completive information is, like focus, new to the
addressee, but it differs from focus in terms of its relative significance and hence
is classified −prom. For example, the phrase in the kitchen in (12A) is, by their
classification, completive information:
Bresnan and Mchombo note that in a cleft construction, the clefted constituent
bears both functions: it is the focus in the matrix clause and the topic in the
embedded clause, as shown in (14a). They propose that the same constituent
cannot bear both the topic and focus functions in the same clause, accounting
for the grammaticality judgment assigned to (14b):5
5 Dalrymple (2010) discusses examples from Malayalam that appear to violate this constraint.
representing information structure 375
(15) F-structure for Inna, John claimed that he saw adapted from King (1995:
199):
’
’
In (15), the phrase Inna is displaced from its expected position inside the clausal
complement to the beginning of the sentence. At f-structure, Inna is both the
topic of the clause and the obj of the verb saw. This sentence therefore exhibits
a long-distance syntactic dependency, and is governed by the Extended Coherence
Condition, first proposed by Zaenen (1980) and discussed in detail by Fassi-Fehri
(1988). The Extended Coherence Condition requires the information structural
functions topic and focus to be associated with some syntactic function in the
f-structure; Bresnan and Mchombo (1987) state the condition as follows:
Used in this way, the f-structure features topic and focus are grammaticized dis-
course functions which have a purely syntactic role as overlay functions, consistent
with Bresnan and Mchombo’s original proposal.
However, the use of the f-structure labels topic and focus was extended in
some literature to include not only grammaticized syntactic functions, but also
the information structure roles to which those functions are related. For example,
King (1995) discusses the Russian example (17), in which the object den’gi ‘money’
bears the role of focus in information structure terms but is not involved in a long-
distance dependency and is not syntactically distinguished from other objects.
Since at the time no other way of representing information structure roles like
topic and focus existed, it was perhaps natural to co-opt the related but different
376 information structure
(17) F-structure for prislal muž den’gi according to King (1995: 212):
prislal muž den’gi
sent husband money
‘My husband sent (me) the money.’
’
When topic and focus are used in this way, f-structure no longer represents only
syntactic information; it now includes information structural roles as well. This
undermines the principle of modularity that underpins the grammatical architec-
ture of LFG, discussed in §7.2.6 This approach to information structure representa-
tion was subsequently adopted in other LFG work, predominantly for expediency,
it seems, as its inadequacies have been acknowledged in the literature. For example,
Butt and King (1996b), who discuss information structure and its relation to
word order patterns in Urdu and Turkish, use topic and focus to represent
information structure topic and focus in f-structure, but note that a separate level
of representation for information structure features would be preferable. Likewise,
Choi (1999) represents the information structure features +new and +prom at
f-structure, but expresses similar dissatisfaction.
As noted in §2.1.11, the inclusion of the attributes topic and focus in
f-structure is problematic for other reasons: Alsina (2008) and Asudeh (2004,
2011, 2012) criticize their use in f-structure even for the purely syntactic
representation of displaced constituents in long-distance dependencies. Both
authors point out that the labels topic and focus have some justification owing
to the crosslinguistically widespread connection between displaced phrases in
long-distance dependencies and information structural topic or focus status.
These labels are, in fact, used in a deliberately ambiguous way in such cases
in order to capture a syntactic relation and an information structure relation
simultaneously. However, it is not necessarily the case that a particular type of
long-distance dependency always involves the same information structure role,
and thus collapsing the distinction between the two in this way is undesirable. As
an alternative, Alsina proposes the neutral term op, standing for “operator,” in place
of both topic and focus, as the f-structure attribute for all displaced elements.
6 There is also the question of the other constituents’ information structure status, for example of
muž ‘husband’ and the verb prislal ‘sent’ in (17), which remains unrepresented under such an f-structure
approach.
representing information structure 377
Similarly, Asudeh (2004) proposes the single attribute udf (for “unbounded
dependency function”). These terms permit syntactic analyses of those syntactic
features that are common to all long-distance dependency constructions without
introducing (or apparently introducing) reference to information structure
categories. This is a desirable outcome, since there is no necessary one-to-one
relation between a particular information structure category and a particular
syntactic construction involving a long-distance dependency. For this reason, we
do not use topic and focus as f-structure attributes, but instead reserve these
terms for use solely in relation to information structure. We use the term dis to
represent long-distance dependencies in f-structure in this book. We prefer the use
of dis rather than udf because the term “unbounded dependency” generally refers
to the relation between a filler and a gap, but the feature dis is also used for the
initial constituent in constructions involving left-dislocation and no gap (Beans,
David likes them). We prefer dis to op because the term “operator” is widely used
in semantics with a related but different meaning, and we wish to emphasize the
syntactic rather than semantic nature of dis as a syntactic overlay function.
(18) a. Was it the ex-convict with the red shirt that he was warned to look out for?
b. No, it was an ex-convict with a red tie that he was warned to look out for.
(King 1997: 8, citing Jackendoff 1972: 232)
(19) F-structure for ex-convict with a red tie according to King (1997: 8):
’
In this particular exchange, the (contrastive) focus in (18b) is tie, which is only
part of the value of the obj attribute in the f-structure in (19); there is no
f-structure consisting only of the focused material which can serve as the value of
a focus attribute in the f-structure. To solve this problem, King argues for treating
378 information structure
(20) c-structure
f-structure i-structure
s-structure
semantics
(22) IP −→ XP I!
(ι(ˆ⃰) topic) = ι(⃰) ι(ˆ⃰) = ι(⃰)
(↑ gf) = ↓ ↑=↓
7 This view is somewhat similar to the proposal of Engdahl and Vallduví (1996) in the Head-Driven
Phrase Structure Grammar framework, whereby the features focus and ground are represented
within a distinct structure called info-struct. O’Connor (2006) and Mycock (2006) provide detailed
discussions of these issues.
representing information structure 379
This rule states that in Hindi-Urdu an IP dominates a phrase of any category (XP)
and an I! constituent. The XP has some syntactic role at f-structure (gf), and at the
same time has the information structure role topic at i-structure. As discussed in
§4.2.1, the IP corresponds to the same f-structure and i-structure as its head, I! . This
rule implies the configuration of c-structure, f-structure, and i-structure shown in
(23), with the ι function from c-structure nodes to i-structures represented by a
dashed arrow:
2 3
(23) gf [ ] (f-structure)
𝜙
IP
XP I!
ι 2 3
topic [ ] (i-structure)
Butt and King (2000b) illustrate their proposal by reference to (24).8 Although
projected independently from the c-structure, i-structure functions and
f-structure functions are not entirely dissociated from each other. The structures
in (23) respect the Extended Coherence Condition given in (16), since the
i-structure topic Naadyaa has the grammatical function subj at f-structure, and
the i-structure focus bazaar-mẽ is a member of the adj set at f-structure.
Butt and King (2000b), like King (1997), assume that i-structure is a projection
from c-structure. A similar proposal is made by Mycock (2006), who argues that
information structure is directly related to both c-structure and prosodic structure.
The shared conception to these proposals is that there is a direct connection
between i-structure and c-structure, and a correspondingly indirect connection
between i-structure and f-structure. The primary motivation for this view is
King’s (1997) observation, discussed above, that f-structure constituents often do
not correspond to information structure constituents, since f-structures are often
either too small or too large to define information structure roles. This is labeled
the granularity problem by Dalrymple and Nikolaeva (2011: 92). The analysis of
constituent questions proposed by Mycock (2006) raises a similar issue, insofar
as it requires reference to units that do not match f-structure constituents, and so
cannot be defined in f-structure terms.
The solution to the granularity problem which King (1997) proposed, and which
was adopted by Butt and King (2000b) and Mycock (2006), is that i-structure
is projected directly from c-structure. This overcomes the granularity problem
because c-structure is more fine-grained than f-structure. For example, although
the f-structure of a modifier is contained within the f-structure of the head, the
c-structure nodes of the modifier may map to a different i-structure from that of
the head. A direct projection from c-structure to i-structure therefore appears to
permit more fine-grained distinctions.
IP
NP I
N S
Naadyaa AdvP NP VP
Nadya
abhii N PP V
now ι
.t bazaar-mẽ xarid rahii thii
buy
10.4.1 Overview
Dalrymple and Nikolaeva (2011) adopt King’s (1997) view, shared by Butt and
King (2000b) and many other LFG researchers, that information structure should
be treated as a separate level of representation, independent from f-structure.
Dalrymple and Nikolaeva propose that information structure contains four
attributes, topic, focus, background, and completive, representing the four
information structure roles originally proposed by Butt and King (1996b), as
discussed in §10.2 above. These information structure roles are determined
by discourse context, and, as discussed in §10.1, can be signaled by various
different means, such as word order and phrase-structure position.9 An important
feature of Dalrymple and Nikolaeva’s proposal is that the elements categorized at
information structure, on the basis of these four features, are meaning constructors
(see Chapter 8), following Mycock’s (2009) insight that information structure and
s-structure are closely related.
9 More discussion of these issues is provided by Mycock (2006), Erteschik-Shir (2007), Féry and
Krifka (2008), and references cited in those works.
382 information structure
In (25), the structured meaning is a triple: the first member represents the focus
Bill , the second member contains a set of alternatives to the focus A , and the third
member contains the background meaning ‘introduced to Sue’. The set of alter-
natives A represents all the relevant individuals that might have been introduced
to Sue, including Bill. For example, in the context under consideration A might be
{Bill, David, Chris, …} . In a similar way, our approach also assumes that meanings of
the parts of an utterance are separated and classified according to their information
structure roles in context, though the representation used is somewhat different.
Dalrymple and Nikolaeva (2011) criticize certain assumptions commonly associ-
ated with structured meaning approaches to information structure; for example,
the transformation of structured meanings into non-structured (“standard”)
meanings in the presence of “focus-sensitive” operators such as only depends
on a considerably more restricted view of the structuring of meanings, and plays
no part in the analysis that Dalrymple and Nikolaeva propose. Their approach
assumes that the information structuring of utterance meaning is relevant for all
utterances, regardless of the presence or absence of operators like only; in addition,
it gives no preferential role to the information structure category of focus, but
considers all information structure categories to be relevant to the structuring of
meaning.
IP
NP I ’
s d
N VP
c
N V
David V NP
λy.λx.select (x, y) : c ⊸ [d ⊸ s ]
{ David : d }
selected N s ι
Chris : c
N
Chris
Here, in addition to the familiar c-structure and f-structure, we see the i-structure,
labeled s𝜎ι , which represents the organization of the meaning constructors from
(26) according to their status as topic or focus in the context provided.
As noted above, Butt and King (2000b) defined a function ι from c-structure
nodes to information structures. In contrast, Dalrymple and Nikolaeva (2011)
define ι as a function from semantic structures to information structure. In this
way, i-structure is projected directly from s-structure, and essentially organizes
meaning constructors, the building blocks of s-structure, according to i-structure
status. An expression like s𝜎ι , which labels the i-structure shown in (27), is
384 information structure
We can also use the labels in (29) to recast the full representation given in (27) in
a more reader-friendly format:
David V NP
{ [David] }
[selected]
selected N s ι
N [Chris]
Chris
NP I
’
N VP s d
N V c
David V NP
{ [David] }
selected N s ι
{ [selected-Chris] }
N
Chris
In this representation, meaning constructors are categorized according to their
information structure role by virtue of being included in one of the set of values for
an s-structure attribute (topic, focus, background, or completive). In (32), the
meaning associated with the phrase David fills the topic role, while the meaning
associated with the phrase selected Chris fills the focus role. Meaning constructors
contributed by all parts of an utterance are categorized in this way according to
their information structure contribution, and appear in the relevant category at
i-structure.
10.4.4.1 Semantic structure features Liao (2010) proposes the use of features
to represent the activation and accessibility of discourse referents in context.
The features she uses are adopted from Lambrecht (1994): status, with values
identifiable and unidentifiable; actv (“activation”), whose values include
active, accessible, and inactive; and the binary feature anchored. Liao uses
these features to analyze the distribution of overt and null anaphora in Mandarin
Chinese and to determine information structure roles. Example (33) repeats and
augments (88) from Chapter 8 in line with the view that s-structure is the proper
level for the representation of these features. In (33), we see the f-structure and
meaning constructor for David in a context where David is identifiable and active
in the discourse:
386 information structure
The meaning constructor for David pairs the meaning David with a semantic
structure in which an array of semantic and pragmatic features of that meaning are
represented; other features and values might be relevant in other contexts (§8.9).
The value of the df feature is not specified in the lexical entry for David (nor any
similar lexical entry, with the possible exception of question words; see §17.7) since
it is not an intrinsic lexical property of David that it plays a particular information
structure role. Rather, the information structure role of David depends, on each
occasion of its use, on the linguistic context in which it appears. As discussed
in §10.1, information structure roles are determined by the context of utterance;
they can be signaled linguistically in a variety of ways, such as by agreement
or casemarking, phrasal position, or prosody. It is also possible for information
structure roles to be associated by default with particular grammatical functions.
For example, in many languages, including English, the subject is the default topic
(Dalrymple and Nikolaeva 2011; Bresnan et al. 2016). In this way, any of a number
of components of the grammar may determine the df value of a particular meaning
in an utterance.
The relation between the df feature at s-structure and the categorization of a
meaning at i-structure depends on special constraints appearing in lexical entries
in Dalrymple and Nikolaeva’s model. The lexical entry for David, for example,
contains at least the following information:
modeling information structure 387
The first line of this lexical entry is entirely familiar. It is the specification in
the second line that is crucial for determining the correct information structural
configuration. This specification involves the meaning constructor David : d𝜎 ,
abbreviated with the label [David]. The functional description specifies that this
meaning constructor is required to be a member of the set of meaning constructors
specified by the value of the s-structure attribute df. The expression can be
paraphrased as follows:
For example, if the value of the semantic feature df in the s-structure correspond-
ing to ↑𝜎 is topic, then the meaning constructor must be a member of the topic set
at i-structure; see, for example, (32). If the value of df is specified as focus, then the
meaning constructor is a member of the focus set at i-structure, and likewise for
background and completive. Thus for any meaning constructor, when a value is
specified for the s-structure feature df, the information structure categorization of
that meaning constructor, whether topic, focus, background, or completive,
is concurrently specified.10 If no value is specified for the df feature, the meaning
constructor is not integrated into the information structure of the sentence, and the
resulting meaning for the sentence is semantically incomplete, incoherent, or both.
(37) IP −→ NP I!
(↑ subj) = ↓ ↑=↓
↑𝜎ι = ↓𝜎ι
((↓𝜎 df) = topic)
In the sentence David selected Chris, David is the subject, appearing in the specifier
of IP. As mentioned previously, there is a well-known default relationship between
the grammatical function subject and the information structural role topic. This
is captured by the phrase structure rule in (37) combined with a lexical entry
like (35). Taking the two together, we obtain the following partial configuration,
representing here only the c-structure and f-structure:
10 The way in which this is specified is similar to the use of the pcase feature in the specification of
the grammatical role of a prepositional phrase; see §6.2.
388 information structure
(38) IP
NP
(s subj) = d
I!
s𝜎ι = d𝜎ι 2 4 53
((d𝜎 df) = topic) s subj d pred ‘David’
N
David
(d pred) = ‘David’
[David] ∈ (d𝜎ι (d𝜎 df))
In the annotations under the phrase structure nodes and in the lexical entry in
(38), the ↑ and ↓ metavariables have been instantiated to the f-structure names s
and d. The ↑ = ↓ annotation on N, which ensures that the NP and its head David
correspond to the same functional structure, has been left implicit. The arrows
represent the familiar 𝜙 function from c-structure nodes to f-structures. In the
following exposition, we omit the c-structure in the interests of readability, retain-
ing only the functional descriptions harvested from the annotated c-structure. The
full f-description obtained from the c-structure in (38) is:
(39) (s subj) = d
s𝜎ι = d𝜎ι 2 4 53
((d𝜎 df) = topic) s subj d pred ‘David’
(d pred) = ‘David’
[David] ∈ (d𝜎ι (d𝜎 df))
The first line of (39) requires the f-structure d to be the subject of s, which is true
for the f-structure shown. The second line is crucial: it requires the information
structure corresponding to s and d to be the same. All nonhead daughters in the
phrase structure rules of a language bear such a specification, ensuring that all
members of a clause share the same information structure.11
The third line provides the optional, default discourse function topic associated
with the subject. If its discourse function is not otherwise specified, and as long
as compatible specifications are provided by the linguistic context (as discussed
below), and the prosodic and discourse prominence features of David are con-
sistent with the topic role, then the subject is associated with the i-structure role
topic.
The fourth line is contributed by the lexical entry for David, and requires that the
subject f-structure, d, have a feature pred with value ‘David’. Again, this is true for
the f-structure in (39). The fifth line, also contributed by the lexical entry, specifies
11 Heads need not be explicitly marked with this specification. The head of a phrase corresponds to
the same f-structure as the mother node according to the Head Convention (§4.2.1). Therefore, since
s-structure is projected from f-structure, and i-structure is projected from s-structure, a phrase and its
head correspond to the same f-structure, s-structure, and i-structure.
modeling information structure 389
that the meaning constructor [David] must bear the role specified by (d𝜎 df) at
i-structure, as discussed above.
If we assume that the default equation ((d𝜎 df) = topic) holds, we can simplify
the final line of these constraints as follows, according to the equality s𝜎ι = d𝜎ι :
2 4 53
(40)
s subj d pred ‘David’
(s subj) = d
(d𝜎 df) = topic 4 5
d𝜎 df topic
(d pred) = ‘David’
[David] ∈ (s𝜎ι topic) 2 3
s𝜎ι topic { [David] }
10.4.4.4 An example A more complete set of phrase structure rules for a simple
English clause is:
(41) IP −→ NP I!
(↑ subi) = ↓ ↑=↓
↑ ι =↓ ι
((↑ df) = topic)
I! −→ I VP
↑=↓ ↑=↓
VP −→ V
↑=↓
V −→ V NP
↑ = ↓ (↑ obj = ↓
↑ ι =↓ ι
We are now able, using these rules, to model the information structure of a sentence
such as David selected Chris in (27). The annotations on these rules should be
familiar and unremarkable, except for the annotation ↑𝜎ι = ↓𝜎ι on the NP daughter
of V! , which specifies that the object’s information structure is the same as the
information structure for the entire utterance. We augment the lexical entries for
selected and Chris with the requirement for their meaning constructors to bear
an information structure role: they are required to be a member of some set of
meaning constructors at i-structure, the set determined by the value of the attribute
df at s-structure:
390 information structure
(43) IP
’
NP
(s d s d
I
s ι=d ι
c
((d
N VP
N V
David NP
(d V (s c
[David] ∈ (d ι (d s ι=c ι
selected
(s ’ N
[selected] ∈ (s ι (s
N
Chris
(c
[Chris] ∈ (c ι (c
12 The discourse prominence features and prosodic contour must reinforce these assignments, or at
least must not conflict with the assignment of these roles. The contribution of prosody is explored in
Chapter 11. As yet there exists no widely agreed LFG model of how discourse context contributes to
the determination of information structure roles.
modeling information structure 391
(44) s ’
’
(s d
(d s d
(s c c
(c
(B) (d d :
(s s :
(c c :
(C) [David] ∈ (d ι (d
[selected] ∈ (s ι (s
[Chris] ∈ (c ι (c
s ι=d ι
s ι=c ι
We simplify the equations in (C), using the equalities in (B) and in the last two lines
of (C), to produce the compact description in (45) of the information structure for
this utterance:
Since [selected-Chris] can be derived by linear logic proof from [selected] and
[Chris] (as shown in the proof in 30), we can, once again, simplify this further to:
(46) { [David] }
s ι
{ [selected-Chris] }
392 information structure
Example (47) is an augmented version of the diagram in (43), showing the full
configuration and relationships between structures. The dashed arrows represent
the 𝜎 function from f-structure to s-structure, while the solid arrows represent the
ι function from s-structure to i-structure:
(47) IP
NP
(s d
I
s ι=d ι
((d
N VP
N V
David NP
(d V (s c
s ι=c ι
selected
N
(s ’
N
Chris
(c
’
f-structure: s d
c
David : d
λy.λx.select (x, y) : d ⊸ (c ⊸ s
s-structure: Chris : c
)
{ David : d }
Chris : c
λy.λx.select (x, y) : d ⊸ [c ⊸ s ]
i-structure:
We see in this diagram that this clause has a single i-structure in which meaning
constructors are classified according to their information structure role. This
i-structure is formed on the basis of information in the s-structures corresponding
to individual meaning constructors. For instance, the discourse function value
topic in the s-structure corresponding to David ensures that the relevant meaning
modeling information structure 393
constructor is a member of the topic set at i-structure (in this case, the only
member).
This equation states that the meaning constructor must appear in the i-structure
set which is specified by the value of (↑𝜎 df). Very often, this value is not specified
grammatically, but is determined by the linguistic and discourse context. Even in
cases where the value of (↑𝜎 df) is not grammatically specified, the equation in
(48) requires a value to be found. In this way, all meaning constructors play a
role at i-structure. This contrasts with claims that some elements can be entirely
excluded from information structure. For example, Kwon and Zribi-Hertz (2008)
assume that certain unmarked subjects and objects in Korean have no information
structure role. In Dalrymple and Nikolaeva’s model, such an analysis is impossible:
all parts of an utterance meaning are categorized at information structure. Infor-
mation structure does not simply pick out certain parts of a sentence and specify
particular roles for them, therefore, but partitions the entire sentence meaning
according to the relevant criteria.
394 information structure
1 Bögel’s (2012, 2013, 2014, 2015) p-diagram approach, grounded in details of the speech signal,
represents an alternative approach within the LFG framework; this is briefly discussed in §11.7.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
396 prosodic structure
is to account for the attested variability in intonation that can relate to factors such
as speech rate, speaker, style, and dialect, inter alia (see, for example, Arvaniti and
Ladd 2009; Arvaniti 2016). For a general introduction to phrasing and intonation,
which also includes data referred to in this chapter, see Hayes and Lahiri (1991).
The central concerns of a theory of the relation between prosody and other parts
of the grammar are the understanding and analysis of the relation between the
phonological features of a word or utterance and its abstract grammatical prop-
erties. It is clear enough, for example, that some connection must exist between
the phonological realization of a word and the c-structural, f-structural, and s-
structural features and structures that correspond to that phonological realization.
However, the precise nature of that connection and how it is constrained are less
clear. The problem is particularly acute above the word level: how is a particular
phonological or prosodic domain related to a particular syntactic or semantic unit?
This is not merely a matter of incorporating an additional, ancillary aspect of
linguistic analysis into an otherwise complete model of grammar. Prosodic features
such as pitch and duration can, and very often do, make crucial contributions to the
interpretation of an utterance. For example, in Japanese a question is distinguished
from a declarative sentence in writing by the presence of a sentence-final particle
such as ka. In spoken Japanese, the inclusion of this particle is optional (Hinds
1986); what is critical is the intonation pattern that is used. If the version of (1)
without ka is spoken with a final fall in intonation, the utterance is interpreted as a
declarative; if there is a final rise in intonation, it is interpreted as a yes-no question.
In the latter case, the contribution to meaning that prosody makes effectively
“overrides” the apparently unambiguous declarative syntax of the sentence. This
means that the pitch movement associated with the end of the utterance is crucial
to the interpretation of (1) when it does not include the question particle. Such
data illustrate how prosody can make an independent contribution to meaning:
Prosody may also be used to resolve ambiguity. For example, it has been claimed
that differences in prosodic phrasing can be used to distinguish between candidate
prosody: an independent level of structure? 397
Which of the two syntactic analyses is preferred may depend upon the location
of cues such as pauses which signify boundaries between prosodic constituents:
when a pause occurs after gradually, the interpretation is the one given in (3a), i.e.
gradually modifies the verb learn in the subordinate clause; when a pause occurs
after learn, the interpretation is the one given in (3b), i.e. gradually is part of the
main clause. Such data indicate that a close relationship can exist between prosodic
phrasing and syntactic structure (though one should be cautious about claiming
that disambiguation is the result of a one-to-one relationship between the two).
Furthermore, they demonstrate the importance of integrating prosody into any
account of the form-meaning relation.
Debate in the literature on prosody has centered on one fundamental issue in
particular, which we too must address before presenting the approach that we
adopt, namely whether the phonological/prosodic component is an independent
module with its own internal structure.
syntax and phonology are separate aspects of linguistic structure and as such are
each unable to interpret units, features, and structures specific to the other.
Even setting aside this incompatibility, the direct reference approach has been
criticized in the light of data which show that syntactic and prosodic struc-
ture are not, as one would predict under such a model, necessarily isomorphic;
for discussion of relevant data, see Chen (1987), Selkirk and Shen (1990), and
Bošković (2001). Lahiri and Plank (2010) seek to emphasize, based on observations
about Germanic languages that date back at least to Steele (1775/1779), that
the prosody-syntax relation is in fact characterized by extensive misalignment;
a lack of isomorphism between the two is more the rule than the exception (see
example 12).
The alternative to direct reference is an indirect reference approach, according
to which an interface relation serves to connect the syntax and phonology macro-
modules. An indirect reference approach copes well with data which demonstrate
a lack of isomorphism between syntactic and prosodic structure. While the two
are systematically related to one another via a mapping algorithm, under such an
approach there is no expectation that isomorphism is the result. With respect to
modularity and domain specificity, the distinction between syntax and prosody can
be maintained by assuming indirect reference: the objects and structures native to
one module cannot be interpreted or manipulated by the other module.
An indirect reference approach to the syntax-prosody interface is not by def-
inition restricted to one particular theoretical framework, but it clearly fits well
with the parallel grammatical architecture and general commitment to modular
specificity of LFG (§7.2). Of course, such an approach requires an independent
level of representation with its own primitives and organizing principles, which
provides the domains of application relevant to postlexical phonological processes.
The issue of precisely how to define and represent prosodic structure continues
to be the subject of research and debate in the wider literature. In the following
section, we review the ideas which have been most influential in the modeling of
prosody and its interfaces within the LFG framework.
Foot (Ft)
Syllable (S)
The syllable is the smallest unit of the Prosodic Hierarchy that is of concern to us.3
In many languages, syllables are the smallest units with which certain phonological
features, such as stress, are associated.4 A foot is a prosodic constituent consisting
of one or more syllables. Patterns in the alternation of stressed and unstressed
syllables in languages such as English are often stated by reference to rules of
foot formation. One or more feet make up a prosodic word. Feet are crucial in
the formation of prosodic words, but beyond this the foot is not relevant for the
purposes of modeling prosody’s interfaces with other modules of the grammar.
(For this reason, and in the interests of clear presentation, the foot level is omitted
in the analyses of prosodic structure which we present in this book.) Prosodic
words are usually assumed to be the domain within which lexical phonological
processes, and often other processes such as cliticization, apply. One or more
prosodic words make up a phonological phrase, the unit within which a single
prosodic contour may apply. The intonational phrase consists of one or more
2 For an interesting critical review of the Prosodic Hierarchy and Prosodic Phonology in general, see
Scheer (2011).
3 Syllables consist of one or more morae, the smallest prosodic units, but we will not make reference
to morae here. Morae themselves are analyzed as consisting of smaller phonological units, segments and
features, units which are generally considered to be sub-prosodic.
4 As it is not our intention to provide a comprehensive introduction to prosody and phonology, we
set aside the issue of precisely how stress should be defined. For an overview, see Fox (2000).
400 prosodic structure
prosodic phrases. This is the unit within which the smaller prosodic contours of
prosodic phrases are subject to larger overarching processes, such as the gradual
declination of median pitch; the boundaries of intonational phrases are therefore
generally characterized by pitch reset. Finally, the utterance is the largest prosodic
constituent, comprising any single unit of continuous speech consisting of one or
more intonational phrases.5
As with constituent structure, it is possible to state rules or constraints on the
formation of valid prosodic structures. These constraints (see Selkirk 1984 for
details) are usually subsumed under the “Strict Layer Hypothesis” (Nespor and
Vogel 1986: 7; Selkirk 2011: 347), which constitutes a general wellformedness
condition on prosodic structure. The Strict Layer Hypothesis requires that an
utterance be parsed exhaustively into non-recursive prosodic constituents, which
form “layers” that correspond to categories of the Prosodic Hierarchy. Thus, an
Intonational Phrase (IntP) can immediately dominate only Phonological Phrases
(PhPs), and a PhP can immediately dominate only Prosodic Words (PWs).6 For
example:
(5) Utt
IntP
PhP PhP
PW PW PW
5 There are many alternative definitions of these prosodic constituents. See, for example, Selkirk
(1978), Nespor and Vogel (1986), Levelt (1989), Wheeldon (2000), and Frota (2012). As stated
previously, we seek to define these constituents in purely prosodic terms, whereas many of these authors
define them in a combination of syntactic and prosodic terms.
6 This strict view of the constraints on prosodic domination has been called into question by a
number of authors, including Inkelas (1989) and Itô and Mester (2003). In particular, it has been
noted that exhaustivity and the bar on recursivity may be best viewed as tendencies, given that they
can be violated in a number of languages. For a recasting of the Strict Layer Hypothesis in Optimality-
Theoretic terms that takes this into account, see Selkirk (1995). Bögel (2015) provides an alternative
approach within LFG which explicitly avoids the prosodic hierarchy as a basis for prosodic structure.
representing prosodic structure in lfg 401
When spoken, the two readings of the sentence in (6) are associated with different
phrasings: the idiomatic reading ‘was startled’ is licensed when the VP corresponds
to a single PhP (7a), while the literal reading ‘saw a ghost’ emerges when the verb
form and bh ut ‘ghost’ constitute two distinct PhPs (7b), as indicated by the labeled
brackets:
7 Note that the p-structure feature p-form is not the same as the f-structure feature pform encoding
the form of a preposition (§2.5.4).
402 prosodic structure
Butt and King (1998a) propose that p-structure is projected from c-structure just
as f-structure is, and that prosodic constituents such as IntPs and PhPs are derived
from syntactic constituents by means of mapping processes. The relevant map-
ping processes permit not only isomorphism, but also some degree of difference
between the two structures (for example, by increasing or decreasing the levels of
embedding in p-structure relative to the c-structure input). With respect to the
ambiguity of (6), Butt and King argue that the prosodic distinction between the
two readings corresponds to a difference in c-structure. In the case of the idiomatic
reading, Butt and King analyze the N bh ut ‘ghost’ and the V dekh lam ‘saw’ as sisters
at c-structure; together they constitute a V! , as shown in (10). This single bar-
level syntactic constituent corresponds to a single PhP in the relevant prosodic
representation (8). By contrast, in the literal reading the NP bh ut ‘ghost’ occupies
specifier position in the VP, as shown in (11). The syntactic relationship between
the NP and the verb is not as close as in the case of the idiomatic reading, and
this is also the case in the prosodic representation (9): the two relevant syntactic
constituents correspond to two separate PhPs rather than forming one PhP.
(10) ami bh ut dekh -l-am. (11) ami bh ut dekh -l-am.
I ghost see-pst-1sg I ghost see-pst-1sg
‘I was startled.’ ‘I saw a ghost.’
S S
NP VP NP VP
Pron V! Pron NP V!
ami ami
N V N V
I I
bh ut dekh lam bh ut dekh lam
ghost saw ghost saw
representing prosodic structure in lfg 403
8 For an introduction to AM/ToBI see Beckman et al. (2005) and Ladd (2008); for ToBI analyses of
data from a range of languages, see the papers in Jun (2005, 2014).
404 prosodic structure
order to account for the type of phenomena exemplified in (2), which show that
prosody can make a crucial contribution to meaning. O’Connor (2006) takes a
different view, according to which p-structure has a direct relationship only with a
level which he refers to as discourse structure. The architecture which O’Connor
proposes is non-standard and has not been adopted beyond his work; we do not
consider it further here.
Perhaps surprisingly, given LFG’s co-description architecture and the general
commitment to modularity, the next major proposal concerning the analysis of
prosody, put forward by Bögel et al. (2009), did not include a separate level of
prosodic representation. As a result, this work has less in common with the indirect
reference approaches outlined in §11.2.1 than the LFG analyses which preceded
it. Bögel et al. (2009) proposed a “pipeline architecture” as a way of capturing
facts about the (mis)alignment of prosodic and syntactic constituents. Rather than
treating prosodic structure as projected from c-structure, Bögel et al. argue that
prosodic information feeds into c-structure analysis. In their model, the input to
c-structure is a prosodically bracketed string. Phrase structure rules then apply
as normal to produce a valid c-structure for the string, but with one significant
augmentation: phrase structure rules are formulated so as to make reference not
only to syntactic categories but also to prosodic brackets. These rules are used in
Bögel et al. (2009) to capture the misalignment which characterizes the syntax-
prosody relation in Germanic languages that, as Lahiri and Plank (2010) discuss, is
often due to the differing behavior of function words in the formation of syntactic
and prosodic constituents. Lahiri and Plank exemplify using an old advertising
slogan:
The bracketing in (12) shows that the English function words a and of group with
a constituent to their right in the syntax, but to their left in the prosody. (In fact,
the slogan appeared in writing as Drinka pinta milka day, reflecting the prosodic
bracketing shown.) In their analysis of this example, Bögel et al. (2009) propose that
the phonological string undergoes prosodic parsing, which introduces the relevant
prosodic boundaries into the string. This “prosodically annotated string” is then
the input to the syntactic component:
A key feature of this pipeline architecture is that syntax can interpret the prosodic
boundaries that are inserted into the string. Though this is not relevant for the
representing prosodic structure in lfg 405
example in (13), it is crucial to Bögel et al.’s (2009) account of the role of prosody
in resolving syntactic ambiguity. For example, the phrase in (14) is ambiguous,
depending on what exactly the AdjP old modifies, and is compatible with either of
the syntactic analyses indicated:
Bögel et al. capture this fact about alignment with their Principle of Prosodic
Preference: “syntactic structures with constituent boundaries that do not coincide
with prosodic boundaries are dispreferred.” While this approach successfully
accounts for the role that prosody can play in resolving syntactic ambiguity
and captures cases of mismatch between syntax and prosody, it does so at the
cost of introducing prosodic information into the syntax macromodule. In an
important respect, this violates the grammatical principle of modularity (§7.2):
the syntactic module should not be able to interpret nonsyntactic objects. This
is circumvented by proposing that a prosodic boundary has an additional role
as one of two syntactic categories (left boundary, LB, or right boundary, RB),
with the result that prosodic objects are effectively “disguised” as being syntactic:
under this approach, LB and RB are terminal nodes in the c-structure tree. This
does not represent a strictly modular approach to the issue. Moreover, the rather
simplified prosodic representation involving LB and RB that Bögel et al. (2009)
adopt does not allow easy reference to the different types of units that belong
to the Prosodic Hierarchy, something which is essential in the analysis of many
phenomena. Presumably, reference to distinct prosodic units could be made via a
separate, additional prosodic component introducing a large number of differently
labeled prosodic brackets into the string, but this adds further complexities. Bögel
et al. (2009) do not explore the details or the wider implications of including
such a component. Subsequent work on prosody and its interfaces within LFG has
moved away from the “pipeline” architecture proposed by Bögel et al. (2009) and
returned to an approach in which prosodic structure represents a distinct level
in the parallel architecture, a position more compatible with strict modularity. In
later work such as Bögel (2015), Bögel rejects the prosodic hierarchy as the basis for
p-structure.
406 prosodic structure
NP I!
N VP
π V
π
David arrived
modeling the prosody-syntax interface 407
The string can be thought of as a grammatical signal parsed into minimal units.
The majority of analyses assume this view of the string: it is regarded as compris-
ing minimal syntactic units which are then subject to phrase-structural analysis
in the c-structure, via the π projection. However, this ignores the fact that a
string can also be parsed into minimal phonological/prosodic units. The string,
as Asudeh (2009: 107) observes, thus represents part of the syntax-phonology
interface.
Dalrymple and Mycock (2011) were the first to make a detailed proposal con-
cerning this dual nature and function of the string. They argue that the string must
be analyzed as having two distinct aspects: one syntactic, the s-string, the other
phonological/prosodic, the p-string. Any linguistic signal can be parsed in two
ways: the s-string represents the parsing of a signal into minimal syntactic units,
and so corresponds to the conception of the string in most earlier works, while
the p-string represents the parsing of a signal into minimal phonological/prosodic
units. Just as the s-string is the basis of the syntactic phrase-structural analysis of
an utterance, so the p-string is the basis of the prosodic phrase-structural analysis.
To give an example, in Germanic languages units are formed according to the
types of rhythmic principles highlighted by Lahiri and Plank (2010), along with the
structure of the Prosodic Hierarchy; see §11.2. The units of p(rosodic)-structure
are native to the phonology macromodule, and the projection of p-structure is
based on phonological and prosodic features; there is no sense in which p-structure
is derived from syntactic phrase structure, and no assumption of a necessarily
close correlation between syntactic and prosodic phrase structure. The two are,
of course, related to one another, but it is the string which represents the point of
interface; thus, modularity is respected. Dalrymple and Mycock (2011) illustrate
their proposals using the sentence Anna was studying at the university, providing
a “double-tree” analysis of the two aspects of the string. The s-string, p-string,
c-structure, and p-structure which they propose—with some emendations to bring
it in line with proposals by Mycock and Lowe (2013)—are shown in (17). The
units of the s-string are each related to a terminal node of the c-structure via the π
projection. The prosodic units related to these syntactic units form the p-string.
Parallel to the syntactic analysis of this utterance, the units of the p-string are
each related to a terminal node of the p-structure via the β projection, with “S”
standing for syllable. Thus, the two aspects of the string receive equivalent but
distinct analyses:9
9 In (17), each PW is also a separate PhP, but it could be the case that, for example, the PWs are
grouped into two PhPs, giving (æ nə wəz stʌ di ɪŋ ət ðə ju nə)PhP (vɜ: sə ti)PhP . Our approach is flexible
in order to account for the attested variability in intonation mentioned in §11.1. While for our purposes,
nothing in (17) hinges on the placement of PhP boundaries, their inclusion is important because they
can be associated with other aspects of prosody. For instance, voicing assimilation in Bengali is bounded
within PhPs (Hayes and Lahiri 1991). Thus, even if a PhP comprises a single PW, this does not mean
that the two can or should be conflated.
408 prosodic structure
(17) IP
I!
VP
V!
PP
P!
!
NP V NP
N!
N I V P D N
π π π π π π
s-string Anna was studying at the university
PW PW PW
IntP
Utt
The relation between s-string and p-string units or, to put it another way, between
the syntactic and prosodic parses of an utterance, is represented in (17) using
dotted lines. Dalrymple and Mycock (2011) propose that the s-string–p-string
relation is defined by information stored in lexical entries. Underpinning their
proposal is a more fine-grained understanding of the nature and content of lexical
entries, according to which a lexical entry specifies both a s(yntactic)-form and a
p(honological)-form, as well as the c-structure category and f-description. The co-
occurrence of s-forms and p-forms in lexical entries constrains the analysis of a
string by requiring that any lexical item identified as a syntactic unit necessarily
corresponds to a phonological/prosodic unit or units, in the same relative position
in the string. The role of the lexical entry in mediating between the s-string and
the p-string is illustrated in (18). The diagram in (19) exemplifies the portion of
the grammatical architecture for the noun university which appears in (17):
modeling the prosody-syntax interface 409
(18) C-structure
Lexical entry
π
s-form s-string
p-form p-string
β
P-structure
(19) Lexical entry, s-string, and p-string: Dalrymple and Mycock (2011)
C-structure
Lexical entry
π
university university
/ju.nɪ.vɜ:.sɪ.ti/ /ju nə vɜ: sə ti/
β
P-structure
Dalrymple and Mycock (2011) therefore use the p-string, the s-string, and the
lexicon to model the interface between the syntax and phonology macromodules.
Under this analysis, information of various kinds can be associated with s-string
and p-string units. Mycock and Lowe (2013) develop Dalrymple and Mycock’s
(2011) approach to lexical entries by proposing that s-string and p-string units are
not atomic, but in fact are bundles of information which they represent as feature
structures.
A key component of any feature structure representing a unit of the s-string or
p-string is the feature fm (form), whose value is related to the s-form or p-form
of a lexical entry. For example, the lexical entry for university includes:
The symbol • is used to refer to s-string units (Mycock and Lowe 2013). In the case
of an s-string fm feature such as that shown in (20), the relation with the s-form in
the relevant lexical entry (i.e. university) is one of identity. That is, the s-form of a
word as it appears in a lexical entry is identical to the value of the fm feature in any
s-string instantiation of that word. In the case of the p-string, on the other hand,
this is not the case. For instance, the p-form of university includes five syllables
410 prosodic structure
(shown separated by periods), of which the second and fourth syllables contain
the vowel /ɪ/:10
(21) /ju.nɪ.vɜ:.sɪ.ti/
Notice that in the p-string in (17) the vowels in those same syllables are realized
as /ə/. Phonological processes related to speech tempo and other contextual
factors may apply, rendering the relation between p-forms and p-string fm fea-
tures opaque. (It is important that the underlying vowels are retrievable though,
as speakers can clearly access and produce them in instances of slow, careful
production.) In informal terms, it is not difficult to see how the application of
such phonological processes affects the derivation of p-string fm features from
p-forms, and likewise how the corresponding undoing of those same rules affects
the reconstruction of p-forms from p-string fm features. For the present purposes,
however, we make no proposals regarding the precise formalization of this aspect
of phonology.11
The other respect in which fm features differ from the s-forms and p-forms
of lexical entries is the possibility of many-to-one correspondences. Again, this is
clearly seen by comparison of (21) with (17). The single p-form of the lexical entry
in (21) corresponds to five distinct units in the p-string representation. Under a
feature-structure approach, each of these p-string units is represented as a distinct
p-string feature structure with its own p-string fm feature; that is, there is one
p-string unit, represented as a separate feature structure, for each syllable. We
assume that the specification of p-string units is part of the lexical information
stored in lexical entries.12
Of course, fm features and their values are not the only components of a lexical
entry. Each lexical entry also includes a c-structure category and f-description. At
this point, we must note an inconsistency in the formal details of the presentation
so far. Following long-established tradition in LFG, we have been using the
variable ↑ in f-descriptions in lexical entries. Thus, for example, the f-structure
number feature for university is specified by the f-description (↑ index num) = sg.
However, as discussed in §5.3.1, the variable ↑ is an abbreviation for 𝜙(ˆ⃰), that
is, the f-structure projected, via the 𝜙 function, from the mother of the current
c-structure node. This variable was appropriate when a syntactic element such
as university was conceived of as a terminal node of the c-structure tree, with
the preterminal node N as its mother. In that case, ↑ refers to the f-structure
corresponding to the N node, as required. However, as discussed in §3.5 and at
the start of the current section, most major approaches to the LFG architecture
follow Kaplan (1987, 1995) in understanding the (s-)string to be distinct from the
10 This is a simplification. As Bögel (2015) observes, following Levelt et al. (1999), segments and the
metrical frame are most likely stored separately in the lexicon. See Bögel (2015) for more on the precise
contents of the p-form portion of a lexical entry.
11 For a formalization of some postlexical phonological processes in an LFG setting, see Bögel (2015).
12 This issue affects not only the p-form–p-string relation, but also the s-form–s-string relation, since
single lexical entries can specify multiple s-string elements; see Lowe (2016a).
modeling the prosody-syntax interface 411
(22)
The lexical entry for university (with its parts labeled) is therefore:13
While a lexical entry of this type is technically more accurate, the use of ↑ is a
well-established tradition in LFG, and is certainly more readable. We therefore
retain this traditional notation, on the understanding that ↑ appearing in a lexical
entry refers to a different function from ↑ when it appears in phrase structure
annotations. In a lexical entry, ↑ is defined as 𝜙(π(•)), where the π function maps
from units of the s-string to terminal nodes of the c-structure. In a phrase structure
rule, ↑ is defined as 𝜙(ˆ⃰) (or, to be completely explicit, as 𝜙(M(⃰)), in terms of the
mother function M on c-structure nodes).
13 On the function λ, which relates a node of the tree to its label, see §5.1.2.
412 prosodic structure
the years, for instance the end-based (also known as the edge-based) approach
of Selkirk (1986) and the Align family of constraints in Optimality-Theoretic
approaches including Selkirk (1995). Following Dalrymple and Mycock (2011),
Mycock and Lowe (2013) assume that string units appearing in particular contexts
are associated with sets of labels. These labels represent the left and right edges of
the constituents in the associated structure that belongs to the relevant module
of the grammar (i.e. c-structure or p-structure). So, in (17), the s-string element
‘Anna’ is associated with the left edge of the IP in the c-structure, and also with
both the left and right edges of an NP and N. This is represented in the relevant
string unit by means of features l (left edge) and r (right edge) whose values are
sets comprising information concerning which constituents this particular s-string
unit represents the left or right edge of. Thus, because the N Anna is the left edge
of the IP, NP, and N constituents and the right edge of the NP and N constituents,
the representation of the s-string unit for Anna in (17) is:
(24) fm Anna
l { IP, NP, N }
r { NP, N }
There are two p-string elements corresponding to the s-string unit for Anna: ‘æ’
and ‘nə’. The first of these is associated with the left edge of a PW, a PhP, the IntP,
and the Utt in the p-structure. As is the case for edge information related to s-string
units, this edge information about p-string units appears as values of the features
l and r within p-string attribute structures, as shown in (25).14
The features fm, l, and r are not the only features that have been proposed for
s-string and p-string attribute structures. Mycock and Lowe (2013) also include
prosodic information associated with particular p-string units in p-string feature
structures. For example, they assume the feature syllstress with value p in order
to represent the location of primary stress. Thus, the full representation of the two
p-string units for Anna in (17) is:
(25) fm æ
fmnə
syllstress p
l { }
l { Utt, IntP, PhP, PW }
r { }
r { }
14 For a different approach to capturing prosodic phrasing and its encoding in LFG, see Bögel’s work
on the p-diagram. Bögel (2015) provides full details; see also §11.7.
modeling the prosody-syntax interface 413
of the node to which it applies; it is therefore the inverse of the mother relation
M discussed in §5.3.1. The leftmost immediate daughter of the node in question
can then be defined as the member of the set of daughter nodes which is not
preceded, in linear terms, by any other member of the same set. Likewise, the
rightmost immediate daughter of the node can be defined as the member of the
set of daughter nodes which is not followed, in linear terms, by any other member
of the same set. These functions, from nodes to leftmost and rightmost immediate
daughters, are labeled Dl and Dr respectively, and are defined, in relation to c-
structure nodes, as follows:
This is shown graphically in (27), where the mother node is represented as c1 , the
leftmost daughter is Dl (c1 ), and the rightmost daughter is Dr (c1 ):
(27) c1
The leftmost and rightmost terminal nodes dominated by a node can be defined
by the recursive application of these functions until a node is reached that has no
daughters. Mycock and Lowe (2013) label the functions that find these terminal
nodes as Tl and Tr respectively. The definitions they provide are given in (28).
These rules can be informally read as follows: the leftmost/rightmost terminal
node from the current node is the current node if the current node has no daugh-
ters; otherwise, it is the leftmost/rightmost terminal node of the leftmost/rightmost
daughter (respectively) of the current node. The rule applies recursively to find the
appropriate terminal descendant of any node:
(
(28) a. Tl (⃰) ≡ ⃰ if D(⃰) = ∅
else Tl (Dl (⃰))
(
if D(⃰) = ∅
b. Tr (⃰) ≡ ⃰
else Tr (Dr (⃰))
In (29), we display a sample tree in which the current node (⃰ in the definitions
above) is represented as c1 , the leftmost terminal daughter is Tl (c1 ), and the
rightmost terminal daughter is Tr (c1 ):
Tr (c1 )
Tl (c1 )
414 prosodic structure
Finally, the s-string elements corresponding to these terminal nodes are straight-
forwardly obtained by applying the inverse function π−1 from terminal nodes
of the c-structure to elements of the s-string. In (30), the leftmost s-string ele-
ment of the node labeled c1 is π−1 (Tl (c1 )), and the rightmost s-string element is
π−1 (Tr (c1 )):
(30) Sample configuration: π−1 (Tl (c1 )) is the leftmost s-string element of c1 , and
π−1 (Tr (c1 )) is the rightmost s-string element
c1
Tr (c1 )
Tl (c1 )
π−1
π−1
π−1 (Tl (c1 )) π−1 (Tr (c1 ))
Mycock and Lowe (2013) propose to use arrows as abbreviations for the paths
from c-structure nodes to leftmost and rightmost string elements. The arrow ⇙
abbreviates the function from a c-structure node to the leftmost s-string element
dominated by it, and the arrow ⇘ abbreviates the function from a c-structure node
to the rightmost s-string element dominated by it:
Tr (c1 )
Tl (c1 )
π−1
π−1
3 4 3 4
fm sform1 fm sform2
modeling the prosody-syntax interface 415
The tree in (35) explicitly shows these specifications on all c-structure and p-
structure nodes. Notice that the labels appearing in the sets which are the values
of each string unit’s l and r features correspond exactly to the specifications made
by the arrows in the c- and p-structures.
As mentioned previously, in addition to the features fm, l, and r, Mycock
and Lowe (2013) include prosodic information associated with particular p-string
units in p-string feature structures. The feature syllstress with value p, which
appears in (35), represents the location of primary stress:15
15 Lowe (2016a) also proposes a feature clitic, which can appear in both s-string and p-string
units, to distinguish syntactic or prosodic clitics from non-clitics at the level of the string, and an s-
string or p-string feature field, which contains labels referring to all syntactic categories dominating a
particular element in the c-structure or to all the prosodic categories dominating a particular element
in the p-structure.
416 prosodic structure
NP I!
NP ∈ (⇙L), NP ∈ (⇘R)
VP
VP ∈ (⇙L), VP ∈ (⇘R)
N V!
N ∈ (⇙L), N ∈ (⇘R)
V NP
V ∈ (⇙L), V ∈ (⇘R) NP ∈ (⇙L), NP ∈ (⇘R)
π N
π N ∈ (⇙L), N ∈ (⇘R)
π
fm Anna fm hit fm Norman
l { IP, NP, N } 5 ! 6 l { NP, N }
s-string l I , VP, V! , V
5 6
r { NP, N } r {V} r IP, I! , VP, V! , NP, N
fm fm
æ fm nə fm hıt nɔː fm
syllstress p mən
syllstress p syllstress p
5 6 l { } 5 6
l 5 PhP, PW 6
l { }
p-string l
IntP, PhP, PW
5 6 l PhP, PW
5 6
r PhP, PW r IntP, PhP, PW
5 6
r { } r PhP, PW r { }
β β β β β
S S S S S
PW PW PW
PW ∈ (#L), PW ∈ ($R) PW ∈ (#L), PW ∈ ($R) PW ∈ (#L), PW ∈ ($R)
IntP
IntP ∈ (#L), IntP ∈ ($R)
The approach to prosody and its interfaces developed by Dalrymple and Mycock
(2011) and illustrated in (35) makes crucial reference to edge information asso-
ciated with string elements. Dalrymple and Mycock (2011) themselves note that
“passing information about the edges of all major constituents into the p-string
and s-string . . . may seem excessive,” but also point out that in at least some cases,
for instance when speech is slower and more careful, a greater degree of alignment
exists between syntactic and prosodic constituents, meaning that “information
modeling the prosody-syntax interface 417
NP VP
Pron V!
N V
PW PW PW
PhP PhP
IntP
b. Reading 2: ‘I saw a ghost.’
[(ami)PhP ]NP [[(bh ut)PhP ]NP (dekh lam)PhP ]VP
S
NP VP
Pron NP V!
N V
PW PW PW
IntP
Each phrasal boundary in the syntax (NP or VP) has a corresponding PhP boundary
in the prosodic analysis.16 (There is no “double marking” or “double edge effect” if
an edge of the VP also represents an edge of an NP.) The two readings are therefore
associated with distinct c-structure and p-structure pairs. As proposed by Dal-
rymple and Mycock (2011), information about the edges of all major constituents
is recorded as labels in l or r feature-value sets, by means of requirements like
those given in (34). The string representations for the Bengali utterances, based
on the analysis presented by Butt and King (1998a) (which itself is based on that
of Hayes and Lahiri 1991), are as shown in (38) and (39). Under neutral focus, a
16 These boundaries are crucial in accounting for different voicing assimilation possibilities in these
two examples; see Hayes and Lahiri (1991).
modeling the prosody-syntax interface 419
fm bh ut fm dekh
syllstress p syllstress p fm lam
l { }
p-string tone h l { PW }
r { IntP, PhP, PW }
l { PhP, PW } r { }
r { PW }
fm bh ut fm dekh
syllstress p syllstress p fm lam
p-string l { }
l { PhP, PW } tone h
r { PhP, PW } l { PhP, PW } r { IntP, PhP, PW }
r { }
In examples (38) and (39), the principle of Interface Harmony, coupled with
the rules of alignment in Bengali, mean that XP labels in an s-string feature
structure are matched with phrase labels in the values of the l or r attributes of the
corresponding p-string feature structure. This means that, ordinarily at least, the
s-string in (38) is not paired with the p-string in (39), for example. Such constraints
on wellformedness at the interface between the syntax and phonology macro-
modules, situated in the string, account for the relationship between c-structure
and p-structure in cases of prosodic disambiguation. They also account for the
ambiguity that remains when only the written form of the sentence is available,
because without p-structure constituent edges to match with, in principle either of
the two c-string/s-string analyses is possible.
When it comes to Interface Harmony, determining exactly what prosodic infor-
mation (for example, boundary type) is important to the syntax, and vice versa, is a
key area for future research. For instance, in the Bengali examples we see that each
NP forms a separate PhP comprising the relevant PWs, with the attendant result that
in the literal reading (Reading 2), the V dekh lam forms a PhP on its own. However,
as we will see in the following sections, information about the location of syntactic
and prosodic constituent boundaries is not the only sort of information that may
be associated with string units and may therefore appear as values of the l and r
420 prosodic structure
17 In what follows, we assume that a declarative question has the same interpretation as an equivalent
polar interrogative that is syntactically marked as such by subject-auxiliary inversion, for example Was
Anna studying at the university? This is a simplification, as declarative questions exhibit contextual
restrictions which polar interrogatives do not; on this, see, for example, Gunlogson (2003). Our aim
here is to illustrate the mechanism whereby prosody can have an effect on meaning within the LFG
framework. The full details of an LFG analysis of declarative questions await further research.
declarative questions 421
defined as β−1 (Tr (+)), the rightmost p-string element. β is the function from p-
string elements to terminal nodes of the prosodic structure, and β−1 is the inverse
of β. An example configuration is:
(41) β−1 (Tl (+)) is the leftmost p-string element of the prosodic constituent +,
and β−1 (Tr (+)) is the rightmost s-string element
+
Tr (+)
Tl (+)
β−1
β−1
β−1 (Tl (+)) β−1 (Tr (+))
Thus, we define the relations Tls and Trs , which pick out the stressed syllable closest
to the left and right edge of a constituent +, as follows:18
According to the first line of this definition, Tls (+) is the leftmost syllable Tl (+) in
the prosodic phrase if the p-string element corresponding to Tl (+) contains the
feature syllstress with value p. A sample configuration meeting these require-
ments is given in (43):
Tr (+)
Tl (+)
β−1
3 4
β−1 (Tl (+)): syllstress p
The second line of the definition of Tls applies in case the leftmost syllable does not
correspond to a stressed element of the p-string. In that case, we check the next
p-string element to the right, and continue rightwards until we encounter a
p-string element containing the feature syllstress with value p.
As usual, we can introduce a convenient abbreviation for these concepts, using
arrows superscripted with s:
18 The function N, which appears in (42), finds the next element in linear order when applied to
string elements, as discussed in §6.11.2; N −1 finds the preceding element.
422 prosodic structure
In sum, the definitions of the functions #s and $s are similar to the definitions
of # and $ in (33). The only difference is the presence of an additional specifica-
tion: these functions do not necessarily find the absolute leftmost and rightmost
syllables, but rather the leftmost and rightmost syllables that have the feature-value
pair which we use to represent primary stress, i.e. syllstress p. These two pairs of
arrows are all that we require to analyze declarative questions and prosodic focus
marking, which is the subject of §11.6.
For the prosodic part of the analysis of a declarative question, the interrogative
tune is a property associated with an IntP. The specification of this tune is intro-
duced by a prosodic phrase structure rule, parallel to the syntactic phrase structure
rules with which we are familiar:
Line 1 of the annotation below the final PhP in (45) means that the value of n_tone
is l for the leftmost stressed syllable (i.e. the leftmost p-string unit including the
feature-value pair syllstress p) in this PhP. Line 2 requires the value of the
rightmost p-string unit in this PhP to have an rb_tone (right boundary tone)
feature whose value is h. (45) captures the intonational contour we are concerned
with, but what it does not do is link it to the semantic contribution associated with
this interrogative tune. To capture this, we must turn now to the syntactic part of
the analysis.
As discussed in Chapter 8, s(emantic)-structure is projected from f-structure,
that is, from the syntax. As a consequence, it is necessary to associate the meaning
constructor for interrogative semantics with a syntactic unit representing the
clause as a whole. We assume a constructional meaning, introduced in the IP
phrase structure rule, which contributes the relevant meaning constructor (§8.6).
For present purposes, we define the semantics of polar interrogativity as follows
(which represents a simplification, but suffices for the purposes of illustration in
this case):
(47) IP −→ XP I!
↑=↓
PolarIntSem ∈ (⇘ r)
[PolarInt]
declarative questions 423
This rule ensures that a label, “PolarIntSem,” appears as a member of the set value
of r in the rightmost s-string unit corresponding to the root IP. This specifica-
tion always appears together with the polar interrogative meaning constructor
[PolarInt], as illustrated in (52).
Returning now to the prosodic part of our analysis, just as on the syntactic
side, the interrogative tune specification is always accompanied by a further
specification which defines a label PolarInt. This label appears as a member of the
set value of r in the rightmost p-string unit in the IntP:19
(48) PolarInt ∈ ($ r)
Putting together (48) and (45), we have a specification of the interrogative tune
which combines its tonal features with a label relating to polar interrogativity:20
Following the analysis of Lahiri and Plank (2010), adopted here, the last three
syllables of university form a prosodic word (PW). A partial representation of
the p-structure and p-string for these three syllables illustrates the way this rule
determines certain feature values:
(50) fm vɜː
syllstress p fm sə fm ti
rb_tone h
n_tone l l { }
p-string l { }
l { PhP, PW }
r { }
r { Utt, IntP, PhP, PW, PolarInt }
r { }
β β
β
S S S
PW
We are now in a position to put together the syntactic and prosodic aspects of our
analysis of declarative questions. Crucial to this analysis are the feature values in the
rightmost string feature structures, specifically “PolarIntSem” in the final s-string
unit (see 47) and PolarInt in the final p-string unit (see (49)), which are shown in
this partial representation:
19 We follow Mycock and Lowe (2013) in distinguishing syntactic labels like “PolarIntSem” from
prosodic labels like PolarInt by putting the latter in italics. Note that these are simply labels, and do
not imply the presence of semantic properties in either the c-structure or p-structure: PolarInt could
equally well be labeled XYZ, or “PolarIntSem” could be “XYZ.”
20 Note that we do not assume a one-to-one relationship between a tune and meaning.
424 prosodic structure
(51) NP
N!
N
π
fm university
l { N }
s-string
r { IP, VP, PP, NP, N, PolarIntSem }
fm ti
rb_tone h
p-string l { }
r { Utt, IntP, PhP, PW, PolarInt }
β
S
PW
But what is the connection between the labels “PolarIntSem” and PolarInt? Both
are clearly intended to link with polar interrogativity, which is characteristic of
this construction, but the labels and the syntactic/prosodic structures with which
they are associated are not directly dependent on one another: neither is derived
from the other and their sphere of application is entirely separate, in line with
strict modularity. And yet it is clear that the two must be related if the analysis
is to achieve its aim of pairing the interrogative tune with the appropriate meaning
constructor. Once again, the key assumption is the principle of Interface Harmony.
Interface Harmony requires that information about a meaning constructor
introduced at c-structure (for example [PolarInt]), which has its presence recorded
via a label in the s-string (i.e. “PolarIntSem”), must be matched with an equivalent
label in the corresponding p-string unit (i.e. PolarInt). As in other cases, both
labels should be associated with the same feature (i.e. l or r) in the relevant
string units. Because the labels “PolarIntSem” and PolarInt are intrinsically related
to one of the two separate phrase structure rules which introduce either the
relevant meaning constructor or the relevant tune, the appropriate semantics and
intonational contour must also co-occur. In the case of a declarative question,
then, the principle of Interface Harmony applies so that, if “PolarIntSem” appears
as a member of the set value of r in any s-string unit, then the equivalent label
PolarInt must appear in the corresponding p-string unit as a member of the set
value of r (along with the feature values associated with the rest of the interrogative
tune phrase structure rule, as shown in (49) and (50)). In the case of (52), the
relevant labels co-occur and correspond to one another; if they did not, Interface
Harmony would be violated and the resulting structures would be ungrammatical.
The phrases to which these labels are related (IP on the syntactic side, IntP on the
prosodic side) are also wellformed according to (47) and (49). This means that the
structure shown in (52) is grammatical. In this way, the contribution of prosody
to syntax and semantics can be modeled in a framework that does not require a
direct relation between syntactic and prosodic structure, and does not blur the
distinction between these two modules of grammar:
426 prosodic structure
In this section, we have seen how to model p-structure and account for the ways
in which it interacts with syntax and semantics. In the next section, we turn to
another aspect of interpretation that can have a close relationship with prosody:
information structure.
The answer sentences in (53) and (54) exemplify narrow focus: the word that
bears prosodic focus marking is also the only element that bears focus status
at information structure. However, it is also possible for a single word to bear
prosodic focus marking when more than one word has focus status at information
structure. This is known as broad focus, and is exemplified in (55). In (55a), the
whole VP in the answer sentence has focus status, but the Nuclear Tone is associated
with Norman (specifically, with the initial stressed syllable of Norman).21 In (55b),
21 Here, we set aside the issue of an event, but not the type of that event, being presupposed in the
context of What happened? or What did X do (to Y)? type questions. When such a question is asked, it
prosodic focus marking 427
the subject NP some old woman is the focus of the answer sentence, but the Nuclear
Tone is associated with the initial stressed syllable of woman; cf. (54):
Although the distinction between broad and narrow focus is generally accepted in
the literature on focus marking, in an important sense the analysis that we adopt
here transcends the distinction between broad and narrow focus. This distinction
depends on different correlations between focus in the syntax and information
structure on the one hand, and the prosodic marking of focus status on the other. In
the modular grammatical architecture which we assume, these different aspects of
focus and its encoding are specified separately. Important to understanding these
inextricably linked dimensions of the notion “focus” are two concepts which we
refer to as Extent of Focus (Foc-Extent) and Exponent of Focus (Foc-Exponent).
Foc-Extent (also known as the Focus Domain) refers to the portion of a sentence
which can be said to have focus status in information structure terms, while Foc-
Exponent is the indication at some level of representation, for example prosody
(p-structure), of the focus status of part or all of a sentence.22 In the examples in
(55), square brackets enclose the syntactic elements that constitute the Foc-Extent,
while the h Nuclear Tone annotation indicates the word (given in bold) which
bears the focus marking (the main stress in the sentence).
Of course, the precise definition of Foc-Extent depends on the general approach
taken to information structure and its relation to other aspects of linguistic
structure within the grammar. We define the Foc-Extent as the set of meaning
constructors corresponding to syntactic elements that are associated with focus.
In the approach introduced in Chapter 10, elements of a sentence’s meaning—i.e.
meaning constructors, which appear in bold in (56)—are categorized according
to their discourse function (df) at s(emantic)-structure, and consequently belong
to the relevant set (for example topic, focus) at the level of i(nformation)-
structure.
is in fact only the type of the event that is in focus: the occurrence of some event is presupposed. On
this, see Mycock (2006). For the purpose of illustrating the analysis of prosodic focus marking in LFG,
we make the simplifying assumption that a single meaning constructor is associated with a verb, and
thus that the verb is associated with a single discourse function at information structure. If one were to
adopt a neo-Davidsonian approach to verb meaning instead, the distinction between event and event
type could be captured using the s-structure attributes rel and event (see §9.10.5.2), and these two
aspects of verb meaning could be analyzed as belonging to different information structure categories.
22 Of course, information structure may also be indicated at other levels of structure, for example
c-structure, but our concern here is with prosodic marking.
428 prosodic structure
In (56), Anna is the subject of a clause, and is also the topic at i-structure by virtue
of appearing in the specifier of IP. Key to this analysis is the attribute-value pair df
topic included in the s-structure for ‘Anna’, a𝜎 . This information, combined with
the annotations on the terminal node and the specifier of IP node in the c-structure,
serves to categorize the relevant meaning constructor as belonging to the topic set
in the clause’s i-structure, f𝜎 ι :
(56) IP
NP
(f a
I
f ι =a ι
((a f : a:
N
a :
Anna
(a f ι: [Anna] }
[Anna] ∈ (a ι (a
The Foc-Exponent, captured by the rule in (58), serves to delimit the Foc-Extent.
Foc-Extent is expressed via separate c-structure rules which, together with the
p-structure rule in (58) and the principle of Interface Harmony, play an equally
important role in the analysis of prosodic focus marking as an interface phe-
nomenon.
In the case of the sentence in (53), the focus constituent in the answer sentence
(Norman) could be identified with either the NP, the N! , or the N which dominates
the word Norman in the c-structure. Following Mycock and Lowe (2013), we
assume here that the phrase in question is the NP, since focused constituents often
correspond to XP categories, but it makes no difference to the analysis of the present
example.
We propose the following node annotation principle for Extent of Focus in
English:
(59) Any c-structure node may optionally be annotated with the following two
specifications:
DFFoc ∈ (⇙l)
DFFoc ∈ (⇘r)
These constraints specify that the label “DFFoc” appears as a member of the
set value of the features l in the leftmost string element (⇙l) and r in the
24 As with PolarInt above, note that DFFoc is simply a label; it does not situate discourse features in
p-structure.
25 We specify and require harmony for only the r features in the s-string and p-string. This reflects
the fact that focus marking in English is fundamentally right-edge based. We make no claims as to
whether or not the same is true for other languages.
430 prosodic structure
(60) a. inclLeft(_α) ≡ _α ∈c (• l)
b. inclRight(_α) ≡ _α ∈c (• r)
c. sdf(_α) ≡ (↑𝜎 df) = _α
d. nsdf(_α) ≡ (↑𝜎 df) /= _α
The template inclLeft ensures that its argument _α is a member of the l set of
the current string element, and similarly for the template inclRight. The template
sdf specifies that its argument _α is the value of the df attribute of the s-structure
↑𝜎 , and nsdf specifies that the df value of ↑𝜎 is different from _α. These basic
templates are used in the definition of the templates lfocus and rfocus:
Recall that we use the up arrow ↑ in lexical entries to refer to the c-structure node
accessible from an s-string element via the π−1 relation (§11.4.1):
NP
(62)
DFFoc ∈ (⇙l)
3 4 3 4
N pred ... df focus
𝜙 𝜎
π−1
53 46
l DFFoc
Similarly, rfocus holds of a string element if “DFFoc” is a member of its r set, and
the df value of its mother’s semantic structure is focus:
NP
(63)
DFFoc ∈ (⇘r)
3 4 3 4
N pred ... df focus
𝜙 𝜎
π−1
53 46
r DFFoc
(64) ΣFoc ≡ { Σ Σ⃰ Σ | Σ }
@lfocus @sdf(focus) @rfocus @lfocus
@rfocus
The rule in (65) permits one or more sequences consisting of any number of
string elements that are not associated with focus at s-structure (by the template
@nsdf(focus)), optionally followed by a sequence fitting the definition of ΣFoc ,
followed by any number of string elements that are not associated with focus
at s-structure. That is, for any string in English the rule in (65) licenses one or
more substrings of which every element is associated with df focus at s-structure,
and of which the leftmost and rightmost elements are marked by the appearance
of “DFFoc” in their l or r value sets respectively. Following the approach to
information structure presented in Chapter 10, the association with df focus at
s-structure has the effect that the meaning constructors corresponding to the s-
string unit(s) in this sub-string appear in the focus set at i-structure.
The p-structure rules in (57) and (58), the c-structure constraint in (59), and
the s-string constraint in (65) are the basis for the analysis shown in (66) for the
example given in (53).
In this utterance, the stressed syllable of the word ‘Norman’ /nɔː/ is associated
with the Nuclear Tone (n_tone h). The prosodic phrase structure rule in (58)
may thus apply to the Prosodic Word dominating that syllable, in this case the
final Prosodic Word in the p-structure pronounced /nɔː.mən/. As a result of
(58) applying, the label DFFoc appears as a member of the set value of r in the
syllable which has primary stress within this Prosodic Word, i.e. the first syllable of
‘Norman’. On the syntactic side, the phrase-structure rule in (59) applies to the NP
that corresponds to the s-string element ‘Norman’. By this rule, the label “DFFoc”
must appear in the l and r sets in the corresponding leftmost and rightmost
s-string elements respectively, which in this case are the same element. By the
string constraint in (65), the meaning constructor corresponding to this s-string
element, that is the meaning constructor introduced in the lexical entry for the
name Norman, appears in the focus set at information structure. The principle of
Interface Harmony requires that an s-string element which has a feature “DFFoc”
in its r value set must be associated with a p-string element that has the feature
DFFoc in its r value set. In (66), this requirement is satisfied, and the analysis
succeeds. If, say, the c-structure rule in (59) were to apply not to the final NP in
the c-structure but to the first one (meaning that Anna was in focus at information
structure), the analysis would fail because “DFFoc” would appear in the r set in
the s-string element Anna but would not be associated with a DFFoc feature in a
corresponding p-string element. That is, prosodic and syntactic specifications are
effectively required to match up, such that whichever prosodic element bears the
Nuclear Tone necessarily corresponds to the syntactic element associated with the
meaning constructor(s) that represent the focus at information structure.
prosodic focus marking 433
n :
NP I
ι
VP Norman : n
h ι :
.. ..
. .
N V
V NP
DFFoc ∈ (⇙L)
π
DFFoc ∈ (⇘R)
π
N
π
{ NP, N, DFFoc }
s-string { IP, NP, N }
I , VP, V , V
{ NP, N } {V} IP, I , VP, V , NP, N, DFFoc
ɔː
ə ən
{ } { }
p-string IntP, PhP, PW
PhP, PW
PhP, PW
PhP, PW IntP, PhP, PW
{ } PhP, PW
{ DFFoc }
β β β β β
S S S S S
PW PW PW
( s ⇒
DFFoc ∈ ( s
IntP
Precisely the same set of rules can equally well account for instances of broad
focus, where the Foc-Extent in the c-structure and s-string spans more than one
word. Example (67) shows the analysis for (55a), an example of broad focus in
which the Foc-Extent is hit Norman. The only difference from the structure in (66)
is that the c-structure rule in (59) applies at the VP level, meaning that the label
“DFFoc” appears in the l set in the s-string element corresponding to the word hit.
This in turn means that the meaning constructor for hit appears in the focus set
at i-structure along with that for Norman, whose r set in the s-string contains the
label “DFFoc” due to the annotation DFFoc ∈ (⇘r) on the VP node.
434 prosodic structure
n :
NP I h :
ι
Norman : n
VP
h ι : hit : h
DFFoc ∈ (⇙L)
.
.. ..
DFFoc ∈ (⇘R) .
N
π
V NP
π N
π
{ NP, N }
s-string { IP, NP, N }
I , VP, V , V, DFFoc
{ NP, N } {V} IP, I , VP, V , NP, N, DFFoc
ɔː
ə ən
{ } { }
p-string IntP, PhP, PW
PhP, PW
PhP, PW
PhP, PW IntP, PhP, PW
{ } PhP, PW
{ DFFoc }
β β β β β
S S S S S
PW PW PW
( s ⇒
DFFoc ∈ ( s
IntP
In this chapter, we examine the place of morphology in the LFG architecture. Work
on the morphology-syntax interface in LFG (notably by Sadler and Spencer 2001
and Kaplan and Butt 2002) assumes a modular view of the morphological compo-
nent, in line with the overall modular architecture of LFG: the morphological com-
ponent has its own internal structure and obeys universal and language-particular
constraints on word formation that need not be shared by other levels of structure.
As Sadler and Spencer (2001) point out, this view accords well with Beard’s
Separation Hypothesis (Beard 1988, 1995), which argues for a separation between
the representation of the phonological form of a word and the features encoding
the morphosyntactic and morphosemantic contribution that the word makes.
Following Sadler and Spencer (2001), Kaplan and Butt (2002), Spencer (2006,
2013), and many others, we assume that the morphological component associates
a word form with a set of features representing the morphological structure and
grammatical contribution of the word, and that these features are interpreted at the
interface of morphology with the rest of the grammar, producing the lexical entry
for the word form. This view fits well with a realizational theory of morphology1
as proposed by, among others, Anderson (1992), Stump (2001, 2002, 2006, 2012),
and Brown and Hippisley (2012), though, like Sadler and Spencer (2001), Kaplan
and Butt (2002), and Andrews (2005), our proposals are compatible not only
with explicitly paradigm-based models, but with any morphological theory which
relates words to feature sets encoding their grammatical properties and structure,
including finite state theories of morphology (Kaplan and Kay 1994; Beesley and
Karttunen 2003). Our work builds on and extends work on morphology and LFG
by Butt et al. (1996a), Butt et al. (1996b), Frank and Zaenen (2002), Sadler and
Nordlinger (2004), Andrews (2005), Otoguro (2006), Bond (2016), and especially
Sadler and Spencer (2001) and Kaplan and Butt (2002).
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
the morphological component and the lexicon 437
our definition of the lexical entry, and we then show how the morphological
component, together with a theory of the interface between morphology and the
rest of the grammar, produces the complete set of lexical entries for a language.
A note about terminology is in order. Morphologists often use the term “lexical
entry” to refer to information associated with a lexeme rather than a word form,
under the assumption that the contribution of a word form depends on the
contribution of the lexeme, specified in the lexical entry for the lexeme, together
with the features of the particular word form. We follow common LFG practice
in using the term “lexical entry” to refer instead to a word form (here, the plural
noun dogs) and its associated grammatical information, as in (1). We use the term
lexemic entry to refer to a lexeme and its associated grammatical information. The
aim of this chapter is to relate the lexical entry for a word form to the morphological
feature set associated with the word form by the morphological component, given
its lexemic entry.2
For ease of presentation, and since our focus in this chapter is not on the details
of syntactic representation or semantic composition, we will assume simplified
f-descriptions containing only certain f-structure features, and omitting many
details of f-structural constraints as well as semantic and other non-f-structural
2 In this chapter, we describe the task of producing the lexical entry for a word form, including its
f-description, on the basis of its morphological features. Importantly, this is only for ease of exposition,
and there is no directionality intrinsic to the theory; our overall aim is simply to define and constrain the
relation between lexical entries and lexemic entries for a language in light of a theory of morphological
realization for the language. Indeed, the opposite order of exposition is often adopted in the setting of
Paradigm Function Morphology (Stump 2001), where a mapping from f-descriptions to morphological
features is described.
438 the interface to morphology
information. Thus, we will work with a simplified entry such as the one in (2),
which we will use as a stand-in for the full entry in (1):
The lexical entry in (2) encodes the same syntactic information as the traditional-
style LFG lexical entry in (3), but also contains information about the word’s
phonological form p-form.
Kaplan and Butt (2002)3 propose an alternative way of representing lexical entries,
by means of a relation which they call M; to avoid confusion with the function
M from c-structure nodes to their mother nodes in the tree, we call this relation
L. For Kaplan and Butt, L is a three-place relation involving a word form, a
c-structure category, and an f-description: they propose the representation in (4),
incorporating the same information as the lexical entry in (3).
Rewriting the lexical entry in (2) in this format, the L relation for the plural noun
form dogs is:
(6) The lexical entry in (2) as an L relation:
L<dogs,
/dɔgz/,
N,
{(↑ pred) = ‘dog’, (↑ index num) = pl}>
3 We follow the presentation in the unpublished Kaplan and Butt (2002) draft paper, which differs
slightly from the version in the handout for the talk presented at the LFG02 conference.
the morphological component and the lexicon 439
The representation in (6) encodes exactly the same information as the representa-
tion in (2), but facilitates the discussion of the definition of L in the following. Our
task, then, is to provide a way of defining the set of lexical entries L for a language,
given an independently specified set of lexemic entries and an independently
specified theory of morphological realization for the language.
In sum, the aim of this chapter is to elucidate the interface between the morpho-
logical component and the rest of the grammar, and to demonstrate how the set of
lexical entries for a language is produced, given a set of lexemic entries and a theory
of morphological realization for the language. We will provide a general framework
which allows exploration of alternative theories of morphological realization and
its interface with the rest of the grammar in an LFG setting.
Our approach is broadly compatible with Spencer’s view, though it is clear that
word forms may be associated with information of other types as well: for example,
the presence of certain affixes might mark a particular information structure
role (Dalrymple and Nikolaeva 2011). Although the difference between the SYN
and SEM components of the lexemic entry is important to the theory of lexical
relatedness proposed by Spencer (2013), our primary concern in this chapter is to
demonstrate how a realizational theory of morphology can be integrated into the
architecture of LFG. Hence, in our discussion of the morphological component
and its place in the grammar, we do not enforce a distinction in the lexemic entry
between syntactic, semantic, and other kinds of information.
We follow Spencer (2013) in assuming that each lexemic root is associated
with a unique identifier, its Lexemic Index (LI).4 The LI cannot be equated
with the semantic form, at least if we assume that the semantic form defines
subcategorization requirements, since morphologically related forms with the
same LI may have different subcategorization frames. Even if we assume that
the LI corresponds to a subpart of the semantic form, the fn (see §5.2.2.3), it
is unclear whether identity of lexemic index for two different word forms has
any consequences for compositional syntax or semantics, and thus whether the
morphological derivational history of a lexeme should be encoded in the semantic
form. For clarity, we provide mnemonic LIs such as dog1 for the lexeme related
to the noun forms dog/dogs, but we assume that the LI plays a role only within
the morphological component, and is not visible to any other component of the
grammar.
A lexemic entry is, then, a three-place relation LE involving (i) the form of the
root and any non-predictable stem forms, together with morphological features
encoding information about inflectional class, category, and other unpredictable
features; (ii) an f-description that encodes syntactic, semantic, and other informa-
tion associated with all word forms of the lexeme, filling the role of Spencer’s SYN
and SEM; and (iii) the Lexemic Index, which we abbreviate as LI:5
For example, we assume that the lexemic entry that is relevant in the analysis of
the plural noun dogs can be represented as follows, using the LI dog1, and leaving
the morphological features unspecified for the moment:6
(9) Lexemic entry for the lexeme dog1 with morphological features unspecified:
LE <morphological features, {(↑ pred) = ‘dog’}, dog1>
The f-description that is associated with all forms of the lexeme dog1 is
{(↑ pred) = ‘dog’}: this is the portion of the f-description that the word forms
dog and dogs of this lexeme have in common. We now turn to a discussion of
the morphological features which we expect to find as the first argument of this
lexemic entry.
Following Sadler and Spencer (2001) and Spencer (2006), we use the term m-
feature for morphological features, and we use the m-prefix for particular m-
features.
There are several types of m-features. Features that are relevant only within
the morphological component are morphomic features. A standard example of a
morphomic feature is inflectional class. Inflectional class is specified as a part of
5 The lexemic entry relation LE is similar to what Andrews (2005) refers to as the lookup function,
for which he uses the symbol L.
6 Recall that for simplicity, we use the pred feature and its value to stand for the full f-description
associated with a root. In a full treatment, the f-description in the lexemic entry would include more
information, including (at least) the semantic content of the root: [dog] ∈ (↑𝜎ι (↑𝜎 df)), with an
appropriate definition of the meaning constructor abbreviated as [dog].
the morphological component and the lexicon 441
the lexemic entry, and we will see that it can play a role in determining the lexical
entry for a word form.
Besides morphomic features, some m-features are relevant in determining the
syntactic contribution of a word form: these are morphosyntactic m-features,
relevant in defining c-structure or f-structure relations and constraints. Others are
associated with morphological signaling of meaning: these are morphosemantic
m-features. Still others are relevant at other levels of structure: for example, a word
form may bear morphological marking to indicate its information structure role.
A single m-feature can also play multiple roles, determining the contribution of a
word form at more than one level of structure.
Thus, the lexemic entry in (10) associates the lexeme dog1 with an m-feature
m-root which specifies the root form of the lexeme dog1, an m-feature m-
iclass which specifies inflectional class, and an m-feature m-cat which specifies
the lexical category of the lexeme. We do not rely on a particular theory of
morphological realization or inflectional classes, and so we use the arbitrary integer
1 to represent the inflection class (m-iclass) of the lexeme dog1.
The lexemic entry also includes all of the unpredictable stems for a lexeme. For
example, the lexemic entries for the verb roots print and drink are given in (11);
drink differs from print in that its past tense form drank and past participle form
drunk are irregular, and must be lexically specified. Again, we provide arbitrary
inflectional class information for these lexemes:7
7 Unlike the lexemic entry for dog in (10), only the pred fn value for the verbs print and drink is
specified in (11). This is because the lexemic entry for a verb like print or drink produces both active and
passive lexical entries, with different semantic forms, and the correlation between passive morphology
and passive argument structure must be maintained: the semantic form for the active version of print
is ‘print#subj,obj$’, and the passive version is ‘print#oblagent ,subj$’. The f-description for a verbal
lexical entry, including the full semantic form linking thematic roles with grammatical functions, is
specified as required by mapping theory (Chapter 9) on the basis of its functional features, including
voice features. For an explanation of the structure of semantic forms and reference to parts of a
semantic form by features such as fn, see §5.2.2.3.
442 the interface to morphology
theories, we are agnostic about the details of the content and representation of m-
features encoding purely morphological information.
To summarize, every language specifies a set of lexemic entries LE which asso-
ciate a Lexemic Index with a set of morphological m-features and an f-description.
The detailed content of the m-features is dependent on the particular theory of
morphological realization that is assumed, and we place no further constraints on
the form of individual lexemic entries and the relations between them.
To define the lexical entries for the word forms of a language, we require access
to the complete set of m-entries as defined by the morphological realization
component R.8 The realization relation R for a language accounts for all aspects
of the morphological realization of word forms in the language, encompassing a
8 We do not take a position on whether some or all lexical entries for morphologically invariant
forms are stored separately, rather than being defined on the basis of m-entries for those forms: for
example, whether there is an m-entry for the English determiner the from which the lexical entry for
the is derived. It may be that such forms are in fact stored as underived lexical entries, and do not play
a role in the morphological component.
the morphological component and the lexicon 443
D is what Kaplan and Butt (2002) call the “description function,” and what Sadler
and Nordlinger (2004) call a “lexical transducer,” relating m-features and their val-
ues to grammatical specifications. In a Paradigm Function Morphology setting, D
corresponds to what Stump (2002) calls rules of paradigm linkage, or (roughly) the
correspondence relation between the form paradigm (pairing a lexeme with a set
of m-features) and the content paradigm (pairing a lexeme with an f-description)
in work by Stump (2015).
In §12.1.4, we saw that the realization relation R produces an m-entry which
associates the set of m-features {m-cat:noun, m-num:pl} with the s-form dogs,
the p-form /dɔgz/, and Lexemic Index dog1, as shown in (13). For the plural noun
dogs, D interprets the m-features {m-cat:noun, m-num:pl}, mapping them to the
c-structure category N and the f-description {(↑ index num) = pl}:
(15) D <dog1,
{m-cat:noun, m-num:pl},
N,
{(↑ index num) = pl}>
9 Marcotte and Kent (2010) explore the possibility of eschewing the D mapping and requiring the
realizational component R to produce a full f-description rather than a set of m-features. We side
with Sadler and Spencer (2001), Kaplan and Butt (2002), and many others in recognizing the need to
distinguish between the m-features relevant within the morphological component and the f-description
that is relevant outside it.
444 the interface to morphology
Legitimate members of L are those which meet the conditions imposed by LE, R,
and D. First, a lexemic entry LE must provide a partial f-descriptionL associated
with a Lexemic Index LI and a set of m-features. Second, the morphological
realization relation R must include an m-entry relating the LI and the m-features
to an s-form and a p-form. Third, the description function D must map the m-
features to a category and f-descriptionM . The f-description for the lexical entry
is, then, obtained by combining (via set union, ∪) the f-descriptionL intrinsic to
the lexeme and the f-descriptionM obtained from the m-features for the particular
word form in question. We can represent the requirements in (16) graphically, as
in (17):
10 The definition in (16) characterizes a set by describing its elements, as discussed in Foot-
note 14 of Chapter 6. In (16), L is a set whose elements have the form <s-form, p-form,
category, f-descriptionL ∪ f-description M >; elements of that form are included in L if they satisfy the
description in the next three lines.
the morphological component and the lexicon 445
As advocated by Kaplan and Butt (2002), m-features and the Lexemic Index are
crucial in defining the description function D, but neither the Lexemic Index nor
any of the m-features appear in the lexical entries L which are accessible to the rest
of the grammatical system. It is only the information in the lexical entries in L—
the s-form, p-form, c-structure category, and f-description—that is visible outside
the morphological component. This maintains a clean separation between the
morphological component and the other components of the grammar. In this way,
our proposal aligns itself with the Principle of Morphology-Free Syntax (Pullum
and Zwicky 1988; Zwicky 1992), and contrasts with proposals that reject the
Lexical Integrity Principle, including Distributed Morphology (Halle and Marantz
1993; Embick and Noyer 2007) and the Exo-Skeletal Model (Borer 2013).
The set of lexemic entries LE and morphological realization relation R are indepen-
dently specified for each language, in conformance to a suitable realizational theory
of morphology. In the remainder of the chapter, our focus is on the definition of the
description function D, which constitutes the relation between morphology (sets
of m-features) and the information and constraints that are relevant for the rest of
the grammar.
446 the interface to morphology
In their analysis of auxiliary selection, Butt et al. (1996a) and Butt et al. (1996b)
propose a new projection from c-structure, morphosyntactic structure. Building
on this work, but adopting a slightly different projection architecture, Frank and
Zaenen (2002) propose an alternative analysis in which morphosyntactic structure
is projected from f-structure rather than c-structure, and which also requires
the use of complex c-structure categories. As noted in §7.1.1, we do not adopt
morphosyntactic structure as a level of representation elsewhere in this book;
we present these two analyses here simply as illustrations of how the D-mapping
works, and in particular to show how different definitions of the D-mapping from
the same m-features give rise to different lexical entries, depending on the syntactic
analysis that is adopted.
Butt et al. (1996a) and Butt et al. (1996b) argue that Affix Hopping patterns
should not be treated in terms of complementation or embedding at functional
structure. Instead, they propose to encode verbal dependencies at morphosyntactic
structure, which on their approach is related to c-structure through a projection
function μ, represented in (21) as a dashed arrow:
11 §2.5 provides more discussion of the structure and representation of the values of features such as
case and num; here we provide simplified atomic values such as acc and pl.
448 the interface to morphology
(21) be swimming according to Butt et al. (1996a) and Butt et al. (1996b):
! $
vtype inf
" # (Morphosyntactic structure)
dep vtype prs.ptcp
μ
V'
V VP
% &
ˆ⃰μ = ⃰μ (ˆ⃰μ dep) = ⃰μ pred ‘swim#subj$’
↑=↓ ↑=↓ (F-structure)
aspect prog
be V
swimming
For clarity, we have provided the same value prs.ptcp for the m-feature m-vtype
in (23) and the vtype feature at morphosyntactic structure in (21). Impor-
tantly, these features and their values have a very different status: the m-feature
m-vtype:prs.ptcp in (23) is visible and relevant only within the morphological
component, and plays a crucial role in defining the lexical entry for the word form
swimming, while the vtype feature in (21) is relevant in defining and constraining
syntactic dependencies involving verbs.
Given the m-entry in (23), Butt et al.’s analysis requires the morphosyntactic
m-vtype feature that is contributed by swimming to correspond to a vtype
feature at morphosyntactic structure via the μ mapping, and an aspect feature at
f-structure; this allows the required dependencies to be established:
Butt et al.’s analysis also requires the m-cat feature for verbal forms to map to the
phrase structure category V:
(25) D mapping for phrase structure category according to Butt et al. (1996a,b):
D
m-cat:verb ⇒ V
In this way, Butt et al.’s analysis enforces “affix hopping” requirements at mor-
phosyntactic structure while maintaining a simple, monoclausal f-structure with
no embedding.
It is possible for the same theory of morphological realization, and in particular
the same m-entry in (23), to underpin an analysis of the English auxiliary system
that makes very different syntactic assumptions. Frank and Zaenen (2002) provide
an alternative analysis of auxiliary selection, illustrated primarily for French, which
assumes a different relation between morphosyntactic structure and f-structure,
and which also appeals to complex c-structure categories; see § 3.3.5 and §5.1.4 for
definition and discussion of complex categories, including parametrized complex
categories. Frank and Zaenen’s proposal differs from Butt et al.’s in the inventory of
features that appear in morphosyntactic structure: according to Frank and Zaenen,
morphosyntactic structure contains a number of additional features not treated
by Butt et al., including person, number, gender, and case. Constraints on these
features are generally stated in terms of f-structure relations rather than c-structure
configuration: verbs show agreement in person, number, and/or gender with their
subjects or objects, for example, and case is often an indication of f-structure
role. According to Frank and Zaenen (2002), then, morphosyntactic structure is
450 the interface to morphology
best treated as directly related to f-structure rather than c-structure, since this
allows easy reference to f-structural relations in characterizing morphosyntactic
dependencies such as agreement and case assignment. For a discussion of the
grammatical architecture that Frank and Zaenen assume, see § 7.1.1.
Frank and Zaenen (2002) redefine μ as a function from f-structures (rather than
nodes of the c-structure) to morphosyntactic structures:
V[fin,be] VP[nonfin,prs.ptcp]
μ
be V[prs.ptcp]
swimming
(Morphosyntactic
! structure) $
vtype inf
" #
dep vtype prs.ptcp
Since Frank and Zaenen assume that morphosyntactic structure is directly related
to f-structure and only indirectly related to c-structure, their approach cannot
take advantage of c-structural embedding relations to control auxiliary sequencing.
Instead, Frank and Zaenen propose a combination of constraints involving mor-
phosyntactic features and complex c-structure categories to constrain the position
and morphological form of verbs in the verbal complex.
As discussed in §3.3.5 and §5.1.4, complex categories allow certain features to
be realized as a component of a complex c-structure category, and parametrized
rules allow for these features to be passed through the tree. For example, the
parametrized phrase structure rule in (28) states that a V' with a finiteness value
represented as _ftness dominates a V with form ‘be’ and the same finiteness value
_ftness as the V' , followed by a VP headed by a nonfinite present participial
verb. This rule ensures that progressive be is followed by a present participial
complement, as in be swimming:
morphological features and morphological classes 451
(28) V' phrase structure rule according to Frank and Zaenen (2002):
V' [_ftness] −→ V[_ftness,be] VP[nonfin,prs.ptcp]
On Frank and Zaenen’s approach, then, be and swimming have the lexical entries
in (29), containing specifications of vtype for be and dep vtype for swimming at
morphosyntactic structure, as well as complex c-structure categories:
To derive these lexical entries, we again assume the m-entry for swimming given in
(23). However, Frank and Zaenen’s approach requires a different D-mapping from
m-features to f-descriptions. On their approach, the description function D must
establish the following correspondence between the set of m-features for swimming
and a complex c-structure category:
As this example illustrates, more than one m-feature may be involved in a sin-
gle mapping relation: here, the m-features {m-cat:verb, m-vtype:prs.ptcp} are
involved in a single D-mapping specification to the complex category V[prs.ptcp].
Frank and Zaenen’s analysis also involves constraints on morphosyntactic struc-
ture and f-structure. This means that the same m-vtype feature as in (30) maps to
the following f-description:
Note, then, that the same m-features can be relevant in two different aspects of the
D-mapping: here, the same m-feature, m-vtype, is involved in determining both
a complex c-structure category (30) and an f-description (31).
inspecting the paradigms for the noun stolovaja ‘dining room’, the adjective bol' šoj
‘big’, and the noun lampa ‘lamp’:
Given these m-entries, the D-mapping rule in (34) produces the correct category
for these verb forms; notably, for the mixed category stolovaja with m-feature
m-class:mixed-a-n, the m-cat:adj m-feature maps to the c-structure category
N:13
12 The term “mixed category” is also used in a different sense, to refer to categories such as the English
gerund, which exhibit both nominal and verbal syntactic characteristics simultaneously (Bresnan 1997;
Bresnan and Mugane 2006; Seiss 2008; Lowe 2017a). “Mixed lexical categories” in Spencer’s sense are
categories which have morphological characteristics of one category, but syntactic characteristics of
another category.
13 If we assume a morphological theory which makes use of privative features and defaults, the
mapping for the regular nouns bol' šaja and lampa could be treated as a default mapping (§12.2.6).
morphological features and morphological classes 453
12.2.4 Deponency
Börjars et al. (1997) and Sadler and Spencer (2001) discuss the Latin verb loquor
‘I speak’, which has the morphological form of a passive verb but is syntactically
active, as can be seen by comparing the forms of loquor, with active voice, to the
active and passive forms of lego ‘I read’. This verb illustrates deponency (Baerman
2007), a mismatch in which a word has a form implying a particular category (here,
passive voice) but realizing a different category (active voice):
The D-mapping rules given in (37) produce the correct voice specification for these
word forms: active voice for the deponent verb loquitur, and passive voice for the
regular verb legitur:
(37) a. Passive form, active voice for deponent verbs (for example, loquitur):
D
m-voice:pass ⇒ (↑ voice) = active in the presence of the m-feature
m-class:deponent-voice.
b. Passive form, passive voice for regular verbs (for example, legitur):
D
m-voice:pass ⇒ (↑ voice) = pass in the presence of the m-feature
m-class:regular.
(39) a. New needle-free measle vaccine ‘could save thousands of children’s lives’
(headline in The Telegraph, 17 Aug 2009)
b. . . . it is reasonable because though we have never found a measle
germ associated with measle-symptoms we have in cases with like
symptoms found, not indeed measle germs, but things of the same sort . . .
(Wisdom 1968)
One way to capture Acquaviva’s analysis in our setting is by assuming that the m-
feature m-num:pl maps to the syntactic feature (↑ index num) = pl in most cases,
but with particular lexically specified exceptions involving reference to particular
lexemic entries: if the LI is measle1 or a handful of other syntactically singular
nouns with plural morphology, m-num:pl maps to (↑ index num) = sg. Such an
analysis requires the following m-entries and D-mapping rule:
morphological features and morphological classes 455
This analytic possibility may not arise, depending on the particular theory of
morphological realization and the morphology-syntax-semantics interface that
is adopted. There are two alternative analytic possibilities that do not involve
reference to the LI. First, it may be preferable to analyze measles as a root that
belongs to a particular (very small) inflectional class, consisting of deponent nouns
that are morphologically plural but syntactically singular. Under this analysis, the
m-feature m-class for this class, rather than the LI for the lexemic entry, would
be important in the D-mapping for measles. Second, if the realization relation R
specifies measles as having m-num:sg even though it bears plural morphology, the
definition of D need not take the LI into account. If alternative accounts which do
not involve reference to the LI are shown to be available and preferable for all word
forms, we need not include the LI as a component of the description function D.
On the other hand, if the interpretation of a set of m-features is best analyzed as
varying according to the LI, we must take the LI into account in our definition of
D. Future work will show which hypothesis is viable, and whether the LI must be
taken into account in at least some cases of the D-mapping.
14 Note that we neither advocate nor reject the use of privative features or defaults in defining the
D-mapping relation; whether or not such rules are needed depends on the theory of the morphological
realization relation R that is adopted and the syntactic rules and constraints for the language. Privative
features are not used in Paradigm Function Morphology (Stump 2001), and so feature default rules of
this type are not relevant in a PFM setting.
456 the interface to morphology
Under these assumptions, we require a default rule which applies when the m-num
feature is absent:
In §12.3.2.3, we will see how to express default rules within the formal framework
that we adopt.
Here, a set of m-features is relevant for subject agreement for inflected verbs,
encoding first person plural subject agreement. Thus, the D mapping must map
a complex m-feature of this form to the following f-description:
Kaplan and Butt propose that the description function D maps from individual
m-features to component parts of the f-description for a word form:
(47) Kaplan and Butt’s (2002) description function D for each m-feature:
D(m-root:Katze) = {(↑ pred) = ‘Katze’}
D(m-cat:noun) = {N, (↑ ntype) = count}
D(m-gend:neut) = {(↑ concord gend) = n}
D(m-num:sg) = {(↑ index num) = sg}
D(m-num:pl) = {(↑ index num) = pl}
D(m-case:nom/dat/acc) = {(↑ concord case) ∈ {nom,dat,acc}}
D(m-case:nom/gen/dat/acc) = {(↑ concord case) ∈ {nom,gen,dat,acc}}
458 the interface to morphology
They then define the description function D for the set of m-features as the union
(∪) of the result of applying D to each feature independently:
(48) Kaplan and Butt’s (2002) description function D for a set of m-features:
D({d1 , d2 , . . . , dn }) = D(d1 ) ∪ D(d2 ) ∪ …D(dn )
12.3.2 Definition of D
Recall that the D-mapping is a relation involving the following terms:
The D-mapping for the plural noun dogs given in (15) is repeated in (50). It relates
the m-features {m-cat:noun, m-num:pl} to the c-structure category N and the
f-description {(↑ index num) = pl}:
(50) D <dog1,
{m-cat:noun, m-num:pl},
N,
{(↑ index num) = pl}>
The c-structure category for a word form is specified by the Dcat mapping on
the basis of the L(exemic) I(ndex) and the m-features, and the f-descriptionM is
determined by the mapping defined by Dfeats . These mappings are defined on a
language-by-language basis, though there is likely to be a great deal of commonality
15 Andrews (2005) also extends his analysis to case stacking, discussed in §12.3.3.6.
the description function d 459
12.3.2.1 Dcat : defining the c-structure category If we assume that English does
not make use of complex c-structure categories, and furthermore that there are
no “mixed” categories in English with differing morphological and syntactic
categories, the definition of Dcat for English is:
An informal rule was given in (34a) for a complex category which is morphologi-
cally an adjective but syntactically a noun. That rule can be restated in the current
format as in (54):
more than one, all of the m-features are passed on to Dfeats for determination of
the f-description of the word form.
In this definition, Dfeats (with four arguments) relates a L(exemic) I(ndex), a set of
m-features {m1 , m2 , . . . mn }, an f-structure f , and an f-description d1 ∪ d2 ∪ . . . dn
corresponding to those m-features. Dfeats is defined in terms of two other func-
tions: Ddefault defines a set of default features d0 which appear when certain m-
features are absent, and the Dfeat functions map individual m-features m1 . . . mn to
f-descriptions d1 . . . dn . That is, Dfeat (with five arguments) specifies the mapping
relation for a particular m-feature (for example, m1 ) in the context of the LI, the
complete set of m-features {m1 , m2 , . . . mn }, and a target f-structure f , mapping
m1 to the f-description d1 . According to this definition, then, each m-feature
is considered in the context of all of the other m-features, and the appropriate
f-description of the f-structure f is added to the full f-description. This requires
a definition of Ddefault for the language and of Dfeat for each m-feature, as we now
show.
The definition of Dfeats is formulated so as to allow the D-mapping to specify
constraints on any relevant f-structure. Usually, the relevant f-structure is ↑, as
specified in (51); we discuss the D-mapping for complex m-features and case
stacking in §§12.3.3.5 and 12.3.3.6, where the f-description may refer not to ↑,
but to another f-structure related to ↑.
17 Privative features are not assumed in Paradigm Function Morphology (Stump 2001). When
such theories of morphological realization are adopted, the Ddefault function is not relevant, and the
definition of Ddefault which we provide in this section can be ignored.
the description function d 461
following, where the default f-description (d0 in the definition in (55)) results from
applying each of the n rules in turn:
Notice that we have specified f as the relevant f-structure in the third argument
position of the definition of Ddefault and, in turn, of the individual default rules
D1default . . . Dndefault , to allow the D-mapping to impose constraints on the rele-
vant f-structure, whether it is ↑ or another f-structure related to ↑. In most cases,
the relevant f-structure is ↑, as specified in the definition in (51), but for complex
m-features or case stacking another f-structure might be relevant, as we show in
§§12.3.3.5 and 12.3.3.6.
Schematically, individual Ddefault rules are stated as in (57); in this definition,
the absence of the m-feature m1 licenses the introduction of the f-description d1
as a constraint on the f-structure f . In other words, if the privative feature m1 is
absent, d1 appears in the default f-description of f ; if m1 is present, d1 does not
appear, and the empty f-description ∅ is introduced:18
As with the Dcat rules, the Ddefault rules are not involved in “feature accounting”:
once the default rules have applied to a set of m-features, that set is passed
unchanged to the Dfeats rule.
This rule does not require or disallow the presence of other m-features in defin-
ing the mapping between m-case and syntactic case; m-case:nom maps to the
f-description {(f concord case) = nom} for the f-structure f independent of
the presence or absence of other m-features, and similarly for the other case
possibilities.
If the mapping between morphological and syntactic case is completely straight-
forward, we can make use of a convenient abbreviation capturing the effects of each
of the four separate D-rules in (59), borrowing the underscore notation for the
argument of a parametrized template (§6.7) to indicate that morphological case
always matches syntactic case:
19 As we have done so far, we use simple values like nom and acc as the value of the case attribute.
Section 2.5.7.5 provides a full discussion of the representation of the values of the case feature.
the description function d 463
For example, there may be no mapping to an f-description for the m-iclass feature
representing inflectional class, though that feature may be important in providing
the proper context for other D-mapping rules, as outlined in §12.2.4 and §12.3.3.2:
(63) Example: D-mapping to the empty f-description for the m-iclass:1 and
m-iclass:2 m-features
Dfeat <LI, m-iclass:1, m-features, f , ∅>.
Dfeat <LI, m-iclass:2, m-features, f , ∅>.
(64) Example: D-mapping to the empty f-description for the m-cat m-feature
with any value, using an abbreviatory convention mentioning only the
m-cat feature
Dfeat <LI, m-cat, m-features, f , ∅>.
The D-mapping rule in (64) maps the m-cat attribute with any value to the empty
f-description ∅.
(65) Schematic mapping for a lexical exception dependent on the Lexemic Index:
Dfeat <LI, m1 , m-features, f , d1 >
if and only if LI = l1 ,
otherwise Dfeat <LI, m1 , m-features, f , d2 >.
To align more closely with the m-feature representations we have been assuming,
we reformulate the complex m-feature in (67) in the following equivalent way:
In this reformulation, the value of the m-agr m-feature is a pair in which the first
argument su is a morphological marker providing information about the syntactic
role of the relevant f-structure for the complex feature specification, and the second
the description function d 465
(70) Example: Mapping between the morphological marker su and the gram-
matical function subj
Dfeat <LI, su, m-features, subj, ∅>
The rule in (70) states a correspondence between su and the grammatical function
subj in the presence of a set of m-features, but does not contribute an f-description.
We can make use of the mapping in (70) in the following D-mapping rule:
In (71), the relevant f-structure is f , and the value of the complex m-arg m-
feature is a pair whose first argument is the morphological marker su. The relation
between su and the grammatical function subj is specified in the D-mapping rule
defined in (70). The second argument is a set of m-features, {m-pers:1, m-num:pl},
which map to the f-description constraining the relevant f-structure. The rule in
(71) states that the set of m-features specified as the second member of the pair
should be interpreted as constraining the Dfeats mapping with respect to the subj
attribute in the f-structure f . That is, the relevant f-structure is (f subj), and the
Dfeats mapping requires the f-description to hold of (f subj).
This rule is a specific instance of the following general pattern:
According to this rule, the value of a complex m-feature is a pair in which the first
argument specifies a morphological marker attr and the second argument is a set
of m-features. The Dfeat mapping for a complex feature with respect to f-structure f
is defined as the Dfeats mapping for the set of m-features, where the morphological
marker attr maps to a grammatical function _gf, and the relevant f-structure for
the constraints specified by the set of m-features is (f _gf).
466 the interface to morphology
(73) thara-ngka-marta-a
pouch-loc-prop-acc
The locative casemarking closest to the root thara ‘pouch’ marks the word as
the head of a locative phrase (‘in (its) pouch’). The proprietive and accusative
casemarking further constrains the syntactic environment within which the phrase
can appear: thara-ngka ‘in its pouch’ must modify a proprietive phrase (‘with X
in its pouch’) which itself modifies an accusatively-marked phrase. According to
Sadler and Nordlinger (2004), the f-structure for the form thara-ngka-marta-a
is the one labeled p in (74), and it must appear in the f-structure environment
shown:20
20 The representation in (74) differs from the proposals of Nordlinger (1998) and Sadler and
Nordlinger (2004) in using the standard set-valued attribute adj for adjuncts, rather than assuming
different adjunct functions indexed by semantic role.
21 A euro is a type of wallaroo, a kangaroo-like animal.
the description function d 467
The f-structure for (75) is given in (76), which meets the requirements given in
(74) for the environment in which thara-ngka-marta-a must appear:
(76)
’
p
Sadler and Nordlinger (2004, 2006b) provide an analysis of case stacking within the
theory of Paradigm Function Morphology (Stump 2001) which, like the complex
m-features discussed in the previous section, appeals to the possibility of an
enriched structure for m-features: the order of the case features in case stacking is
captured by allowing the value of a complex m-feature to consist of an embedded
set of m-case m-features.
The table in (77), based on Sadler and Nordlinger (2004), states the relation
between morphological case and the corresponding f-description. The C and
M subscripts represent a distinction between Core cases, used adnominally and
relationally, and Modal cases, encoding temporal contributions as well as comple-
mentizing and associating functions; for more discussion of this distinction, see
Sadler and Nordlinger (2004, 2006b):
Our goal, then, is to produce the f-description in (78) on the basis of the lexemic
entry for thara, the m-entry for thara-ngka-marta-a, and a D-mapping given the
m-entry.
We propose the following lexemic entry for thara:
The f-description in this lexemic entry provides the first line of the f-description
in (78).
Next, we require an m-entry for the word form thara-ngka-marta-a as pro-
vided by the realizational component R for Martuthunira. We adopt Sadler and
Nordlinger’s insight that the value of the m-case feature in case stacking is best
represented as a recursively embedded set reflecting the order in which the case
affixes appear in the word form:
The task is now to extend the definition of D in such a way that this complex
m-case specification maps to the following f-description, as desired:
D
(81) m-case:{m-caseC :loc, {m-caseC :prop, {m-caseC :acc}}} ⇒
{(↑ case) = loc,
((adj ∈ ↑) concord case) = prop,
((adj ∈ adj ∈ ↑) concord case) = acc,
(obj adj ∈ adj ∈ ↑)}
Unlike the analysis of Sadler and Nordlinger (2004), we do not construct a tree
as a part of the D-mapping; rather, we follow Andrews (2005) in proposing a
recursive definition of D for set-valued features which produces the appropriate
f-description.
22 The f-description in (78) involves inside-out functional uncertainty (§6.1.3) and uses the set
membership symbol ∈ as an attribute (§6.3.1).
the description function d 469
We can now give a slightly lengthier but more revealing restatement of the relevant
f-description, closely following proposals by Andrews (2005) and using local
names for f-structures (prefixed by a percent sign %: see §6.5):
(83) ↑ = %f0
(%f0 concord case) = loc
m-caseC :loc (adj ∈ %f0 ) = %f1
%f1
(%f1 concord case) = prop
{m-caseC :prop} (adj ∈ %f1 ) = %f2
%f2
(%f2 concord case) = acc
{{m-caseC :acc}} (obj %f2 ) = %f3
%f3
The definitions in (84a) and (84b) are relevant for all languages with case stacking.
The definitions in (84c) are particular to Martuthunira, and are specified on a
language-by-language basis.
This chapter explores issues in the syntax and semantics of modification. Since
there is in principle no limit to the number of modifiers that a phrase can have,
we represent modifiers at functional structure as members of a set of modifying
adjuncts adj (§2.4.4). Functional annotations on c-structure rules ensure that at
f-structure, each modifier appears as a member of the adjunct set of the phrase it
modifies.
In this chapter, we concentrate in particular on adjectival modification, since
the syntax and semantics of adjectives is fairly complex and illustrates many of
the issues of interest to us. §13.1 provides an overview of the syntax of adjectival
modification, §13.2 discusses three semantic classes of adjectives and how their
meanings are represented, and §13.3 discusses adjectival modification at the
syntax-semantics interface within the glue approach.
Defining the semantic contribution of a modifier brings up a set of tricky
problems, as first noticed by Kasper (1997). In §13.4, we address these issues and
show that they have a straightforward solution within our framework.
The chapter concludes with a brief examination of the syntax and semantics
of adverbial modification: §13.5 discusses the syntax and semantics of manner
adverbs like skillfully, as well as sentential adverbs like clearly.
(1) a. The girl handed the baby a toy on Tuesday in the morning.
b. ⃰David saw Tony Mr. Gilroy my next-door neighbor.
At f-structure, each modifier is a member of the adj set. In (2), the adjectival
modifier Swedish is treated as a member of the adj set of modifiers of the noun
man:
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
474 modification
In phrases with more than one modifier, the f-structure for each modifier appears
as a member of the adj set:
The lexical entries for the attributive adjectives tall and Swedish and the noun
man contain at least the following syntactic information, including specification of
syntactic form (fm), phrase structure category (the result of applying the labeling
function λ to the terminal c-structure node π(•)), and the value of the pred
feature:1
1 Since we do not provide morphological or prosodic analyses in Part III, we present lexical entries
in the traditional format introduced in §5.3.2, and not the fully-specified format, including the p-form,
which was introduced in §11.4.1.
2 For discussion of postnominal AP modifiers, see Arnold and Sadler (2013).
syntax of adjectival modification 475
NP
N
↑ =↓
N
↑ =↓
A N
↓ ∈ (↑ ↑ =↓
Swedish man
(6) N −→ A
3 N
↓ ∈ (↑ adj) ↑=↓
N
↑ =↓
A N
↓ ∈ (↑ ↑ =↓
tall A N
(↑ ↓ ∈ (↑ ↑ =↓
Swedish man
(↑ (↑
476 modification
This meaning is of type &e → t'. It picks out the set of men—that is, the set of
entities x for whom the proposition man (x) is true. When a meaning like the one
in (8) is modified, the result is a meaning which is of the same type but which
reflects a modified meaning rather than the original unmodified meaning. In this
section, we describe how noun meanings are modified in different ways by different
semantic classes of adjectives.
The type of this meaning is &e → t', just like the unmodified meaning man; the
difference in meaning is that this expression picks out the set of individuals x that
satisfy both the predicate Swedish (x) and the predicate man (x)—the individuals
that are both Swedish and men. Adjectives like Swedish are called intersective, since
the individuals who are Swedish men are those that are in the intersection3 of the
set of individuals that are Swedish with the set of individuals that are men. As Partee
(1995) points out, for intersective adjectives the conclusion in (10c) follows from
the premises in (10a) and (10b):
(10) a. David is a Swedish man. [David is in the intersection of the set of Swedish
individuals and the set of men.]
b. David is a football player. [David is in the set of football players.]
c. Therefore, David is a Swedish football player. [David is in the intersection
of the set of Swedish individuals and the set of football players; that is,
David is both Swedish and a football player.]
3 The intersection of two sets contains all of the members that the two sets have in common: see §6.3.
semantic classes of adjectives 477
Such adjectives are called subsective: the set of big mice is a subset of the set of
mice, but a big mouse is not in the intersection of the set of things that are big in
general with the set of mice. Indeed, it is not clear whether it makes sense to talk
about the set of things that are “big in general”; to the extent that it does, the set
of things that are “big in general” would contain solar systems and galaxies, and
possibly mountains and oceans, but certainly not mice.
There are several kinds of subsective adjectives.4 One type is gradable adjectives
like big or tall, which, as noted by Montague (1970a), Kamp (1975), Siegel (1976),
Kennedy (1997), and many others, must be interpreted relative to some relevant
standard. For example, a particular mouse might count as a big mouse, even though
the same mouse is probably not a big animal or even a big rodent. Similarly, a seven-
year-old boy can be correctly characterized as a tall seven-year-old even if he is not
tall compared to an adult.
We propose the following simplified meaning for big mouse (see Kennedy 1997
for a full discussion of the semantics of gradability and comparison):
The argument P of big represents the contextually salient property that determines
the relevant standard of measurement. If the property P of the individual is that it
is a mouse, modification by the adjective big requires the individual to exceed some
standard of size that is determined by reference to mousehood. In other words, in
order to claim that something is big relative to the property of being a mouse, we
need to know the range of sizes that are usual for mice.
In a neutral setting, the contextually relevant property is often the property
denoted by the modified noun; for example, the contextually salient property P in
an example like big mouse is generally resolved to the property of being a mouse.
However, as pointed out by McConnell-Ginet (1979) and Pollard and Sag (1994),
4 One well-studied type of subsective adjective is exemplified by beautiful in one reading of beautiful
dancer, which can refer either to an individual who is both beautiful and a dancer, or (in the non-
intersective, subsective reading) to an individual who dances beautifully (Siegel 1976). Larson (1998)
proposes to treat this ambiguity by allowing beautiful to specify either the individual variable for dancer,
giving rise to the meaning ‘beautiful and a dancer’, or the event variable associated with the dancing
event, giving rise to the meaning ‘dance beautifully’. Though we will not provide an explicit analysis of
these examples here, such an analysis can be easily formulated in a glue setting by adopting a semantic
representation including event variables, as introduced in §8.10.
478 modification
in certain contexts other interpretations are also possible. Pollard and Sag provide
the following example:
Here the property P relevant to the interpretation of the adjective good is being a
volleyball player, since in this example good linguist means, more or less, linguist
who is good at playing volleyball. Examples such as these show that the property
P need not correspond to the property denoted by the modified noun, but can be
determined contextually.
Of course, modified phrases can undergo further modification. The meaning of
the doubly modified phrase tall Swedish man is:
These adjectives, studied by Kamp (1975) and in more detail by Siegel (1976),
are different from those discussed in the previous section in an important way:
a Swedish man is a man, and a big mouse is a mouse, but a fake gun is not a
gun; instead, it may actually be a toy or a piece of soap. Thus, the meaning of a
phrase with a nonsubsective adjective depends on the meaning of the unmodified
phrase, but the resulting property may hold of an individual even if the unmodified
meaning does not. In other words, the set of individuals referred to by the
modified noun is in general not a subset of the set of individuals referred to by
the unmodified noun: this is why these adjectives are called nonsubsective.
modifiers and semantic composition 479
A former student is one who at some previous time was a student, but who is
no longer a student; the meaning of student is important in understanding the
meaning of former student, but the individuals represented by x in the meaning
given in (17) for former student are not required to be students. Thus, former in (17)
denotes a relation between the property of being a student and some individual
who formerly had that property. Similarly, a fake gun is an entity that is not a gun,
but which has some properties in common (for example, appearance) with entities
that are actually guns; again, although a fake gun is not a gun, the meaning of gun
is important in determining the meaning of fake gun:
Importantly, the resulting meaning still has the same type as the unmodified noun;
nonsubsective adjectives, like intersective adjectives and gradable adjectives, turn
an unmodified &e → t' meaning into a modified &e → t' meaning.6 In fact, the
type of an unmodified meaning is always the same as the type of the modified
meaning, and this is important in our analysis of modification and meaning
composition.
5 Partee (2010) argues that adjectives that have traditionally been classified as nonsubsective are
actually subsective, inducing coercion of the meaning of the modified noun. In Partee’s example (a),
for instance, the denotation of the first underlined occurrence of fur is expanded to include not only
fur, but material resembling fur in some respect, while the second underlined occurrence of fur has its
normal meaning, and refers only to actual fur:
(a) I don’t care whether that fur is fake fur or real fur.
Adopting this view and eliminating the category of nonsubsective adjectives would allow a simplifica-
tion of the compositional semantic treatment that we present in this chapter; for the sake of clarity of
analysis and consistency with much of the literature on adjective semantics, we maintain the traditional
three-way distinction between intersective, (non-intersective) subsective, and nonsubsective adjectives
in our discussion.
6 We ignore the role of intensionality in the interpretation of subsective and nonsubsective adjectives;
for discussion of the semantics of adjectives and intensionality, see Kamp and Partee (1995) and Partee
(2010).
480 modification
A modified noun like Swedish man is associated with a meaning constructor whose
right-hand (glue) side is exactly the same as the meaning constructor for man, but
whose left-hand (meaning) side is associated with a modified meaning rather than
an unmodified one:
In this section, we show how a meaning constructor like the one in (20) is derived
from the meaning constructors for Swedish and man.
Augmented lexical entries for Swedish and man, including meaning construc-
tors, are as follows:7
(21) Lexical entries for man and Swedish, including meaning constructors:
a. man N (↑ pred) = ‘man’
λx.man (x) : (↑𝜎 var) ⊸ (↑𝜎 restr)
b. Swedish 3
A (↑ pred) = ‘Swedish’
λP.λx.Swedish (x) ∧ P(x) :
[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)] ⊸
[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)]
The meaning constructor for man is familiar from our discussion of common
nouns in §8.8.2. The meaning constructor for Swedish uses inside-out functional
uncertainty (§6.1.3) to refer to the semantic structure of the phrase it modifies. The
expression (adj ∈ ↑) refers to the f-structure in which ↑ appears as a member of
the modifier set;8 in (20), this f-structure is labeled m. The expression (adj ∈ ↑)𝜎
7 In §13.4, we will refine this analysis in our discussion of recursive modification (modification of
modifiers).
8 The use of the set membership symbol ∈ as an attribute is discussed in §6.3.1.
modifiers and semantic composition 481
The right-hand side of the meaning constructor for Swedish illustrates the charac-
teristic glue contribution of a modifier: it requires a resource of the form v ⊸ r as
its argument and produces a resource of exactly the same form. The general form
for modifiers is given in (23), where M is the meaning of the modifier and S is the
glue contribution of the phrase it modifies:
The meaning contribution of big mouse given in (26) refers to a mouse that exceeds
the size of individuals that are described by the contextually determined property
P . Since the property P is determined by contextual factors, not by semantic
composition on the basis of syntactic relations, we will not specify a means for
determining P but instead will leave it uninstantiated.
Although the meaning contribution of a gradable adjective like big is not the
same as that of an intersective adjective like Swedish, the right-hand sides of the
two meaning constructors are the same, since the two kinds of adjective play a
similar role in meaning assembly. The lexical entry for big in (27) again includes
a somewhat simplified meaning constructor, pending our discussion of recursive
modification in §13.4.
Like the entry for Swedish given in (21), this entry uses inside-out functional
uncertainty to refer to the f-structure of the phrase it modifies. The lexical entry
for mouse is exactly analogous to the one for man, and will not be displayed.
Instantiating the lexical entries for big and mouse according to the labels in (26),
we obtain the following instantiated meaning constructors:
The meaning constructor for big requires a meaning resource of the form v ⊸ r,
and mouse provides such a resource. The resulting meaning is obtained by applying
the expression λR.λx.big (x, P ) ∧ R(x) to its argument λx.mouse (x). The result is
as desired—from the meaning constructors labeled [big] and [mouse] in (28), we
derive the following meaning constructor for big mouse:
resources on the right-hand sides of the meaning constructors for the modifying
adjectives we have examined.
(34) Swedish
v[ ]
r[ ]
λP.λx.Swedish (x) ∧ P(x) : [v ⊸ r] ⊸ [v ⊸ r]
The meaning constructor for Swedish given in (34) provides information about
how to determine the meaning of the phrase it modifies. It does not provide a
representation for the meaning of Swedish independent of its modifying effect;
instead, it represents only the conjunctive meaning that results from combining
Swedish with the phrase it modifies.
Kasper (1997) shows that this view is inadequate by considering examples like
the following:
In this example, the modifier Swedish is itself modified by the adverb obviously. The
effect of modification by obviously is to modify the proposition that x is Swedish,
Swedish (x), to produce a new proposition obviously (Swedish (x)). However, the
proposition Swedish (x) is not by itself associated with the meaning of the adjective
Swedish, and in fact there is no clear way to disentangle the meaning Swedish (x)
from the rest of the meaning contribution for Swedish in (34).
For a meaning like Swedish (x) to be available, we require an independent,
modifiable characterization of the intrinsic meaning of Swedish, which we assume
recursive modification 485
to be the &e → t' type meaning λx.Swedish (x). We also require a theory of how
this meaning combines with the meaning of the modified noun. Kasper (1997)
provides an analysis of examples like (35) within the framework of Head-Driven
Phrase Structure Grammar; though it is stated in different formal terms, our
analysis has a clear basis in Kasper’s proposal.
For this case as well, the appropriate meaning for Swedish is λx.Swedish (x), the
intrinsic, modifiable meaning of the adjective.
There is an independent difference in English between adjectives used predica-
tively, as in (36), and adjectives used as prenominal attributive modifiers: as shown
in (5), English prenominal attributive adjectives are nonprojecting ( 3A ) categories,
while other, nonprenominal uses of adjectives involve the projecting category A as
the head of AP. In fact, some English adjectives are only nonprojecting, and can
be used only prenominally (late with the meaning ‘deceased’: the late president,
but not The president is late, which means The president is tardy and not The
president is deceased), while other adjectives do not have a nonprojecting form and
cannot be used prenominally (afraid: The man is afraid/the man afraid of spiders,
but not ⃰the afraid man). Thus, the different meaning contributions of attributive
and predicative adjectives in English correlate with their status as projecting or
nonprojecting c-structure categories. Predicative adjectives are associated only
with a type &e → t' meaning, while attributive adjectives contribute an additional
meaning constructor which combines the intrinsic adjective meaning with the
meaning of the noun.
We now assume that the semantic structures of adjectives are internally struc-
tured, containing at least the attribute var. In (37), the f-structure f corresponds
to a semantic structure f𝜎 with the attributes var and restr; as shown earlier,
the values of these attributes are labeled v and r. The f-structure s of the adjective
Swedish also has an attribute var, whose value we have labeled sv:
(37) Swedish
v[ ]
f s f
r[ ]
s sv[ ]
9 See Lowe (2013) for discussion of commonalities and differences in meaning constructors for
attributive and predicative uses of adjectives, and §5.4.5 for discussion of copular constructions.
486 modification
The intrinsic meaning of the adjective Swedish is of type &e → t'. Since we assume
that the basic types e and t are associated with semantic structures, we assign the
type e to sv and the type t to s𝜎 . Thus, the meaning constructor in (38), which
represents the core meaning of the word Swedish, is common to both projecting
and nonprojecting forms of the word:10
The meaning constructor in (39) for the nonprojecting, attributive use of Swedish
contains an additional meaning constructor, whose function is to provide instruc-
tions for combining the first meaning constructor with the noun it modifies:
A Swedish (final):
(39) Lexical entry for nonprojecting 3
Swedish 3
A (↑ pred) = ‘Swedish’
λx.Swedish (x) : (↑𝜎 var) ⊸ ↑𝜎
λR.λQ.λx.R(x) ∧ Q(x) :
[(↑𝜎 var) ⊸ ↑𝜎 ] ⊸
[[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)] ⊸
[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)]]
Instantiating the two meaning constructors in (39) according to the labels given in
(37) makes them much easier to read; we label the first meaning constructor in the
lexical entry in (39) as [Swedish-basic] and the second as [attr-adj]:
10 We assume that in an example like (36) the meaning constructor for the copula manages the
relation between the value of the var attribute in the predicative adjective’s semantic structure and
the subj of the copular construction.
recursive modification 487
Therefore, the two meaning constructors [Swedish-basic] and [attr-adj] can play
exactly the same role in meaning assembly as the simple meaning constructor
[Swedish] discussed in §13.3.1. In particular, from the premises [Swedish-basic],
[attr-adj], and [man], we correctly derive the meaning constructor for Swedish
man given in (24):
We treat other adjectival modifiers similarly: the basic meaning constructor for an
adjective combines with a meaning constructor that reflects the function of the
adjective, such as attributive modification, in the syntactic context of its use.
More generally, the example just presented illustrates that the simple and intu-
itive assumptions we make about meanings and how they combine often turn out
to be largely correct, but may be in need of refinement to account for more compli-
cated examples. In logical terms, the intuitively motivated meaning constructors
often correspond to conclusions resulting from a deduction from a more refined set
of basic meaning constructor premises. It is often easier to work with the simpler
and more intuitive constructors; this is legitimate and theoretically sound, as long
as it can be shown that they follow as a logical consequence from the more basic
premises.
o ov[ ]
λx.obviously (Swedish (x)) ∧ man (x) : v ⊸ r
As with the meaning constructors for Swedish given in (40), which we labeled
[Swedish-basic] and [attr-adj], the meaning contribution of obviously is twofold.
The lexical entry for obviously is therefore:
488 modification
Notably, the right-hand side of this meaning constructor is the same as the right-
hand side of the unmodified meaning constructor [Swedish-basic], repeated here,
and plays the same role in meaning composition:
Thus, our refined theory of the semantic contribution of modifiers not only enables
the clean and intuitive treatment of modification presented in §13.3 but also allows
an analysis of recursive modification, which, as Kasper (1997) shows, has proven
problematic in many other approaches.
A big (final):
(51) Lexical entry for nonprojecting 3
big A
3 (↑ pred) = ‘big’
λx.big (x, P ) : (↑𝜎 var) ⊸ ↑𝜎
λR.λQ.λx.R(x) ∧ Q(x) :
[(↑𝜎 var) ⊸ ↑𝜎 ] ⊸
[[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)] ⊸
[((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)]]
The f-structure and semantic structures for obviously big mouse are:
o ov[ ]
λx.obviously (big (x, P )) ∧ mouse (x) : v ⊸ r
Again, instantiating the meaning constructor premises for big in (51) according to
the labels in (52) makes them easier to read:
o ov[ ]
λx.obviously (fake (x, gun )) : v ⊸ r
We assume the same meaning constructor for obviously as in the previous exam-
ples, but with the semantic structure variables instantiated as in (56):
A fake (final):
(58) Lexical entry for nonprojecting 3
fake A
3 (↑ pred) = ‘fake’
λP.λx.fake (x, P) : [((adj ∈ ↑)𝜎 var) ⊸ ((adj ∈ ↑)𝜎 restr)]
⊸ [(↑𝜎 var) ⊸ ↑𝜎 ]
λQ.λx.Q(x) : [(↑𝜎 var) ⊸ ↑𝜎 ] ⊸ [((adj ∈ ↑)𝜎 var) ⊸
((adj ∈ ↑)𝜎 restr)] :
The most important difference distinguishing fake from the other adjectives we
have examined is that fake requires the meaning of the noun it modifies as an
argument: on the glue side, it requires a meaning contribution of the form [v ⊸ r]
to obtain an intrinsic meaning of the form [fv ⊸ f𝜎 ], which is modifiable by
492 modification
obviously. Although these meaning constructors are quite different from those
contributed by intersective and subsective adjectives, no special versions of the
meaning constructors for adjectival modifiers like obviously or nouns like man,
mouse, or gun are required. Given the meaning constructors for obviously, fake,
and gun, we can perform the following deduction:
Again, the glue side of the meaning constructor for obviously fake gun is [v ⊸ r],
exactly the same as the unmodified noun gun, as with all cases of modification.
AdvP
IP
↓ ∈ (↑ ’
Adv NP I
Clearly N VP
David V
fell
A manner adverb like skillfully can be adjoined to VP, as in (62). Evidence that the
adverb skillfully is adjoined to VP in this example comes from the VP preposing
construction, discussed in §3.4 and illustrated in (63), where a VP appears in
fronted position. If the VP includes an adverb, the latter is also preposed, showing
that the adverb forms a constituent with the VP:
NP I
N VP ’
AdvP
David VP
↓∈ (↑
V Adv
played skillfully
The predicate clearly takes as its argument the proposition David fell, and the
meaning represented in (64) is roughly paraphrasable as It is clear that David fell.
Heny (1973), writing at the time at which Nixon was the president of the United
States, considers the following pair of sentences:
As Heny notes, sentence (65a) is true under the rules of the constitution of the
United States, while sentence (65b) is not true by necessity. In other words, though
it turns out to be true that Nixon is a citizen of the United States, this is not
necessarily the case, since Nixon could have decided to become a citizen of another
country. On the other hand, it is necessarily the case that the U.S. president must be
a U.S. citizen under the laws of the United States. The sentences in (66), containing
the sentential adverb clearly, differ from one another in a similar way:
Even in a situation where the running back fell and David is the running back, it
may not be clear that David fell, since the identity of the running back may not be
certain. Adverbs like clearly and necessarily induce opacity in the subject position
of the sentence they modify, since different ways of referring to the same individual
can affect the truth or falsity of the sentence (Quine 1953).
This aspect of the meaning of sentential adverbs is different from manner
adverbs. Intuitively, a manner adverb like skillfully modifies the action that is
performed, producing a new action that is performed skillfully:
adverbial modification 495
In (67), skillfully is a two-place predicate: its arguments are the person that per-
formed the action (here, David) and the action that is performed skillfully (here,
playing). Unlike the situation with sentential adverbs, the following two sentences
are both true if David is the running back and he played skillfully. Manner adverbs
like skillfully do not induce opacity in subject position, so that if David is the
running back, the sentences in (68) are true in the same circumstances:
In general, a manner adverb like skillfully takes two arguments, one corresponding
to the subject of the sentence and the other roughly corresponding to the English
verb phrase—the action that is performed. For this reason, such adverbs are
sometimes called VP or verb phrase adverbs, and as shown in (62), such adverbs
can be adjoined to VP in English. However, meaning combination with adverbs
like skillfully in fact depends on f-structural relations like subj, not c-structure
constituency relations. This is clearly shown by the existence of manner adverbs
in languages without a VP constituent; Broadwell (2005) provides a detailed
examination of the phrase structure of Zapotec, showing that Zapotec lacks a VP
constituent, and that manner adverbs appear adjoined to S:
Adv S
↓ ∈ (↑ adj) ↑ =↓
Dìáp V NP
strongly ↑ =↓ (↑ subj) = ↓
g-ú’ld Màrìì
pot-sing Maria
13.5.2.2 Adverbs and meaning assembly We assume the syntactic and semantic
structures and meaning constructor in (70) for the sentence Clearly David fell.
The f-structure for clearly is labeled c, and its semantic structure c𝜎 contains the
attribute var whose value we have labeled cv:
496 modification
From now on, we will simplify our representations by displaying only semantic
structures whose internal structure is of interest in the constructions we are consid-
ering. Therefore, we do not display the semantic structures d𝜎 or f𝜎 corresponding
to the subject f-structure d and the sentence f-structure f .
We propose this lexical entry for the sentential adverb clearly:
As in the previous sections, the lexical entry in (71) uses inside-out functional
uncertainty to refer to the f-structure of the phrase it modifies. The expression
(adj ∈ ↑) refers to the f-structure modified by clearly, which in (70) is f . The
first meaning constructor in (71) contributes the intrinsic, modifiable meaning
of the adverb, and the second meaning constructor combines with the first one
to produce a meaning capable of modifying a proposition. We assume that both
of these meaning constructors are part of the lexical entry of the adverb, as in
our analysis of attributive adjectives. Unlike adjectives, there is no independent
predicative use of the intrinsic meaning of an adverb, as (72) illustrates, and so
there is no other version of the adverb lexical entry which contains only the
intrinsic meaning of the adverb and not the second meaning constructor:
The instantiated meaning constructors for the sentence Clearly David fell are given
in (73): the meaning constructors contributed by clearly are labeled [clearly1] and
[clearly2], and the meaning constructors [David] and [fall] follow the proposals
for proper names and intransitive verbs given in Chapter 8:
adverbial modification 497
Since the modifying adverb clearly is not itself modified, we first combine the two
meaning constructor premises [clearly1] and [clearly2] to obtain this meaning
constructor:
As described in §8.5.2.1, we can combine the premises labeled [David] and [fall]
to obtain this meaning constructor:
Again, we assume a bipartite semantic contribution for the adverb skillfully. The
lexical entry for skillfully is given in (78), and the instantiated meaning constructor
premises for this sentence are given in (79):
498 modification
The right-hand side of the meaning contribution of the intransitive verb play,
d𝜎 ⊸ p𝜎 exactly matches the requirements of [skillfully]. We combine [skillfully]
and [play], and obtain this meaning constructor:
1 The analyses presented in §§14.3.4 and 14.4 are the result of joint work with Dag Haug.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
incorporated pronouns and “pro-drop” 501
In this lexical entry, the pred value of the obj of zi-ná-wá-lum-a ‘bite’ is
unambiguously specified by the equation (↑ obj pred) = ‘pro’: the object of this
verb is pronominal. In contrast, the pred value of the subj is optionally specified,
as indicated by the parentheses. In the presence of an overt subj phrase, the subject
marker specifies only agreement information (here, information about noun class),
and the pred value of the subj is provided by the overt subject phrase, as shown
in (2):
N V
↑=↓ ↑=↓
zi-ná-wá-lum-a
njûchi
bees (↑ ’
(↑ ((↑
(↑ (↑
(↑
(↑
502 anaphora
In contrast, (3) shows that when there is no subject phrase, the specifications
associated with the verb provide the pred value for the subj:
(3) zi-ná-wá-lum-a
subj-pst-obj-bite-ind
‘They bit them.’
S
’
VP
↑=↓
V
↑=↓
zi-ná-wá-lum-a
(↑ ’
((↑
(↑
(↑
(↑
c pro
obl
In the f-structure in (4), the antecedent David of the reflexive pronoun himself
is the subj of the f-structure labeled c, and the reflexive pronoun is the oblto
of the same f-structure. The semantic antecedency relation, which establishes
504 anaphora
coreference between the anaphor and its antecedent, is defined and discussed
in §14.4.3; here we concentrate on syntactic factors that constrain the anaphor-
antecedent relation.
The antecedent of himself need not be an argument of the same predicate. In
(5), David is an acceptable antecedent of himself, even though himself is the object
of the preposition around, and David is the subject of the verb wrapped:
(7) The antecedent of the English reflexive pronoun himself must appear in the
Minimal Complete Nucleus containing the pronoun.
binding relations 505
According to this definition, the antecedent of the anaphor himself must appear
in the smallest f-structure that contains both the anaphor and a subj. We call the
domain in which the antecedent of the anaphor must appear the binding domain
of the anaphor, and we say that the binding domain of himself is the Minimal
Complete Nucleus.
Languages with multiple anaphors provide evidence for expanding the range
of constraints that anaphors can obey and also demonstrate that constraints
on anaphoric binding must be specified lexically, not universally or on a per-
language basis: different anaphoric elements in the same language may obey
different anaphoric binding constraints. The pronominal system of Norwegian is
particularly rich.2
Hellan (1988) shows that although both the Norwegian reflexive anaphor seg
selv and the reciprocal hverandre must be locally bound, the binding domain for
the reciprocal hverandre is larger than the domain for seg selv. The reflexive seg selv
must be bound to a coargument, an argument governed by the same pred as the
reflexive. In contrast, the reciprocal hverandre, like the English reflexive pronoun
himself, must be bound in the Minimal Complete Nucleus.
In (9), the verb fortalte ‘told’ has three arguments: subj, obj, and an oblabout
argument marked by the preposition om. The antecedent of the oblabout seg selv
is the subj, Jon (Hellan 1988: 67):
2 Hellan (1988), Strand (1992), and Dalrymple (1993) provide further discussion of binding con-
straints on Norwegian anaphors. In a very interesting series of articles, Lødrup (1999a, 2007, 2008a)
challenges some of Hellan’s generalizations and provides alternative analyses of binding constraints for
the anaphoric elements in Norwegian, while maintaining the generalization that binding constraints
can differ for different anaphoric elements.
506 anaphora
In (9), the antecedent of seg selv is a coargument, as required: both seg selv and Jon
are arguments of the verb fortalte ‘tell’. In contrast, (10) is ungrammatical:
In (10), the verb kastet ‘threw’ takes only two arguments, subj and obj; the phrase
fra seg selv is an adjunct. The reflexive pronoun seg selv is the obj of the preposition
fra ‘from’, and the intended antecedent is the subj of the verb kastet ‘throw’. The
reflexive and its intended antecedent are not coarguments of the same pred, and
the sentence is unacceptable.
Example (11) contains the Norwegian reciprocal pronoun hverandre, whose
antecedent is the coargument subj de ‘they’ (Hellan 1988: 67). The f-structure for
(11) is roughly the same as the one displayed in (9) for the reflexive seg selv:
These examples show that different anaphors in the same language may be required
to obey different binding constraints. Thus, these constraints must be lexically
specified for each anaphoric element.
Crosslinguistic examination of anaphoric binding patterns reveals four domains
relevant for anaphoric binding, defined by the f-structure properties pred, subj,
and tense, and by the entire sentence:
These binding requirements are specified as part of the lexical entry of each
anaphoric element (Dalrymple 1993). They limit the possibilities for pronoun
antecedency, which, as we show in §§14.3.2 and 14.3.3, may be further constrained
by other levels of structure.
Formally, we can define the syntactic domain in which an anaphor must find its
antecedent by means of expressions involving inside-out functional uncertainty. As
discussed in §6.1.3, inside-out functional uncertainty allows reference to enclosing
structures, those in which a particular f-structure is contained. In the case at hand,
we can use inside-out functional uncertainty to define the binding domain of an
anaphor, that is, the f-structure domain within which the anaphor is required to
find its antecedent.
Assuming that the f-structure for the pronoun is p, we can define each of the
anaphoric binding domains as follows:
508 anaphora
These expressions constrain the f-structure domain within which the antecedent
of the anaphor can appear. Recall that gf is an abbreviation for any grammatical
function (§6.1.2). In the expressions in (15), the path leading to the anaphor
through the binding domain is as follows:
For clarity, we use the abbreviation gfpro for the grammatical function borne by
the anaphoric pronoun, which can be any grammatical function. Constraints on
the domain within which the anaphor must be bound are stated by means of
off-path constraints on the path to the anaphor.3 In the case of the Coargument
Domain definition, for example, the path leading through the binding domain to
the anaphor is as follows:
This expression is defined in terms of a series of attributes gf, each of which must
obey the off-path constraint ¬(→ pred). In this expression, the symbol → refers to
the f-structure value of the attribute gf, and the expression ¬(→ pred) represents
a negative existential constraint preventing that f-structure from containing the
attribute pred. The effect of this constraint is that the path through the binding
domain leading to the anaphor may not pass through an f-structure containing
the attribute pred. If the path were to pass through such an f-structure, the binding
domain would incorrectly extend beyond the Coargument Domain.
The other binding domains are similarly defined by off-path constraints that pre-
vent the path from passing through an f-structure of a certain type. For anaphors
subject to the Minimal Complete Nucleus constraint, the path from the pronoun
to its antecedent may not pass through an f-structure with a subj attribute, and for
the Minimal Finite Domain constraint, an f-structure with a tense attribute may
not be crossed. For the Root Domain, the path is unconstrained and may be of
arbitrary length. In this way, anaphors that obey positive constraints are required
to find an antecedent within a certain f-structural domain.
Although the antecedent of the anaphor must appear within its binding domain,
antecedents that are too deeply embedded within the binding domain are not
acceptable:
In this expression, as above, p is the f-structure for the anaphor, and (gf⃰ gfpro p)
defines the binding domain containing the anaphor and its antecedent. The
expression in (19) refers to some f-structure within the binding domain which
bears the unspecified grammatical function gfante . The f-command requirement
follows from the form of this expression, since the expression in (19) picks out all
and only the f-structures that f-command the anaphor within the binding domain.
In the expression in (19), the grammatical function of the antecedent gfante
is unconstrained, and any grammatical function may be chosen. In some cases,
however, the grammatical function of the antecedent may also be constrained in
that the antecedent of the anaphor may be required to bear either the function
subj or the function poss within the binding domain. Hellan (1988) shows that
the antecedent of the Norwegian possessive reflexive sin must be a subject:
4 See Culy (1991) for discussion of the Fula pronominal ɗum, which is unusual in not obeying a
command condition.
510 anaphora
F-structures with the subj function in other domains are picked out similarly; see
Dalrymple (1993) and Bresnan et al. (2016) for details.
Other conditions on the antecedent must also be met in some cases. Culy
(1996) discusses the pronominal systems of several varieties of Fula in which
certain pronouns place very specific syntactic requirements on their antecedents:
for instance, that the antecedent must be a pronoun. Culy analyzes this as a type of
agreement between the pronoun and its antecedent.
(24) The antecedent of the English pronoun him must not appear in the Coargu-
ment Domain containing the pronoun.
In the case of positive constraints, the anaphor is required to corefer with some
element picked out by the constraint; in the case of negative constraints, the
anaphor must be noncoreferent with all elements picked out by the constraint.
The antecedent of ham selv must appear in the Minimal Complete Nucleus
containing it. This accounts for the unacceptability of (27), since the intended
antecedent Jon does not appear in the Minimal Complete Nucleus relative to
ham selv:
512 anaphora
However, ham selv may not corefer with a coargument subject, even one that is in
its Minimal Complete Nucleus:
Thus, the Norwegian pronoun ham selv obeys two binding conditions, one nega-
tive and one positive: it must be noncoreferent with a subj coargument, and it must
be coreferent with an argument in the Minimal Complete Nucleus. Both of these
requirements must be satisfied. This is accomplished by including both a positive
constraint and a negative constraint in the lexical entry for ham selv.
Similarly, in (30) the pronoun him is the object of believe and also the subject of like,
and himself is the object of like. This means that him and himself are coarguments
of like, but this does not lead to ungrammaticality:
binding and prominence 513
As these examples show, coreference with certain coarguments does not violate
the negative binding constraints which him obeys. In contrast, him in (31a) cannot
corefer with the object of compare, and him as the object of like cannot corefer with
the subject of like, David (31b):
Only arguments which are superior on the grammatical function hierarchy are
relevant for the binding constraints associated with him. Bresnan et al. (2016:
230) refer to this requirement as the syntactic rank condition, and require binding
constraints to be defined in terms of syntactic rank:
In (35), an overt pronoun is not acceptable if coreference with the matrix subject
is intended:
However, pronominals that are not overtly realized at c-structure do not obey such
ordering restrictions. The subordinate clause in (36) contains a null pronoun that
may corefer with the matrix subject kut..ti ‘child’:
514 anaphora
Kameyama (1985) examines similar data from Japanese, discussed in §6.11.4, and
proposes the following generalization, valid for pronouns in Japanese, Malayalam,
and many other languages:
(38) F-precedence:
f f-precedes g ( f <f g) if and only if for all n1 ∈ 𝜙−1 (f ) and for all n2 ∈
𝜙−1 (g), n1 c-precedes n2 .
This definition, together with the generalization in (37), predicts the patterns
of acceptability for the Japanese and Malayalam examples examined above: null
pronouns f-precede and are f-preceded by every element in the sentence, so
no matter where the antecedent of a null pronoun appears in the sentence, the
condition in (37) is satisfied. In contrast, overt pronouns are not permitted to
f-precede their antecedents, accounting for the unacceptability of coreference in
examples (34) and (35).
Bresnan (1995, 1998) and Bresnan et al. (2016) also discuss linear precedence
conditions on anaphoric binding with particular attention to weak crossover vio-
lations in extraction, providing a different definition of f-precedence. We discuss
these proposals in §17.4.2.
However, when the object contains a possessive reflexive whose antecedent is the
subject of the passive verb, only one reading is possible (Hellan 1988: 162):
We might expect that either argument in (41) could be interpreted as the theme
and either as the malefactive, in the same way that there is flexibility in (40).
In fact, however, the only possible construal of this sentence is one where the
subject barnet is the malefactive argument, and the object sine foreldre is the theme.
Assuming that the goal and malefactive arguments occupy the same position on
the thematic hierarchy, the malefactive argument outranks the theme argument.
In the permissible reading of (41), then, the antecedent barnet outranks the phrase
containing the pronoun sine foreldre on the thematic hierarchy. The other reading,
where the phrase containing the pronoun thematically outranks the antecedent, is
not available.
516 anaphora
In (42) and (43), gfante is the grammatical function whose value is the potential
antecedent of the anaphor, labeled f in (44) and referred to as → in (43). The
f-structure labeled c in (44) is a coargument of gfante , and is given the name
%coarg in the first constraint in (43). According to the second constraint, there is
a path5 gf⃰ through c leading to the f-structure p of the anaphor; in other words, c is
the coargument of f which contains the anaphor. We can now require the potential
antecedent f to be more prominent than its coargument c by specifying a particular
definition for the relation prominent:
. /
(44) gfante f[ ]
gf c[ ... p[ ]]
5 If the anaphor and its potential antecedent are coarguments, the path gf⃰ is empty, and c and p are
the same f-structure.
anaphora and semantic composition 517
We can then define the template gf-prominent as encoding exactly this set of
constraints:
And we can state the binding conditions on him by means of the expression in (47),
which uses this template:
The expression in (47) ranges over elements in the coargument domain which
are higher on the grammatical function hierarchy than the f-structure of the
anaphor: these are the f-structures which are relevant for the application of binding
constraints. Argument structure and f-precedence prominence constraints are
encoded similarly, by altering the definition of prominent in (43).
in assuming that the glue approach not only accounts for resource-sensitive aspects
of meaning assembly within sentences, but also manages contextual contributions
in discourse interpretation. In contrast to their approach, however, Dalrymple
(2001) provides a theory of coreference and anaphoric binding that is closer to
Discourse Representation Theory (DRT: Kamp 1981; Kamp and Reyle 1993) and
Dynamic Predicate Logic (Groenendijk and Stokhof 1991).
An alternative approach to using the glue language itself to manage contextual
contributions across discourse is to combine glue with a meaning language which
is more dynamic than standard predicate logic, and thus better able to handle the
variability in coreference involved in anaphora. Kokkonidis (2005) proposes to
pair glue expressions with meanings couched in a compositional version of DRT
which is capable of dealing with anaphoric resolution. Haug (2014b) discusses
anaphoric resolution in compositional DRT in detail, and develops a Partial version
of Compositional DRT (PCDRT) which provides a clear separation of the semantic
and pragmatic aspects of anaphora. Haug’s proposals are further developed within
the glue approach by Haug (2013), Belyaev and Haug (2014), and Dalrymple et al.
(2018). This is the approach that we adopt and build on in this book.
In a DRS, discourse referents are listed in the “universe,” the upper portion of the
“box”; the discourse conditions, that is the truth-conditional requirements placed
anaphora and semantic composition 519
on the discourse referents, are represented in the lower portion of the box. In (48),
there is only one discourse referent in the universe: x1 , introduced by David.
In our analysis, we assume the Partial Compositional DRT (PCDRT) approach
of Haug (2014b), which differs from earlier versions of DRT in its treatment of
anaphora. The standard DRT assumption is that a sentence containing a pronoun
cannot be interpreted until the pronoun is resolved. This is problematic, as Haug
points out: if a sentence with an unresolved pronoun is uninterpretable, the content
of the sentence cannot play a role in determining how the pronoun is resolved.
On the PCDRT view, anaphoric resolution is assumed to be part of the non-
monotonic, pragmatic content of discourse. This also allows for the fact that as
the hearer acquires additional information, it may become clear that the initial
resolution for the pronoun was incorrect and must be revised.
In PCDRT, as in standard DRT, non-anaphoric phrases introduce discourse
referents into the discourse context, as shown in (48).6 Pronouns also introduce
discourse referents, but unlike standard DRT, these are not directly resolved to
their antecedents in the truth-conditional meaning representation. In the PCDRT-
based DRS for the two sentences in (49), both the proper name and the pronoun
introduce discourse referents, x1 and x2 respectively:
In (49), there are two discourse referents in the universe: x1 , introduced by the
proper noun David, and x2 , introduced by the pronoun he. The referent introduced
by the pronoun is marked with an overbar (x2 ), representing the fact that x2 is a
discourse referent whose reference must be resolved. The descriptive content of
the anaphoric pronoun he, in this case simply the information that the discourse
referent is male, is embedded under the presupposition operator, ∂ (Beaver 1992),
representing the fact that this aspect of the meaning constrains the possibilities for
anaphoric resolution. The presupposition connective ∂ maps ∂(𝜙) to true if 𝜙 is
true and to the undefined truth value otherwise. In this example, the only available
antecedent for x2 is x1 , but in more complex examples there may be other ways to
resolve the reference of x2 .
Anaphoric resolution is represented as a function, A, from anaphoric discourse
referents to their antecedents: in (49), the relation A holds between x2 and x1 . All
anaphoric discourse referents must obey the requirement in (50) for coreference
6 Of course, discourse referents can also be introduced by various other kinds of phrases and can
correspond to events, to plural individuals formed from the individuals relevant in the context, and so
on; see Kamp and Reyle (1993) for discussion of these issues. Haug’s (2014b) approach can also manage
these aspects of context update; see Dalrymple et al. (2018) for discussion of these issues.
520 anaphora
The core of pragmatic anaphora resolution is then the function R, which maps
indices to antecedent indices; crucially, as we will see, indices are accessible to syn-
tactic representations and constraints. This allows us to keep the simple idea under-
lying the coindexation approach, namely that anaphoric relations are just relations
between syntactic tokens, but without presupposing that the resolution actually
takes place in the syntax. We often omit the subscript s on R, as we do with A.
We thus have the following set-up: indices, which are syntactically accessible,
introduce discourse referents; by mapping from the discourse referent of an
anaphor to its index, then from the anaphor index to the index of its antecedent,
and finally from the antecedent index to the antecedent discourse referent, we
obtain a mapping between the discourse referent of an anaphor and its antecedent
discourse referent (in a particular state s; as stated above, we often omit specifica-
tion of s to avoid clutter):
I −1 I
index1 index2
R
anaphora and semantic composition 521
If the discourse starting with the sentence in (53) continues with the sentence He
yawned, we have Haug’s (2014b) PCDRT representation in (54). The discourse
referent x2 for the pronoun he is anaphoric, as indicated by the overbar. We assume
that it takes as its antecedent the discourse referent x1 for the phrase a man, as
represented by the A relation:
As in standard DRT, quantifiers like nobody and everyone are different from
proper names and indefinites in their effect on the context. They do not introduce
new discourse referents into the wider discourse; instead, they effectively “trap”
discourse referents and do not allow them to contribute to the global context.
522 anaphora
x1
¬ person (x1 )
arrive (x1 )
x1
⇒
person (x1 ) arrive (x1 )
In these examples, the universe of the main discourse structure is empty. This
explains the infelicity of the sentences in (57), as there are no discourse referents
available in the main discourse structure to resolve pronouns in subsequent
discourse:
This observation is reflected in the fact that the universe of the main discourse
structure for Nobody saw a mouse does not contain discourse markers that repre-
sent either a mouse or a person:
7 The symbol ¬ in front of the sub-DRS in (55) is the negation symbol. On the treatment of negative
quantifiers such as nobody in DRT, see Corblin (1996).
8 For a brief discussion of quantifier scope, including narrow and wide scope readings, see §8.4.1.4.
anaphora and semantic composition 523
x1 x2
person (x1 )
¬
see (x1 , x2 )
mouse (x2 )
I
(60) 2 3
0 1 x1
d pred ‘David’ 𝜎 d𝜎 index 1
David (x1 )
Similarly, the index introduced by a pronoun appears as the value of the index
attribute in the s-structure of the pronoun, and is related via the function I to the
discourse referent of the pronoun. This is important in allowing the statement of
syntactic constraints on anaphor-antecedent relations, since it allows the semantic
relation between anaphors and their potential antecedents to be syntactically
mediated and constrained.
9 The index attributes at f-structure and semantic structure have a very different status: the value of
the f-structure index feature is a syntactic agreement feature bundle specifying pers, num, and gend
features (§2.5.7.1), while the value of the semantic index feature is a semantic index.
524 anaphora
This means that we can constrain anaphoric resolution possibilities for an anaphor
by referring to the index values of the anaphor and of its potential antecedents. We
assume that the appropriate prominence constraints are also imposed, as described
in §14.3, though for the sake of simplicity of explanation we do not explicitly
specify prominence constraints in this section.
As discussed in §14.2.1, positive binding constraints define the permissible
syntactic relations between a pronoun and its antecedent. For instance, we have
seen that the antecedent of the English reflexive pronoun himself must appear
in the Minimal Complete Nucleus relative to the pronoun, that is, the minimal
f-structure containing the pronoun and a subj function. In the sentence David
compared Chris to himself, there are two potential antecedents for himself in the
Minimal Complete Nucleus: David and Chris. From a syntactic perspective, either
of these antecedents may be chosen; here we assume a context in which David
is chosen as the antecedent for himself. Example (62) illustrates a syntactically
and semantically wellformed binding relation between himself and David. For
clarity, we represent the relation R by an arrow from the pronoun to its potential
antecedent:
d
R
p
p
If we assume that p is the f-structure for himself, as in (62), the expression in (63)
represents the set of permissible f-structure antecedents of himself, including David
(the chosen antecedent) as well as Chris (a syntactically permitted antecedent, but
anaphora and semantic composition 525
not the one chosen in this context). The expression in (63) reflects the definition
of the Minimal Complete Nucleus binding condition given in (15), according to
which the f-structure path gf⃰ delimiting the binding domain is not allowed to
pass through an f-structure with a subj function:
We can now give a complete and explicit formulation of the positive Minimal
Complete Nucleus binding constraint as it applies to the reflexive pronoun himself.
The specification in (64) is a part of the lexical entry of himself, and constrains the
R relation between the index of himself and the index of its antecedent:
The equation in (64) requires the index of the antecedent of the pronoun,
R(↑𝜎 index), to appear as the value of index in the s-structure of a syntactically
permitted antecedent for the pronoun. In (62), this condition is met, and the
binding relation is permitted.
Example (65) illustrates an impermissible binding relation:
d
R
p
c
p
The index for David, 1, cannot be chosen as the antecedent of the reflexive pronoun
himself, and so cannot be the value of R(2), because the f-structure for David
(labeled d) does not stand in a syntactically permissible antecedent relation to the
pronoun f-structure p: the Minimal Complete Nucleus for the pronoun p is the
526 anaphora
(66) ×
Davidi thought that Chrisj had seen himi,⃰j .
Negative binding constraints are more complex to state than positive constraints, in
that negative constraints cannot be stated simply as constraints on the antecedency
relation R. To see this, consider the following sentence:
The antecedent of him may not be he, since him and he are coarguments. However,
as in (66), the antecedent of him may be David:
(68)
Davidi thought that he had seen himi .
Independently, David can serve as antecedent for he since, again, David does not
stand in a coargument relation to he:
(69)
Davidi thought that hei had seen him.
However, as observed by Wasow (1972), Lasnik (1976), and Higginbotham (1983),
if he takes David as its antecedent, then David cannot also serve as antecedent for
him, since in that case he and him would corefer, in violation of the requirement
for a pronoun like him not to corefer with a coargument.
(70) ×
coarguments, but also any index which is related to one of its coarguments via
an antecedent relation.
In order to impose the proper constraints, we define a function R⃰, a recursive
version of R. R⃰ follows the R path back from index to antecedent index, stopping
only when it finds a index which does not have an antecedent:
We can now formalize the negative constraint on him in the following way:
This complex constraint appears in the lexical entry of the pronoun him. Notice
that the negation scopes over the disjunction over grammatical functions in the
coargument domain, giving universal force: R⃰(↑𝜎 index) may not be equal to R⃰
applied to any of the superior coarguments of ↑ . We examine each part of this
complex constraint in turn.
The constraint in (72) is repeated in (73), with the framed portion delimiting
the noncoreference domain of the anaphor. The off-path constraint ¬(→ pred)
defines the noncoreference domain as the Coargument Domain (compare exam-
ple 17): the minimal f-structure containing a pred.
The framed portion of the constraint in (74) picks out the semantic structures of
the coarguments of the anaphor:10
Since indices appear as the value of the index attribute in semantic structure, the
framed portion of the constraint in (75) refers to the indices of the coarguments of
the anaphor; these indices and any indices related to them via the R relation may
not corefer with the anaphor or any of its antecedents:
10 Recall that we leave the prominence constraint implicit (see §14.3): the fully specified constraint
would refer only to more prominent coarguments of the anaphor, not to all coarguments.
528 anaphora
Thus, the constraint in (72) states that the discourse referent which is the
antecedent (of the antecedent (of the antecedent . . . )) of the pronoun may not be
identical with any coargument of the pronoun, or an antecedent (of an antecedent
(of an antecedent . . . )) of any coargument of the pronoun. In the case of (70), this
constraint produces the desired result: David is not an appropriate antecedent for
him because David is the antecedent of he, and he is a coargument of him.
In the case of (76) and (77), this constraint prevents him from coreferring with
Chris (76), but does not prevent him from coreferring with David (77), since David
is neither a coargument of him, nor the antecedent of a coargument of him.
there is no syntactic relation between the f-structure for David arrived and
the f-structure for He yawned, and therefore no syntactic relation between the
f-structure for the anaphor he and the f-structure for its antecedent David. In such
cases, the function R is not constrained by syntax. We assume that speakers have
access to s-structures (and indices) from the preceding discourse, enabling them
to make pragmatic connections between discourse referents belonging to different
sentences and utterances. We can therefore use R for intersentential anaphora in
exactly the same way as for intrasentential anaphora.
The first sentence of the discourse in (78) is David arrived, with the following
f-structure, s-structure, and DRS:
The discourse continues with the sentence He yawned; the f-structure and s-
structure for He yawned and the DRS for (78) are:
(80) He yawned.
’
x 1 x2
David (x1 )
y h arrive (x1 )
h
∂(male (x2 ))
yawn (x2 )
Although there is no syntactic relation between the f-structure for he and the f-
structure for its antecedent David, we assume that the discourse representation
built up so far is available to the conversational participants, and that it is possible
for the function R to relate the index 2 of the pronoun to the index 1 of its
antecedent. Importantly, the negative constraint on pronouns given in (72) allows
this, because the antecedent David is not coreferent with a more prominent
syntactic coargument of the pronoun he. Example (81), repeated from (49), shows
the resolved meaning representation for the discourse under discussion, with
the anaphoric resolution function R relating the index for he to the index for
David:
We now show how the DRS in (81) is produced from its component parts.
The lexical entries for David and arrived are given in (83); we type-raise David,
treating it as a quantifier:
The symbol ⊕ represents the merge operation for combining DRSs by taking the
union of the discourse referents in the universe and the union of the discourse
conditions on the discourse referents. Instantiating the meaning constructors in
(83) according to the labels in the f-structure in (82), we get:
x1
(84) [David] λP. ⊕ P(x1 ) : ∀H.(d𝜎 ⊸ H) ⊸ H
David (x1 )
[arrive] λx. : d𝜎 ⊸ a𝜎
arrive (x)
x1
(85) λP. ⊕ P(x1 ) : ∀H.(d𝜎 ⊸ H) ⊸ H
David (x1 )
This is the meaning constructor [David]. On the
glue side, if we find a semantic resource of the
form d𝜎 ⊸ H for some semantic structure H, we
consume that resource and produce a semantic
resource for H. On the meaning side, we apply the
x1
function λP. ⊕ P(x1 ) to the meaning
David (x1 )
associated with d𝜎 ⊸ H.
λx. : d𝜎 ⊸ a𝜎
arrive (x)
This is the meaning constructor [arrive]. On the
glue side, it contributes an implicational resource
d𝜎 ⊸ a𝜎 , of exactly the form required by [David].
On the meaning side, it contributes a function
from individuals x to a constraint requiring x to
be an individual who arrived.
x1
[λP. ⊕ P(x1 )](λx. ) : a𝜎
David (x1 ) arrive (x)
By applying the meaning of [David] to the mean-
ing of [arrived], we have derived a meaning for
the semantic structure a𝜎 . In the next step, we
simplify this result by replacing all occurrences of
P in the meaning contributed by [David] with the
meaning contributed by [arrive].
x1
⊕ λx. (x1 ) : a𝜎
David (x1 ) arrive (x)
We can now simplify again, by replacing all occur-
rences of x in the meaning of [arrive] with the
discourse referent x1 .
x1
⊕ : a𝜎
David (x1 ) arrive (x1 )
The final simplification consists of merging the
two DRSs connected by ⊕ into a single DRS, as
required by the ⊕ merge operator.
x1
David (x1 ) : a𝜎
arrive (x1 )
We have produced the desired meaning for David
arrived.
(86) a. For a sequence of sentences s1 s2, where s1 immediately precedes s2, the
value of the context attribute of s1 is assigned as the initial value of the
context attribute of s2. If a sentence is the first sentence in a discourse
and there is no preceding context, its context value is initially associated
with the null context.11
b. Each sentence in a discourse contributes a context update meaning con-
structor which updates its initial context to produce a new context
which can be further updated in subsequent discourse.
We begin our discussion of context update with an analysis of (78), in which the
sentence David arrived is the first sentence in the discourse. In this circumstance,
we associate the value of the context attribute of David arrived with the null
context, represented as a DRS with an empty universe and no conditions. This is
shown in (87) for the s-structure a𝜎 for the sentence David arrived:
Since sentences in DRT are treated as context modifiers, we associate root clauses
with a meaning constructor enforcing modification of the current context. This is
shown in (88):
This meaning constructor consumes the meaning resource ↑𝜎 of the root sentence
to produce a modifier of its context. For the sentence David arrived, the up arrow
↑ in (88) is instantiated to a, as shown in (89a). Combining the context update
meaning constructor in (89a) and the meaning of David arrived given in the last
line of (85), repeated in (89b), produces the meaning constructor in (89c):
11 This is a simplification; we set aside representation of aspects of the context such as world
knowledge, location and time of the speech situation, and the identities of the interlocutors.
anaphora and semantic composition 533
x1
c. λQ. David (x1 ) ⊕ Q : (a𝜎 context) ⊸ (a𝜎 context)
arrive (x1 )
Note that the meaning constructor in (89c) has the form of a context modifier
(Chapter 13), consuming the meaning of the context of a sentence and producing
an updated context meaning.
We now combine the meaning constructor in (89c)—which updates the input
context with the meaning of the current sentence David arrived—with the input
context in (87) representing the null context, since David arrived is the first
sentence in the discourse. This produces the updated discourse context in (90),
incorporating the effect of the sentence David arrived on the null context:
This provides the input context for any subsequent sentence in the discourse.
(91) He yawned.
’
y
h
x1
(92) [he] λP. ⊕ P(x1 ) : ∀H.(h𝜎 ⊸ H) ⊸ H
∂(male (x1 ))
[yawn] λx. : h𝜎 ⊸ y𝜎
yawn (x)
The meaning constructor for he differs from the one for David in only two ways.
First, the discourse referent x1 is marked as an anaphoric discourse referent, as
indicated by the overbar. This means that the discourse in which this referent
appears can only be evaluated for truth in the context of an anaphoric resolution
function R. Secondly, the condition requiring x1 to be male is in the scope of the
presupposition operator ∂. The meaning constructor for yawn differs from the one
for arrive only in referring to an act of yawning rather than to an act of arriving.
534 anaphora
Combining these meaning constructors along the same lines as in (85) produces
the meaning in (93) for the sentence He yawned:
(93) He yawned.
x1
∂(male (x1 )) : y𝜎
yawn (x1 )
Since He yawned is a root sentence like David arrived, it also contributes the
meaning constructor in (88), reflecting its role as a context modifier. Combining
the meaning constructor in (93) with the meaning constructor in (88) produces the
meaning constructor in (94), which requires the sentence to modify the context in
which it appears:
x1
(94) λQ. ∂(male (x1 )) ⊕ Q : (y𝜎 context) ⊸ (y𝜎 context)
yawn (x1 )
The discourse context for the sentence He yawned is the context inherited from the
previous sentence David arrived, shown in (90); that is, the initial context value
of the sentence He yawned is the same as the context value of David arrived. Thus,
(y𝜎 context) is initially associated with the same context as in (90):
x1
(95) David (x1 ) : (y𝜎 context)
arrive (x1 )
The meaning constructor in (95) must combine with the meaning constructor in
(94) to update this context. When DRSs are combined, their discourse referents
may have to be renumbered to prevent a clash; see Haug (2014b) for discussion
of this point. Linear precedence determines which discourse referents are renum-
bered: referents introduced earlier in the discourse retain their subscript. In this
example, then, x1 in the meaning for He yawned must be renumbered as x2 when
its DRS is composed with the DRS for David arrived, since a discourse referent x1
is already present. The result is:
x1 x2
David (x1 )
(96) arrive (x1 ) : (y𝜎 context)
∂(male (x2 ))
yawn (x2 )
anaphora and semantic composition 535
x1 x2
David (x1 )
(97) arrive (x1 ) : (y𝜎 context), R(2) = 1
∂(male (x2 ))
yawn (x2 )
Note that anaphoric resolution does not remove discourse referents from the
context, or restrict the extent to which a particular discourse referent may function
as an antecedent to further anaphors. For example, if the discourse above is
continued by the sentence He fell asleep, the fact that x1 serves as the antecedent
to x2 (by virtue of the resolution R(2) = 1) does not prevent either x1 or x2
from serving as the antecedent to the discourse referent x3 introduced by the new
pronoun.
After this sentence is uttered, the context (the universe of the DRS) contains the
discourse referent x1 representing an individual who is a person and who arrived.
To derive this meaning, we assume this lexical entry for someone:
This is similar in form to the lexical entry for the type-raised proper name David
in (83).
In the derivation of the meaning of Someone arrived, the premises in (100) are
relevant:
536 anaphora
[arrive] λx. : s𝜎 ⊸ a𝜎
arrive (x)
x1
[someone] λP. ⊕ P(x1 ) : (s𝜎 ⊸ a𝜎 ) ⊸ a𝜎
person (x1 )
These meaning constructors combine in the way shown in (85) to produce the
complete and coherent meaning in (101), as desired:
x1
(101) [someone], [arrive] 1 person (x1 ) : a𝜎
arrive (x1 )
x1
(102) person (x1 ) : (a𝜎 context)
arrive (x1 )
(104) He yawned.
’ x1
y ∂(male (x1 )) : y
h
yawn (x1 )
The meaning constructor premises in (105) are involved in the derivation of the
meaning of the discourse Someone arrived. He yawned. as follows:
anaphora and semantic composition 537
[null-context] : (a )
x1
[someone-arrive] person (x1 ) : a
λP.λQ.P ⊕ Q : a ⊸ (a ) ⊸ (a
arrive (x1 )
)
x1
[he-yawn] ∂(male (x1 )) : y
yawn (x1 )
λP.λQ.P ⊕ Q : y ⊸ (y ) ⊸ (y )
x1
(106) person (x1 ) : (a𝜎 context)
arrive (x1 )
We likewise combine the meaning constructor [he-yawn] with the meaning con-
structor [context-mod2] to produce the following context modifier:
x1
(107) λQ. ∂(male (x1 )) ⊕ Q : (y𝜎 context) ⊸ (y𝜎 context)
yawn (x1 )
As with our previous examples, the context to be modified, (y𝜎 context), is inher-
ited from the previous sentence, here Someone arrived. In other words, the context
for He yawned is the one given in (106), and the initial value for (y𝜎 context) is
inherited from (a𝜎 context). Combining these meaning constructors produces
the complete, updated context in (108) for He yawned, and the meaning for the
discourse as a whole:
Since the subject of each sentence introduces its own discourse referent, there must
be two distinct discourse referents in the combined meaning for the discourse.
Thus the referent labeled x1 in the meaning [he-yawn] must be renumbered as x2
in the DRS in (108) to avoid identity with the referent introduced by the meaning
[someone-arrive].
As with our previous example, since this meaning contains anaphoric discourse
referents, it is uninterpretable in the absence of a resolution function. Since there
is only one possible antecedent for the pronoun, the result is as follows:
x1 x2
person (x1 )
(109) arrive (x1 ) : (y𝜎 context), R(2) = 1
∂(male (x2 ))
yawn (x2 )
. /
pred ‘arrive"subj#’ x1
a 0 1 : a𝜎
subj n pred ‘nobody’ ¬ person (x1 )
arrive (x1 )
No discourse referent is available outside the scope of the negation after this
sentence is uttered; this explains the infelicity of a continuation like He yawned,
which requires an accessible antecedent for interpretation of its pronoun subject.
We assume the following lexical entry for the negative quantifier nobody:
λP. x1 : ∀H.(↑𝜎 ⊸ H) ⊸ H
¬[ ⊕ P(x1 )]
person (x1 )
anaphora and semantic composition 539
The crucial difference between this meaning constructor and the one for someone
is that the discourse referent introduced and the condition placed upon it are
embedded within a larger DRS, under the scope of the negation operator. This
means that the contextual contribution of nobody is to leave the wider context
(that is, the context outside the scope of the negative) unchanged. Thus, like
the earlier proposal of Dalrymple et al. (1997), this analysis has the desirable
property of correctly characterizing interactions between quantifier scope and
bound anaphora: any pronouns bound by a quantifier like everyone or nobody must
appear within the scope of the quantifier, since the discourse referent introduced
by the quantifier is only available in contexts within its scope, not outside it.
For the sentence Nobody arrived, we assume these meaning constructor
premises:
[arrive] λx. : n𝜎 ⊸ a𝜎
arrive (x)
x1
(114) [nobody], [arrive] 1 : a𝜎
¬ person (x1 )
arrive (x1 )
Next, we briefly examine the derivation of the meaning of the sentence Nobody saw
a mouse. On the most accessible reading of this sentence, the indefinite a mouse
appears inside the scope of the quantifier nobody, and does not contribute to the
540 anaphora
overall context. This accounts for the infelicity of a sentence like It squeaked in
subsequent discourse:
The derivation of this reading proceeds on the basis of the meaning constructors
in (116), which have been instantiated according to the f-structure labels in (115):
x2
[a] λP.λQ. ⊕ P(x2 ) ⊕ Q(x2 ) : ∀H.((m𝜎 index) ⊸ (m𝜎 restr))
⊸ (m𝜎 ⊸ H) ⊸ H
The premises [a] and [mouse] combine to produce the meaning constructor in
(117), labeled [a-mouse]:12
x2
(117) [a-mouse] λQ. ⊕ Q(x2 ) : ∀H.(m𝜎 ⊸ H) ⊸ H
mouse (x2 )
Combining the premises labeled [a-mouse] and [see], we have the meaning
constructor labeled [see-a-mouse]:13
12 For a full discussion of how the meaning of a determiner like a or every combines with the meaning
of a common noun like mouse, see §8.8.2.
13 For the sake of space, we omit the introduction and discharging of a hypothetical subject. For
details, see §8.8.1.5.
anaphora and semantic composition 541
x2
(118) [see-a-mouse] λy. mouse (x2 ) : n𝜎 ⊸ s𝜎
see (y, x2 )
x1 x2
(119) [nobody-see-a-mouse] person (x1 ) : s𝜎
¬
mouse (x2 )
see (x1 , x2 )
This reading of the sentence makes no discourse referents available for anaphora
resolution in subsequent discourse. Outside of a special context, it is difficult to
get a wide scope reading for the indefinite, according to which there is a particular
mouse that no one saw.
The quantifier everyone works in an entirely parallel way. We assume the
following meaning constructor for everyone:
For the sentence Everyone saw a mouse, two readings are readily available: the
indefinite can take narrow scope, meaning that each person saw a (possibly
different) mouse, or wide scope, meaning that there is a particular mouse which
everybody saw.14 Under the second reading, a discourse referent corresponding to
a mouse is available for pronoun resolution in subsequent discourse. We assume
the following f-structure:
14 Meaning constructors and deductions of narrow and wide scope readings for quantifiers are
discussed in §8.8.1.5.
542 anaphora
x2
[a-mouse] λP. ⊕ P(x2 ) : ∀H.(m𝜎 ⊸ H) ⊸ H
mouse (x2 )
As before, we combine the premises labeled [a-mouse] and [see] to produce the
meaning constructor [see-a-mouse]:
x2
(123) [see-a-mouse] λy. mouse (x2 ) : e𝜎 ⊸ s𝜎
see (y, x2 )
x2
(124) [everyone-see-a-mouse] x1 : s𝜎
⇒ mouse (x2 )
person (x1 )
see (x1 , x2 )
As with the narrow scope reading of nobody saw a mouse, this makes no discourse
referents available for anaphora resolution in subsequent discourse.
When the indefinite takes wide scope, a discourse referent corresponding to
a mouse is available. Beginning with the premises in (122), we combine the
premises labeled [everyone] and [see], producing the meaning constructor labeled
[everyone-see]:
x2
mouse (x2 )
(126) [everyone-see-a-mouse2] x1 : s𝜎
⇒
person (x1 ) see (x1 , x2 )
We have shown that the glue approach is valuable not only in accounting for mean-
ing contributions in the derivation of the meaning of a sentence, but also in man-
aging contextual contributions across sentences in a discourse. The representation
of context and the DRT-based glue-theoretic treatment of anaphora resolution is of
crucial importance in our treatment of anaphoric control in Chapter 15. Elsewhere,
however, when context and anaphora resolution are not relevant to the semantic
issues we examine, we employ simpler representations using predicate logic. No
loss of generality results from this simplification.
This chapter explores the syntax and semantics of functional and anaphoric
control, constructions in which either syntactic or lexical constraints require
coreference between an argument of the matrix clause and an argument of a
subordinate or modifying adjunct clause. In English, such cases include the classes
of equi and raising verbs. Crosslinguistically, descriptions of such constructions
involve reference to functional syntactic relations such as subj, obj, and so on;
therefore, the syntactic discussion in this chapter is primarily centered around the
f-structures of functional and anaphoric control constructions.
The open grammatical functions xcomp and xadj and the closed function
comp were first introduced by Bresnan (1982a) in a pioneering study of clausal
relations and complementation. Since then, there has been a wealth of work in LFG
building on these proposals. In the following sections, we will review the major
proposals within LFG for the syntactic and semantic treatment of functional and
anaphoric control constructions.
Raising verbs are distinguished by the fact that the “raised” argument, the subj
David in (1), is not a semantic argument of the raising verb. In other words, raising
verbs impose no semantic constraints on the raised argument. Notationally, this is
indicated by the position of the raised argument outside of the angled brackets in
the semantic form of the raising verb, as discussed in §2.4.6.1:
1 Raising verbs are so called because of their analysis in transformational grammar (Kiparsky and
Kiparsky 1970; Postal 1974), in which the subject phrase of the subordinate clause was assumed to have
raised, or moved up, from the subordinate clause to its final position in the matrix clause.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
546 functional and anaphoric control
Raising verbs in English and other languages exemplify functional control. Func-
tional control verbs require as an argument an open complement xcomp (§2.1.7).
In (2), the subj of the raising verb functionally controls the subj of the subordinate
xcomp. This means that the subj of the verb seemed is required to be the same
f-structure as the subj of the subordinate xcomp, as indicated by the line con-
necting the two subject positions in (2); see §2.4.3. Other raising verbs exhibit
functional control by an obj, as shown in (3), where the obj of the matrix verb
is the same as the subj of the xcomp.
(4) a. It is raining.
b. ⃰There is raining.
open complements and functional control 547
(5)
In English and many other languages, raising constructions like these are not
limited to verbal predicates, but are also found with adjectival predicates. For
example, the adjectival predicate be likely displays exactly the same behavior as
subject-raising verbs like seem. Such predicates have a syntactically selected subj
argument which is not semantically selected, and is filled by the “raised” argument
of the embedded complement:
b. Dative subject:
Barninu batnaði veikin.
child.def.dat recovered.from disease.def.nom
‘The child recovered from the disease.’
c. Genitive subject:
Verkjanna gætir ekki.
pains.def.gen is.noticeable not
‘The pains are not noticeable.’
Andrews shows that in a functional control construction, the case of a “raised” obj
depends on the casemarking requirements on the subj of the lower clause:
The position of the parenthesized adjunct í barnaskap sínum ‘in his foolishness’,
which is a matrix clause modifier, shows that the “raised” constituent does indeed
appear as the obj of the matrix clause and not in the subordinate clause. Since
this argument is also the subj of the subordinate xcomp, the constraints on
casemarking that the subordinate xcomp verb imposes on this argument must be
met for the sentence to be wellformed. Andrews (1990a) and Zaenen and Maling
(1990) provide more discussion of default case, quirky case, and raising verbs in
Icelandic.
Jacobson (1990, 1992b) discusses some tests which may be taken to demonstrate
the syntactic characteristics of the open complement xcomp in English. First, VP
complement drop is never possible for the open complement xcomp:
Second, xcomp is not among the syntactic categories that can appear in sentence-
initial position in a long-distance dependency in English:
(16) V& −→ V NP VP
↑ =↓ (↑ obj) = ↓ (↑ xcomp) = ↓
3 With Bresnan et al. (2016), we classify the infinitival to complement as a VP, and we assume that the
word to is head of the verbal projection. Falk (2001b) assumes a different c-structure for to-infinitives,
arguing that to appears in C, as the head of CP.
open complements and functional control 551
These lexical entries contain a control equation specifying the relation between a
particular argument of the matrix clause and the subj of the subordinate xcomp.
The control equations for the above verbs are:
These lexical entries and phrase structure rules give rise to the c-structures and
f-structures shown in (20) and (21) for examples (2)–(3). The theory of controller
selection and how the control equation is determined will be discussed in §15.7.
NP I
(↑ ↓ ↑=↓
N VP
’
↑=↓ ↑=↓
David V
(↑ ↑=↓
V VP
↑=↓ (↑ ↓
seemed
V
(↑
↑=↓
(↑ ↑
V VP
↑=↓ ↑=↓
V
to
↑=↓
yawn
(↑ ’
552 functional and anaphoric control
V NP VP
believed N V
Chris V VP
to V
V NP
know D N
the N
answer
Raising predicates with to-infinitive complements are not the only structures in
English which can be analyzed using xcomp. Bresnan et al. (2016: §§12.1 and 12.2)
discuss predicate complements such as the following:
In (22a), the underlined phrase is an AdjP, in (22b) the underlined phrase is a PP,
and in (22c) the underlined phrase is a participial VP. All of the underlined phrases
in (22) can be analyzed as xcomp complements of their respective matrix verbs,
as shown by Bresnan et al. (2016). Bresnan et al. also discuss interactions between
raising and anaphoric binding in constructions such as these.
Falk (2005) notes that the vast majority of predicates which take xcomp com-
plements require those complements to be headed by a verbal predicate, and
do not permit nominal, adjectival, or prepositional complements. Falk argues
that this is unexpected if all open complements have the same function xcomp,
since subcategorization is primarily functional, not categorial (§2.3). Falk therefore
argues for the differentiation of two additional open argument functions, xobj𝜃
and xobl𝜃 , alongside xcomp. Under this proposal, adjectival or nominal open
open complements and functional control 553
complements are xobj𝜃 , prepositional open complements are xobl𝜃 , and xcomp is
restricted to verbal open complements. As discussed in §2.1.7, the status of clausal
complement functions, in particular comp, is controversial; here we follow the
traditional analysis of xcomp as the grammatical function for all types of open
complement.
In both of these sentences, the four words preceding the finite verbs (versuchte
and schien respectively) constitute a single phrase, meaning that the phrase Ein
Aussenseiter, which is the subject of both verbs, appears within the embedded
clause, rather than at the level of the matrix verb. This phenomenon is called
backward control: rather than an argument of the matrix clause controlling an
argument of the embedded clause, the control relation works “backwards,” an
argument from the embedded clause controlling the missing argument position in
the matrix clause (Polinsky and Potsdam 2002a,b, 2006). The f-structure assumed
for (23b) is:
(24) ’
’
554 functional and anaphoric control
The difference between equality and subsumption in this context can be thought
of in terms of information flow. In the case of equality, information flow is bidi-
rectional: any information specified for the embedded argument is automatically
shared with the matrix argument, and vice versa, since they are equal. This allows
the backward control relation observed for scheinen ‘seem’ in (23b), since a phrase
filling the subordinate xcomp subj role is also required to fill the role of the
main clause subj. In the case of the subsumption relation required by the verb
versuchen ‘try’, however, information flow is monodirectional: if a ) b, then any
information specified for a must be shared with b, but information specified for b is
not necessarily shared with a. With versuchen, then, if the shared argument appears
in the matrix clause as the value of (↑ subj), as desired, its pred also appears as the
value of (↑ xcomp subj), in much the same way as it would if the relation were one
of equality. But if the shared argument appears in the embedded clause, as the value
of (↑ xcomp subj), its pred does not appear as the value of (↑ subj), rendering the
matrix f-structure incomplete. This models the fact that versuchen does not permit
backward control: if the controller appears in the embedded clause, the matrix subj
position is empty, leading to an incomplete f-structure.
Sells (2006b) extends Zaenen and Kaplan’s proposals in an analysis distinguish-
ing forward and backward control, proposing that all functional control relations
can be analyzed in terms of subsumption, and arguing that this provides a good
model of the typology of control constructions. Sells argues that in German both
raising and equi verbs specify the control relation in terms of subsumption; so, the
functional description specifying the control relation for the raising verb scheinen
‘seem’ is exactly the same as that for versuchen ‘try’:
open complements and functional control 555
Why, then, is (23b) grammatical? Sells follows Berman (2003) in assuming that in
German, subjects do not need to contain a pred feature, but may consist solely of
agreement features if and only if they are non-thematic. The controller argument
of raising predicates is, as we have seen above, not semantically selected by the
raising predicate. Thus in the case of (23b), the lack of a semantically contentful
(pred-less) subj at the matrix level does not lead to an incomplete f-structure.
On the other hand, as we discuss in §15.4, the controller argument of an equi
predicate is semantically selected, and so the f-structure is ill-formed if the matrix
subj lacks a pred feature. This means that if control relations are stated in terms
of subsumption, backward control is possible with raising predicates, but not with
equi predicates.
In some languages, there are predicates which allow only backward control.
Following work by Polinsky and Potsdam (2002b), Sells analyzes the Malagasy
examples in (27) as “forward control,” with the controller overtly realized in the
main clause, and the examples in (28) as “backward control,” with the controller
overtly realized in the subordinate clause:
Since all of the constraints on the embedded subject also hold for the matrix
subject, and since there is nothing else which can provide a subj value for the
xcomp, this equation ensures that the controller must appear within the embedded
clause.
Whether a comprehensive account of all functional control relations can be
stated in terms of subsumption rather than equality is not clear. Haug (2011)
discusses backward control phenomena in Ancient Greek, and concludes that
equality, rather than subsumption, provides the most satisfying account (see also
Haug 2017). Backward control in Indonesian is also discussed by Arka (2014).
In this example, the predicate seem holds of the proposition yawn (David ), and the
sentence has more or less the same meaning as the sentence It seemed that David
yawned.
As noted in §8.4.1.4, scope ambiguities are available with the subj argument of
a subject raising verb like seem or appear and with the obj argument of an object
raising verb like believe (May 1977; Williams 1983; Jacobson 1990; Halvorsen
1983): a raising verb allows a narrow scope reading for its “raised” argument. For
example, a sentence like Someone appeared to yawn has two readings, a narrow
scope reading paraphrasable as It appeared that someone yawned and a wide scope
reading on which there is some person who appeared to yawn, as shown in (32):
4 Since the analysis of raising verbs does not depend on our theory of anaphoric binding, we
provide simple meaning constructors rather than the PCDRT-based meaning constructors presented
in Chapter 14.
raising verbs and semantic composition 557
In the next section, we will see how both readings for this sentence are derived.
The presence of the subj argument outside the angled brackets in the matrix pred’s
semantic form in (33) indicates that the subj of seem is not a semantic argument
of the verb and that the sole semantic argument is the xcomp. In keeping with this
intuition, we follow Asudeh (2000, 2002a) in proposing the following lexical entry
for the verb seemed:
The meaning constructor in the last line of this lexical entry requires an xcomp
argument. A meaning contribution corresponding to the subj is not required, since
the subj meaning is not a semantic argument of seem . If the subj contributes a
meaning resource, it must be consumed by the xcomp predicate for the sentence
to be semantically complete and coherent.
We now instantiate the meaning constructor given in (34) according to the
f-structure labels in (33). We also provide instantiated meaning constructors for
the proper name David and the intransitive verb yawn:
The meaning constructor labeled [yawn] requires a meaning for its subject, d𝜎 ,
to produce a meaning for y𝜎 . Thus, we first combine the meaning constructors
[David] and [yawn] to produce the meaning constructor [David-yawn] for y𝜎 .
A meaning for y𝜎 is exactly what the meaning constructor [seem] requires, and a
semantically complete and coherent meaning constructor for the sentence results.
The meaning constructor premises in (39) are relevant for this example. Again,
since we are examining the semantics of raising verbs and are not concentrating
on issues related to context update and anaphoric binding, we provide the simple
meaning constructor for the indefinite quantifier someone which was discussed in
§8.8, rather than the PCDRT constructor discussed in §14.4.3.7:
For the narrow scope reading, we note that the quantifier meaning constructor
[someone] requires as its argument a resource of the form s𝜎 ⊸ H for some
semantic structure H. The meaning constructor [yawn] provides such a resource.
Combining [someone] and [yawn], we have the meaning constructor labeled
[someone-yawn]:
To derive the wide scope reading for the example, we make use of the abstraction
rule given as (74) in Chapter 8. Recall that this rule permits the introduction of
a hypothetical premise on the glue side, which is discharged at a later point in
the deduction; on the meaning side, hypothetical premise discharge corresponds
to abstracting over the variable introduced by the premise. For this example, we
hypothesize the premise x : [s𝜎 ] in the first line:
As shown in (42), we combine the hypothesized premise x : [s𝜎 ] with the premise
[yawn], producing the meaning constructor yawn (x) : y𝜎 . This meaning construc-
tor provides the semantic resource y𝜎 required by [appear]. Combining these two
meaning constructors, we produce the meaning constructor appear (yawn (x)) : a𝜎 .
Finally, the hypothesized premise x : [s𝜎 ] is discharged in the last line of (42),
producing the meaning constructor labeled [appear-yawn]: the variable x is
abstracted over on the left-hand side, producing the function λx.appear (yawn (x)),
and the implicational meaning constructor s𝜎 ⊸ a𝜎 is produced on the right-hand
side.
The meaning constructor [appear-yawn] provides a resource of the form
s𝜎 ⊸ H, which is what the quantifier [someone] requires; for this reading, the
semantic structure a𝜎 is chosen to provide the scope meaning of the quantifier.
Combining the meaning constructors [someone] and [appear-yawn], we obtain
the semantically complete and coherent meaning constructor in (43), which
provides the wide scope reading for this example:
Such sentences are parallel in many ways to sentences involving raising from
infinitival complements: (44a) is similar to They seem to have missed the bus, and
(44b) is parallel to John appears to be tired. The crucial similarity between the
copy raising sentences and the standard raising sentences is that the subject of the
matrix verb is not semantically selected, as is evident from sentences like There
seems like there’s a problem; the crucial difference is that there is a pronoun in the
embedded clause corresponding to the controlled argument of the standard raising
sentences (the “copy”). Thus in contrast to standard raising, in copy raising there
are two different noun phrases corresponding to a single semantically selected
argument position. This leads to a semantic difficulty involving resource surplus:
the pronoun subject of the embedded predication makes a contribution to the set
of meaning resources available for semantic composition, but that contribution is
neither required nor usable, because its role is already satisfied by the subject of
the matrix predicate.
In terms of the syntactic analysis of copy raising, Asudeh and Toivonen propose
that the apparent complementizers in English copy raising, e.g. like in (44a) and as
if in (44b), are not in fact complementizers but rather prepositions which select
for two arguments, a subj and a closed complement comp. The subject of the
preposition is controlled by the subject of the matrix raising verb, just as in the
standard raising examples, meaning that we can use exactly the same lexical entry
and functional descriptions for the seem that occurs in copy raising as for the seem
that occurs in standard raising:
NP I
(↑ ↓ ↑=↓ ’
N VP
↑=↓ ↑=↓ ’
V
(↑ ↑=↓
V PP
↑=↓ (↑ ↓
seem
P
(↑
↑=↓
(↑ ↑
P IP
↑=↓ (↑ ↓
like
(↑ ’ they’ve missed the bus
closed complements and anaphoric control 561
5 Equi verbs are so called because of their participation in the “Equi-NP Deletion” transformation
proposed for them by Postal (1970) and others in a transformational framework, in which an NP in
subject position of the subordinate clause of an equi verb was assumed to be deleted under conditions
of identity with an NP in the matrix clause.
562 functional and anaphoric control
In (46), the subj of the obligatory anaphoric control verb tried anaphorically
controls the subj of the comp, and the sentence is interpreted as meaning that
David tried to bring about a situation where David leaves. The controller in an
anaphoric control construction can also be a matrix clause object if an appropriate
matrix verb appears:
Here, the controller of the subj of leave is the obj of the matrix verb, Chris, and the
sentence means that David convinced Chris that Chris should leave.
It is not necessarily the case that if the matrix control verb has an obj, that
argument must be the controller: with verbs such as promise it is possible for the
matrix subj to control the subj of its comp, even though there is also a matrix obj:
Here, the controller of the subj of leave is the subj of the matrix verb, David, and
the sentence means that David promised Chris that David would leave.6 We discuss
syntactic and semantic constraints on the controller in §15.7.
6 This use of the verb promise appears to be dialect-specific, and is not available for all speakers.
564 functional and anaphoric control
This contrasts with the situation with functional control, illustrated in (13), where
the case specified by the subordinate clause verb must appear on the matrix subject.
Since English lacks such case marking, the evidence for a functional distinction
between raising and equi verbs is less clear.7 As discussed in §15.1.1, Jacobson
(1990, 1992b) investigates certain syntactic differences between English raising
and equi verbs. One such difference is VP complement drop: the VP complement of
many equi predicates may be dropped, but this is impossible with raising predicates
(see also (14)):
(50) Equi:
a. [Did David really leave?] He tried.
b. [Will Chris leave?] If David can convince him.
(51) Raising:
'
to (leave).
a. [Did David really leave?] He seemed
⃰∅
'
to (vote).
b. [Does Chris vote?] David believes him
⃰∅
This is a lexically governed option, which is not possible for all equi verbs:
'
to (leave).
(52) a. [Did David really leave?] He wanted
⃰∅
7 Zaenen and Kaplan (2002) even suggest that case mismatches between controller and controllee
do not necessitate an anaphoric control analysis, since sharing of the case feature can be managed by
use of the restriction operator (§6.9.4). Such an approach is technically possible, but does not result in
a linguistically appealing analysis.
closed complements and anaphoric control 565
In fact, these examples are not acceptable for all speakers. Nevertheless, there does
seem to be a contrast between the examples with sentence-initial infinitives in (53)
and those in (15). The relatively low acceptability of the examples in (53) may in
part be due to the unsuitability of the infinitive in the pragmatic/information struc-
ture role associated with the sentence-initial phrase (Tracy Holloway King, p.c.).
In principle, then, equi verbs can be analyzed in terms of either functional or
anaphoric control, and it is sometimes assumed that equi constructions as well
as raising constructions in English involve functional control (see, for example,
Bresnan 1982a, 2001c; Asudeh 2000, 2002a; Falk 2001b; Bresnan et al. 2016).
However, the evidence presented above is consistent with the proposal that at
least some English equi verbs in fact participate in obligatory anaphoric control.
In the absence of unequivocal evidence in favor of the functional control analysis
for English equi verbs, we adopt the obligatory anaphoric control analysis in this
work. Falk (2001b) provides an illuminating discussion of the difference between
anaphoric and functional control in equi constructions, though the conclusions he
draws are different from those adopted here.
In other languages, more variation is found: some languages have two clearly
distinct types of equi verb, some specifying anaphoric control and some specifying
functional control. Kroeger (1993: Chapter 4) shows that Tagalog has two different
types of equi constructions. The first type involves functional control of a subject
argument in the complement clause:
The second type involves anaphoric control of a term (possibly nonsubject) argu-
ment in the complement clause. In (55), the matrix subject Maria anaphorically
controls the objagent argument in the subordinate clause:
Thus, equi verbs that can specify either functional or anaphoric control can be
found, even within a single language.
(56) V& −→ V NP VP
↑ =↓ (↑ obj) = ↓ (↑ xcomp|comp ) = ↓
The rule in (56) differs from the one in (16) in allowing the VP daughter of V&
to bear either the xcomp or the comp function. There is no constituent structure
distinction between VP complements that are functionally controlled, bearing the
xcomp function, and those that are anaphorically controlled and bear the comp
function; xcomp and anaphorically controlled comp appear in the same position
relative to adverbs and direct objects, for example:
Thus, it is only the functional annotations on the rule that distinguish the two cases,
and not the c-structure configuration.
We propose the following syntactic lexical entries for the English equi verbs tried
and convinced:
closed complements and anaphoric control 567
The last line in each of these lexical entries specifies a pronominal subj for the
subordinate comp. In §15.4, we discuss how the anaphoric control relation is
established between the matrix controller and the subordinate clause subj.
These lexical entries, together with the annotated phrase structure rule in (56),
give rise to the anaphoric control structures in (61) and (62):
NP I
’
(↑ ↓ ↑=↓
N VP
’
↑=↓ ↑=↓
David V
(↑ ↑=↓
V VP
↑=↓ (↑ ↓
tried
V
(↑ ’
↑=↓
(↑
V VP
↑=↓ ↑=↓
V
to
↑=↓
leave
(↑ ’
568 functional and anaphoric control
V NP VP
convinced N V
Chris V VP
to V
leave
(63) ’
’
Since the controller is semantically selected, equi verbs cannot appear with a
subordinate comp verb that selects an expletive or semantically empty subject,
since expletive subjects contribute no semantic content and therefore cannot
satisfy the semantic requirements of the equi predicate or enter into an anaphoric
dependency:
In (65), try denotes a relation between an individual David and the proposition
leave (David ) that David is trying to bring about. The equi verb tried requires
the pronominal subject of the subordinate verb leave to be coreferent with the
controller subject of tried, David .
Asudeh (2000, 2002a) proposes an analysis of English equi verbs within the
glue framework that differs from the analysis presented here in its assumptions
about the syntax and meaning of the complement of the equi verb. Asudeh analyzes
equi verbs as taking xcomp complements, not comp; further, following Chierchia
(1984a,b, 1985), Asudeh claims that the xcomp of an equi verb denotes a property
rather than a proposition. On this view, the meaning of a sentence like David tried
to leave is represented as:
In (66), try denotes a relation between an individual and the property that the
individual aspires to have. Chierchia (1984a) argues for the property theory of
control on the basis of entailments like this one:
Chierchia argues that if the complement of the verb try is treated as a proposition
roughly corresponding to the meaning of a sentence like Ezio jogs at sunrise, an
unwanted entailment seems to follow, namely that Nando tries for Ezio to jog at
sunrise:
570 functional and anaphoric control
However, there are several difficulties with the property theory of control advo-
cated by Chierchia and Asudeh.8
First, Chierchia’s proposal is based on the lack of availability of a strict reading
for the conclusion of the argument in (67): the conclusion means that Nando tries
for Nando (the sloppy reading) and not Ezio (the strict reading) to jog at sunrise.9
However, other cases of sloppy-only readings for arguments similar to (67) are
not amenable to a solution that, like Chierchia’s, relies on the property theory of
control. For example, the argument in (70) is also valid:
Coreference between the subject of the verb practice and the understood subject of
the gerund playing is enforced by real-world constraints on situations of practicing
the piano. Regardless of whether the gerund complement of practice is taken to be
a property or a proposition, it makes no sense to talk about Nando practicing Ezio’s
playing the piano. Therefore, Higginbotham argues, examples like this do not shed
light on the issue of the type of the complement of equi verbs.
8 Other authors to have defended the property theory of control include Chierchia and Jacobson
(1985), Dowty (1985), and Jacobson (1990).
9 The distinction between strict and sloppy readings is best known from analyses of ellipsis (see
Dalrymple et al. 1991 and references cited there); in the following, (b) paraphrases the sloppy reading
of (a), and (c) paraphrases the strict reading:
(a) David rode his bike, and Chris did too.
(b) Chris rode Chris’s bike. (sloppy)
(c) Chris rode David’s bike. (strict)
equi verbs and semantic composition 571
As Higginbotham points out, this sentence has two readings, paraphrasable in the
following way:
On the proposition theory of control adopted here, both of these readings are
readily available (we use the notation of Dalrymple et al. 1998 to represent
reciprocal meaning):
The reading paraphrased in (73a) is also readily available on the property theory:
(76) They expect that they will sit next to each other:
expect (they , λx.RECIP (x, λz.λy.sit.next.to (z, y)))
The problem with this representation is that a predicate like expect denotes a
relation between an individual and the property that the individual expects to have.
However, an individual cannot enter into a relationship involving a RECIP predicate,
which must hold of a group and not an individual. These conflicting requirements
make a coherent account of this reading difficult or impossible to obtain under the
assumptions of the property theory.
In addition to these arguments, Higginbotham (1992) presents a number of
other arguments supporting the proposition theory of control; we conclude with
Higginbotham that the proposition theory has clear advantages over the property
theory of control, and we adopt the proposition theory here.10
10 For further arguments in favor of the proposition theory of control, see Sag and Pollard (1991)
and Pollard and Sag (1994).
572 functional and anaphoric control
x1
David (x1 )
x2 , A(x2 ) = x1
try (x1 , )
leave (x2 )
As in Chapter 14, we assume that both the matrix subject David and the unex-
pressed pronominal subject of the complement verb leave contribute a discourse
referent. The unexpressed pronoun contributes an anaphoric discourse referent,
whose antecedent is constrained to be the subject of try. The contribution of a
discourse referent for the subject of the complement clause allows the resolution
of bound anaphors in examples like (78), where the antecedent of the reflexive
pronoun himself is the unexpressed pronominal subject of vote:
x1
λP. ⊕ P(x1 ) : ∀H.((↑ comp subj)𝜎 ⊸ H) ⊸ H
equi verbs and semantic composition 573
The first and second lines of this lexical entry are familiar from our syntactic
discussion in §15.3.3. The third line establishes the antecedency of the pronominal
subject of the comp argument: semantically, the antecedent of the comp subj is the
matrix clause controller, the subj of tried. As discussed in §14.4.1, this is modeled
in terms of the antecedency relation R between the values of index in the semantic
structures projected from the respective f-structures. In (80), we instantiate this
constraint according to the f-structure labels in (77):
The meaning constructor in the fourth line of the lexical entry, which is labeled
[try] in (81), provides the main predicate try . This PCDRT-based meaning con-
structor is parallel to that for arrive, discussed in §14.4.3.4, except that in addition
to the requirement for the meaning of the subj, it also requires a meaning for the
comp argument, which corresponds to the proposition P on the meaning side.
The final line of the lexical entry is the pronominal meaning constructor
supplied by the equi verb for the interpretation of the pronominal subj of its
complement clause, labeled [pro] in (81). This meaning constructor is similar to
the meaning constructor for he discussed in §14.4.3.6, except that, since it is asso-
ciated with a null pronoun, it imposes no person, number, or gender requirements
on the antecedent: in particular, it lacks the requirement for the antecedent to
be male.
The instantiated meaning constructors in (81) are relevant in the analysis of (77):
[leave] λx. : p𝜎 ⊸ l𝜎
leave (x)
x1
[pro] λP. ⊕ P(x1 ) : ∀H. (p𝜎 ⊸ H) ⊸ H
x1
(82) [pro-leave] : l𝜎
leave (x1 )
574 functional and anaphoric control
We also combine [David] with the meaning constructor [try] to produce the
meaning constructor [David-try]:11
x1
(83) [David-try] λP. David (x1 ) : l𝜎 ⊸ t𝜎
try (x1 , P)
x1
David (x1 )
(84) [David-try-pro-leave] x2 : t𝜎
try (x1 , )
leave (x2 )
We have now produced a complete meaning for (77), but the anaphoric discourse
referent x2 still requires an antecedent. According to the constraint in the third
line of (79), the function R applied to the s-structure index feature corresponding
to the pronoun finds the index feature corresponding to David . Thus, we can
augment the meaning constructor in (84) with the correct antecedency relation:
x1
David (x1 )
(85) [David-try-pro-leave] x2 : t𝜎 , R(2) = 1
try (x1 , )
leave (x2 )
11 As defined in §8.8.1.1, the variable H ranges over semantic structures. In order to apply
the meaning [David] to the meaning [try], it is therefore necessary first to hypothesize (§8.8.1.4)
meanings for the subj and comp of try, and then to discharge the hypothesized subject meaning,
resulting in a meaning: [try-P] λx. try (x, P) : d𝜎 ⊸ t𝜎 . The meaning [David] is then applied to
the meaning [try-P], following which the hypothesized meaning P is discharged, resulting in the
meaning [David-try]. Here, and elsewhere in this chapter, we pass over the details of such abstractions,
describing meanings like [David] as directly applying to meanings like [try].
12 Notice that the discourse referents in (82) have been renumbered to avoid a clash with those in
(83); see §14.4.3.6 for discussion.
equi verbs and semantic composition 575
Since this example involves functional rather than anaphoric control, anaphora
resolution is not a central issue in the derivation of its meaning, and we use simple
meaning constructors rather than PCDRT constructors.
We assume that the xcomp complement of the equi verb nagpilit ‘insist on’
patterns with other equi verbs in denoting a proposition, though we have not been
able to verify the semantic patterns presented in §15.4.1 with a native speaker of
Tagalog. On this assumption, we propose the following lexical entry for nagpilit
‘insist on’:
The first and second lines of this lexical entry enforce the syntactic constraints
appropriate for a subject raising verb. The last line gives the meaning con-
structor for this verb, which requires two arguments: an implicational resource
[(↑ subj)𝜎 ⊸ (↑ xcomp)𝜎 ] corresponding to its xcomp argument and a resource
(↑ subj)𝜎 corresponding to its subject. The xcomp resource takes the form of an
implication because it is an argument that is “missing” its subject (which is also
the subject of the main verb). The subject’s meaning is represented by x on the
left-hand side of the meaning constructor in (87); the xcomp’s meaning is the
property P, which is applied to the subject meaning x to produce the proposition
P(x) filling the second argument position of insist .
The meaning constructor in the lexical entry in (87) is labeled [insist] in (88),
instantiated with the f-structure labels given in (86); we also provide the standard
meaning constructors for the proper names Maria and Ben and for the ditransitive
verb bigy-an ‘give’. Since the internal structure of the noun phrase ng-pera ‘money’
is not at issue here, we make the simplifying assumption that this noun phrase
makes a contribution like that of a proper name and has the meaning money :
576 functional and anaphoric control
We first combine the premises labeled [Ben], [money], and [give] to produce the
meaning constructor labeled [give-Ben-money]:
Finally, we combine this meaning constructor with the subject meaning construc-
tor, labeled [Maria]. The resulting meaning constructor, displayed in (91), provides
a semantically complete and coherent meaning for this example:
(93) Chris told Matty that David had gestured to follow Ken.
[possible interpretation: Chris and Matty follow Ken]
A syntactically remote antecedent is also possible. In (94), the subject of follow can
be interpreted as coreferent with Chris, although the noun phrase Chris is not an
argument of the immediately higher clause:
(94) Chris thought that it was clear that David had gestured to follow Ken.
[possible interpretation: Chris follows Ken]
Example (96) shows the c-structure and f-structure for the example David gestured
to follow Chris:
578 functional and anaphoric control
NP I
(↑ ↓ ↑ =↓
N VP
↑ =↓ ↑ =↓
David V
(↑ ↑ =↓
V VP
↑ =↓ ↑ =↓
gestured
V
(↑ ’
↑ =↓
(↑
V VP
↑ =↓ (↑ ↓
V
to
↑ =↓
V NP
↑ =↓ ↑ =↓
follow N
(↑ ’ ↑ =↓
Chris
(↑
arbitrary anaphoric control 579
The most likely interpretation of the sentence David gestured to leave in (97) is that
David gestured for Chris to leave:
In other contexts, other antecedents are possible. Unlike the situation with obliga-
tory anaphoric control, lexical or syntactic constraints do not determine the refer-
ent of the complement subject. Instead, the pronominal subject of the complement
to leave obeys the same constraints on pronoun resolution as an overt pronoun.
However, lexical constraints can play an important role in constraining the range
of possible referents for the controller in an arbitrary control construction. In
(98), the matrix subject in the second sentence cannot corefer with the subject
of the comp: that is, a sentence like David gestured to leave cannot mean that
David gestured for himself to leave. This is a negative constraint on anaphoric
reference, which in this case is imposed by the verb gesture; negative constraints
were discussed in §§14.2.2 and 14.4.3.
x1
Chris (x1 ) : y𝜎
yawn (x1 )
580 functional and anaphoric control
[yawn] λx. : c𝜎 ⊸ y𝜎
yawn (x)
x1
[Chris] λP. ⊕ P(x1 ) : ∀H.(c𝜎 ⊸ H) ⊸ H
Chris (x1 )
The first two premises in (100) are identical to those presented in (87) and (88) of
Chapter 14 in the analysis of the sentence David arrived at the start of a discourse.
The derivation of the meaning of this sentence proceeds analogously. Combining
the premises in (100) produces the following semantically complete and coherent
meaning constructor, as desired:
x1
(101) [context], [context-mod], [yawn], [Chris] * Chris (x1 ) : (y𝜎 context)
yawn (x1 )
The second sentence in (97), David gestured to leave, has the following f-structure
and meaning constructor:
x1
David (x1 )
x2 :g
gesture (x1 , )
leave (x2 )
The lexical entry for the arbitrary control verb gestured is as follows:
arbitrary anaphoric control 581
The first line of this lexical entry provides the pred value for the verb gestured.
The second line of the lexical entry specifies a ‘pro’ value for the pred attribute
of the complement clause’s subject. The third line introduces a constraint on its
antecedent:
As noted earlier, the subj of gesture cannot corefer with its comp subj: the sentence
David gestured to leave cannot mean that David gestured for himself to leave.
The constraint in (104) enforces this requirement by preventing the complement’s
subject’s semantic structure index ((↑ comp subj)𝜎 index) from being related to
the index of (↑ subj)𝜎 via the R⃰ relation.13
The fourth line in (103) is the meaning constructor that introduces the predicate
gesture , labeled [gesture] in (105); it is analogous to the meaning for try given
in (79).
The final line of the lexical entry in (103) provides a pronominal meaning
constructor for the comp subj of gesture, labeled [pro] in (105). Again, this
meaning constructor is identical to that found in the lexical entry of the equi verb
try in (79), and is similar to the meaning constructor for he discussed in §14.4.3.6,
except that it lacks the requirement that the antecedent be male.
With the context [context] produced by uttering the first sentence in the
discourse, Chris yawned, the premises in (105) are relevant for the deduction of
the meaning of David gestured to leave:
13 For discussion of negative binding constraints and the R⃰ relation, see §14.4.3.2.
582 functional and anaphoric control
[leave] λx. : p𝜎 ⊸ l𝜎
leave (x)
x1
(106) [David-gesture] λP. David (x1 ) : l𝜎 ⊸ g𝜎
gesture (x1 , P)
x1
(107) [pro-leave] : l𝜎
leave (x1 )
x1
David (x1 )
(108) [David-gesture-pro-leave] x2 : g𝜎
gesture (x1 , )
leave (x2 )
Since David gestured to leave is a root sentence like Chris yawned, we must also
make use of the meaning constructor labeled [context-mod] in (105), reflecting
its role as a context modifier. Combining [context-mod] with the meaning con-
structor in (108) produces the context-modifying meaning constructor in (109):
14 As in (84), the discourse referents in (107) have been renumbered to prevent a clash with the
discourse referents in (106).
partial (“quasi-obligatory”) control 583
(109) x1
David (x1 )
[David-gesture-pro-leave-mod] x2 : (g𝜎 context) ⊸ (g𝜎 context)
gesture (x1 , )
leave (x2 )
The discourse context for this sentence is the meaning constructor labeled
[context] in (105), which was derived in (101). The meaning constructor in (109)
functions as a modifier on this context, producing an updated context which
includes the contribution of the sentence David gestured to leave. Combining the
meaning [David-gesture-pro-leave-mod] with [context], and rewriting indices
as needed, we arrive at the following meaning for the discourse:
x1 , x2
Chris (x1 )
yawn (x1 )
(110) David (x2 ) : (g𝜎 context)
x3
gesture (x2 , )
leave (x3 )
Finally, we find a licit antecedent for the anaphoric discourse referent x3 . By the
negative constraint in the lexical entry for gesture, the discourse referent x2 is
not a licit antecedent for x3 . The discourse referent x1 is available to serve as the
antecedent of x3 , so the pragmatic resolution function R may apply to the index
corresponding to x3 to establish an anaphoric relation to the index of x1 :
x1 , x2
Chris (x1 )
yawn (x1 )
(111) David (x2 ) : (g𝜎 context), R(3) = 1
x3
gesture (x2 , )
leave (x3 )
involved in obligatory anaphoric control, but stronger than those involved in arbi-
trary anaphoric control. Partial control is exemplified in the following sentences
(Haug 2013):
In (112a), there is an inclusion relation between the controlled argument and its
controller in the matrix clause: the subject of gather is a group, for example a
committee, which includes the Chair of that committee. The same is true in (112b):
the subject of meet must be a group of at least two individuals, which may or
may not include Mary, but must include John. In (112c), the situation is slightly
different: this sentence is felicitous in a situation where Mary herself does not want
to be directly involved in the bombing, but wants some salient and related referent
to bomb Hanoi, for example her country’s airforce.
Partial control has been investigated by a number of authors, including Landau
(2000, 2004) and Grano (2015). Within LFG, the first author to address this
phenomenon was Asudeh (2005), who proposed an analysis involving func-
tional control. Haug (2013) shows that a functional control analysis is not viable,
and provides a semantically-based anaphoric control analysis within the PCDRT
framework adopted in this work. Haug (2013, 2014a) proposes to account for
partial control by enriching the pragmatic resolution function A in such a way as to
license relations other than identity between anaphor and antecedent, for example
a relation of inclusion. Specifically, Haug proposes that the validity of the relation A
between anaphor and antecedent depends on an additional relation B : the relation
A is valid if and only if an appropriate B relation holds between the antecedent and
the anaphor. For example, we can represent the meaning and control resolution of
example (112a) as follows:
x1
chair (x1 )
(113) x2 , A(x2 ) = x1 ; B = λx.λy.chair(x, y)
prefer (x1 , )
gather.at.six (x2 )
Here, the antecedent relation holds between x2 and x1 , because the relation
λx.λy.chair(x, y) holds between x1 and x2 . In all of the examples discussed in the
previous sections, the A relation indicates coreference, and this can be modeled as a
default interpretation of B as λx.λy.x = y. The possibilities of anaphoric resolution,
which are significantly increased by the extension of B beyond coreference, are
constrained in different ways by the specific semantic properties of different
control verbs; for example, a constraint associated with prefer ensures that the
controlled argument in (112a) includes the logophoric center of the control verb,
in this case the subject.
the controller in anaphoric or functional control 585
The first requirement is that the controller is required to be a term: either subj,
obj, or obj𝜃 . As demonstrated by Bresnan (1982a), this makes several strikingly
correct predictions.
First, it accounts for what is known as Visser’s Generalization (Visser 1963–1973;
Bresnan 1982a), according to which there is no passive version of a verb involving
subject control:15
15 As discussed by van Urk (2013), impersonal passives such as the following constitute an exception
to this generalization:
(a) It was decided to leave.
(b) Es wurde versucht, Eichhörnchen zu fangen.
it was tried squirrels to catch
‘It was tried to catch squirrels.’
Such examples constitute an exception to Visser’s Generalization, but do not violate the constraint in
(114a): there is no explicit controller, so whatever the correct analysis of such sentences, they cannot
involve functional or obligatory anaphoric control.
586 functional and anaphoric control
Bach’s Generalization follows from the constraint in (114a), since the controller
must be present as a term argument in a control construction.16 Of course, Bach’s
Generalization does not apply to arbitrary anaphoric control constructions, since
there are no syntactic constraints on the determination of the controller of the
subject of the subordinate comp in an arbitrary anaphoric control construction:
16 Many verbs which show object control entirely lack a detransitivized version, even when they have
no controlled complement:
(a) Louise believed ⃰(Tom) regarding his accident.
(b) Louise convinced ⃰(Tom) of her innocence.
Nevertheless, some such verbs do permit detransitivized versions outside control contexts, supporting
the validity of Bach’s Generalization:
(c) Louise taught (them) about anaphoric binding.
control in adjuncts 587
for a controller is obj𝜃 if there is one, otherwise obj if there is one, and otherwise
subj. The following control constructions obey this rule:
Verbs like promise constitute exceptions to this default, since the subj and not the
obj is the controller:
Both of the generalizations in (114) are best thought of as constraining the deter-
mination of grammatical functions for verbs involving functional or anaphoric
control; for discussion of the theory of mapping between semantic roles and
grammatical functions, see Chapter 9.
In (123) and (124), the dative form of the open xadj adjuncts einum ‘alone’ and
drukknum ‘drunk’ appears. Andrews (1990a) shows that the subjects of these open
adjuncts are functionally controlled by a term argument of the matrix verb. In
particular, in (124) the object Sveini ‘Svein’ appears in dative case, as the verb mœtti
‘met’ requires, and the subj of the dative adjunct drukknum ‘drunk’ is controlled
by the matrix dative obj Sveini.
We propose the rule in (125) for xadj in Icelandic, which allows for the xadj
phrase drukknum ‘drunk’ to be adjoined to a VP:18
(125) VP −→ VP AP ⃰
↑=↓ ↓ ∈ (↑
(↑ | | }) = (↓
According to this rule, the subj of the open adjunct xadj is identified with a term
argument of the matrix clause: a subj, obj, or obj𝜃 . We also propose the lexical
entry in (126) for the adjective drukknum ‘drunk’, which requires a dative subject:
The case constraint imposed by drukknum is satisfied by the dative noun phrase
Sveini:
NP A
N drukknum
Sveini
Like other adjuncts, an open adjunct xadj such as walking the dog combines with
the clause it modifies, producing a new modified meaning of the same semantic
type. In (129), the predicate during relates the interval at which the subordinate
clause event occurs to the interval at which the main clause event occurs, and the
sentence means that during the interval at which Chris was walking the dog, the
event of Chris seeing Tracy occurred.19 The f-structure and meaning constructor
for (129) are:
VP
↓ ∈ (↑ xadj)
IP
(↑ subj) = (↓ subj)
[during-xadj]
V& NP I&
V NP N VP
saw N
Tracy
The VP in the rule in (132) has three annotations that are crucial for our current
discussion. The set-membership expression ↓ ∈ (↑ xadj) requires the f-structure
for the VP to appear as a member of the xadj set of the mother IP. The equation
(↑ subj) = (↓ subj) means that the subj of the xadj phrase walking the dog is
the same as the subj of the matrix clause. The third annotation, [during-xadj],
abbreviates the following meaning constructor:
(135) Meaning constructor premises for Walking the dog, Chris saw Tracy:
[walk] λx.λy.walk (y, x) : d𝜎 ⊸ [c𝜎 ⊸ w𝜎 ]
[dog] dog : d𝜎
[see] λx.λy.see (y, x) : t𝜎 ⊸ [c𝜎 ⊸ s𝜎 ]
[Chris] Chris : c𝜎
[Tracy] Tracy : t𝜎
[during-xadj] λP.λQ.λx.during (P(x), Q(x)) : [c𝜎 ⊸ w𝜎 ] ⊸ [[c𝜎 ⊸ s𝜎 ] ⊸ [c𝜎 ⊸ s𝜎 ]]
We begin by combining the premises labeled [walk] and [dog], producing the
meaning constructor labeled [walk-dog]:
(137) [xadj-walk-dog]
λQ.λx.during (walk (x, dog ), Q(x)) : [[c𝜎 ⊸ s𝜎 ] ⊸ [c𝜎 ⊸ s𝜎 ]]
Next, we combine the premises labeled [see] and [Tracy], producing the meaning
constructor [see-Tracy]:
control in adjuncts 593
(139) [xadj-walk-dog-see-Tracy]
λx.during (walk (x, dog ), see (x, Tracy )) : c𝜎 ⊸ s𝜎
Finally, we combine the meaning constructor in (139) with the remaining meaning
constructor [Chris], producing the following semantically complete and coherent
meaning constructor, as desired:
IP
NP I
N I S
karnta ka-rla V NP S
wangka-mi N NP V
ngarrka-ku N jarnti-rninja-kurra-(ku)
ngurra-ngka-rlu
Simpson and Bresnan (1983) provide evidence from case agreement to demon-
strate that the kurra construction involves anaphoric rather than functional control
and therefore that the modifying adjunct phrase ngurra-ngka-rlu jarnti-rninja-
kurra-(ku) ‘trimming it in camp’ is an adj and not an xadj. As they point out, (141)
contains an adjunct phrase ngurra-ngka-rlu ‘in camp’ which has erg casemarking
-rlu. Such adjunct phrases agree in case with the subj of the clause they modify.
Here, the phrase ngurra-ngka-rlu ‘in camp’ modifies the subordinate adjunct
clause jarnti-rninja-kurra-(ku) ‘trimming it’. Therefore, the erg casemarking on
the modifier ngurra-ngka-rlu ‘in camp’ shows that the subj of the subordinate
clause jarnti-rninja-kurra-(ku) ‘trimming it’ is also erg.
control in adjuncts 595
However, the matrix obj phrase ngarrka-ku ‘man’, which anaphorically controls
the subordinate subj, is dat and not erg, in accordance with the case requirements
imposed by the matrix verb wangka-mi ‘speak’. This difference in case require-
ments between the matrix controller and the subordinate clause controllee shows
that the f-structures of the controller and the controllee are different, and that
anaphoric and not functional control is involved.
In (142), the subj of the adjunct phrase is anaphorically controlled by the matrix
subj ngarrka ‘man’, as required by the affix karra on the subordinate clause verb
jarnti- ‘trim’: the example means that the man is whistling while he is trimming
a boomerang. Since this analysis involves anaphoric control, we represent the
meaning of (142) in DRS form, as follows:
x1 , x2 , x3
(143) , A(x3 ) = x1
during ( boomerang (x2 ) , man (x1 ) )
trim (x3 , x2 ) whistle (x1 )
The treatment of the predicate during in (143) is the same as in §15.8.2. It is the ver-
bal affix karra which contributes the meaning during , specifying (approximately)
that the interval corresponding to the event of trimming the boomerang occurs
during the interval corresponding to the whistling event. We assume a slightly
simplified representation of the noun phrase man, treating it as an indefinite (like
boomerang).
596 functional and anaphoric control
The first line of this lexical entry specifies the pred value for the verb jarnti-rninja-
karra ‘trim’, and the second line provides the trim predicate: like any transitive verb,
this verb requires a meaning for its obj and a meaning for its subj in order to
produce a meaning for the entire sentence.
The third line of this lexical entry represents part of the meaning contribution
of the affix karra, the during predicate:
20 The expression (adj ∈ ↑) exemplifies inside-out functional uncertainty, discussed in §6.1.3; the
use of the set membership symbol ∈ as an attribute is discussed in §6.3.1.
control in adjuncts 597
meaning of the adj clause, and Q corresponds to the unmodified meaning of the
main clause.
The fourth line of this lexical entry provides the f-structure ‘pro’ value
for the pred of the subj. The fifth line establishes the antecedency relation
between the s-structure index value of the subj of the adjunct clause,
((↑ subj)𝜎 index), and the s-structure index value of the subj of the matrix
clause, (((adj ∈ ↑)subj)𝜎 index):
Finally, the last line represents a pronominal meaning constructor [pro] for the
subj of jarnti-rninja-karra ‘trim’, defined as follows:
x1
(148) [pro] λP. ⊕ P(x1 ) : ∀H.((↑ subj)𝜎 ⊸ H) ⊸ H
This is identical to the meaning [pro] used in the analyses of obligatory and
arbitrary anaphoric control provided previously. Instantiating this meaning con-
structor according to the f-structure labels in (144), we obtain the meaning
constructor in (149), for which we reuse the label [pro]:
x1
(149) [pro] λP. ⊕ P(x1 ) : ∀H.(p𝜎 ⊸ H) ⊸ H
The meaning constructors in (150) are relevant in the analysis of example (142):
[whistle] λx. : m𝜎 ⊸ w𝜎
whistle (x)
x1
[man] λP. ⊕ P(x1 ) : ∀H.(m𝜎 ⊸ H) ⊸ H
man (x1 )
x1
[boomerang] λP. ⊕ P(x1 ) : ∀H.(b𝜎 ⊸ H) ⊸ H
boomerang (x1 )
598 functional and anaphoric control
In order to obtain the correct reading for this sentence, it is necessary first to intro-
duce hypothetical meanings for the arguments of the main verb (see §8.8.1.4).21
We therefore make use of the hypothetical meaning constructors a : [p𝜎 ] and
b : [b𝜎 ] to obtain a complete meaning for the predicate trim, labeled [trim-ab]:
(151) [trim-ab] : t𝜎
trim (a, b)
We likewise hypothesize the meaning constructor c : [m𝜎 ] for the subject of the
verb whistle, to produce the meaning constructor [whistle-c]:
(152) [whistle-c] : w𝜎
whistle (c)
The predicate during can then apply to these two propositions, to yield the meaning
constructor labeled [whistle-trim]:22
(153) [whistle-trim] : w𝜎
during ( , )
trim (a, b) whistle (c)
This meaning constructor is of the correct form for the meaning constructor [man]
to apply to it; combining these produces the meaning constructor [man-w-t]:
x1
man (x1 )
(155) [man-w-t] : w𝜎
during ( , )
trim (a, b) whistle (x1 )
21 This is necessary in order to ensure that the predicate during does not scope over the discourse
referents x1 , x2 , and x3 .
22 We simplify here by treating during as a relation between DRSs, which incorrectly allows for the
possibility for discourse referents to scope inside the DRS arguments of during . The problem would
disappear if we adopted a more sophisticated analysis in which during is a relation between event
variables (as introduced in §8.10), since there would then be no subordinate DRSs as arguments to
during .
control in adjuncts 599
x1
man (x1 )
(156) [man-w-t-b] λy. : b𝜎 ⊸ w𝜎
during ( , )
trim (a, y) whistle (x1 )
Now the meaning constructor [boomerang] can be applied to the meaning con-
structor [man-w-t-b], producing the meaning constructor [man-bmr]:
x1 , x2
man (x1 )
boomerang (x2 )
(157) [man-bmr] : w𝜎
during ( , )
trim (a, x2 ) whistle (x1 )
x1 , x2
man (x1 )
: p𝜎 ⊸ w𝜎
boomerang (x2 )
(158) [man-bmr-a] λz.
during ( , )
trim (z, x2 ) whistle (x1 )
The resulting meaning constructor is of the correct form for the meaning construc-
tor [pro] to be applied to it, producing a complete and coherent meaning for the
clause:
x1 , x2 , x3
man (x1 )
boomerang (x2 )
(159) [man-bmr-w-t-a] : w𝜎
during ( , )
trim (x3 , x2 ) whistle (x1 )
x1 , x2 , x3
man (x1 )
boomerang (x2 )
(160) : w𝜎 , A(x3 ) = x1
during ( , )
trim (x3 , x2 ) whistle (x1 )
In this example, the obj of the verb, him, also functions as the possessor of the
obl obj the cheek. Lødrup (2009) discusses possessor raising constructions in
Norwegian, analyzing them in terms of functional control. Further discussion of
control structures in Norwegian is provided by Lødrup (2001, 2002a,b, 2008b,
2017a,b), among others.
further reading and related issues 601
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
clausal coordination 603
IP Cnj[main] IP
NP I! and NP I!
N VP N VP
Chris V David V
yawned sneezed
The c-structure and f-structure in (1) and (2) are constrained by the rule in
(3) for IP coordination. We adopt the complex c-structure category Cnj[main]
(with parameter main) for conjunctions like and; below, we introduce the category
Cnj[pre] for preconjunctions such as both in both Chris and David. The preliminary
rule in (3) does not account for certain aspects of the f-structure of coordinate
structures, including features contributed by the conjunction and; in §16.3, we pro-
vide a coordination rule which incorporates this aspect of the syntax of coordinate
structures:
This rule makes use of the Kleene plus operator +, which licenses one or more
occurrences of IP. Thus, this rule allows a coordinate IP to consist of one or more
IPs, followed by a conjunction, followed by the final IP conjunct.
604 coordination
The functional annotations on this rule use the set-membership relation symbol
∈ to specify that each f-structure corresponding to an IP conjunct is a member
of the set of f-structures corresponding to the mother IP; set descriptions are
discussed in more detail in §6.3. The diagram in (4) shows the relation between
the c-structure and the f-structure of the example under discussion:
NP I and NP I
N VP N VP
Chris V David V
yawned sneezed
Given the rule in (6), the c-structure and f-structure for the coordinated verbs
selected and hired are:
Both of these verbs are transitive, requiring a subj and an obj. In the example under
discussion, the coordinate V is the head of V! , and the f-structure corresponding
to V! (with obj structure not yet included) is the set labeled c, as the annotations
in (8) indicate:
V V
Cnj[main] N
↓∈↑ ↓∈↑
In (8), the equation on the NP node dominating David refers to the f-structure
corresponding to the V! node—the set c—and requires the f-structure for David
to be the obj of that set. These requirements are summarized in (9), where the
f-structure for David is labeled d, and the annotation on the NP node is instantiated
to (c obj) = d:
(c obj) = d
Bresnan et al. (1985b) and Dalrymple and Kaplan (2000) provide a definition of
function application to sets that allows us to interpret an equation like (c obj) = d
when c is a set: in such a situation, d is required to be the obj of each member of
606 coordination
the set c. This is because governable grammatical functions like obj are distributive
features, as defined in (189) of §2.5.8 and (64) of §6.3.2, repeated here:1
Thus, the constraints in (9) entail that d is the obj of each f-structure in c:
By the same reasoning, Chris is the subj of the set c in (5) and is thus the subj of
each member of c, as shown in (5).
Peterson (2004a) observes that the treatment of grammatical functions as dis-
tributive features nicely accounts for this pattern:
The verbs fax and email require a subj (Kate), an obj (the results), and an oblgoal
(to Paul), and these arguments are shared in the coordinate structure in (13a),
as shown in (14a). In contrast, (13b) is ungrammatical for the same reason that
(13c) is ungrammatical: dislike requires an obj but not an oblgoal . The incoherent
f-structure for (13b) is shown in (14b):
1 Peterson (2004a: 655) provides a definition similar to (11), but restricted to situations in which a
is a grammatical function: in other words, for Peterson, the only distributive features are grammatical
functions.
predicate coordination 607
(15) [In den Wald ging der Jäger] und [fing einen Hasen].
into the forest went the hunter and caught a rabbit
‘The hunter went into the forest and caught a rabbit.’
608 coordination
The annotated phrase structure and f-structure for (15) are shown in (17), with the
subject der Jäger ‘the hunter’ distributed over both conjuncts:
(17) [In den Wald ging der Jäger] und [fing einen Hasen].
into the forest went the hunter and caught a rabbit
‘The hunter went into the forest and caught a rabbit.’
CP
CP
CP
↓∈ ↑ Cnj[main]
↓ ∈↑
(↓ ↑
PP
↓∈ (↑ C und C
(↑ ↓ and
In den Wald C VP C VP
into the forest
NP NP
ging
(↑ ↓ (↑ ↓
went caught
einen Hasen
der Jäger
a rabbit
the hunter
’
−
syntactic properties of coordinate structures 609
This allows for the correct c-structural and f-structural analysis of such examples,
with only a minimal addition to the annotation on the standard c-structure rule
for CP coordination.
Frank (2006) also discusses the discourse-functional properties of this construc-
tion, showing that the special properties of this construction are related to general
discourse subordination effects also found in modal subordination (Frank 1997);
this provides an explanation of some of the other syntactic, semantic, and discourse
properties of the construction.
2 Huddleston and Pullum (2002: Chapter 15) provide more discussion of preconjunctions and
what they call correlative coordination: both . . . and, either . . . or, neither . . . nor. Constraints on this
construction are complex: for example, although both is not permitted in IP coordination, either is
permitted:
(a) ⃰Both [it is raining] and [the sun is shining].
(b) Either [it is raining] or [the sun is shining].
We leave a full treatment of the syntax of correlative coordination for future work.
610 coordination
The features preconj and conj are classified as nondistributive features. Treating
the conj feature as a nondistributive feature of the coordinate structure also
ensures that different conjunctions do not appear in the same coordinate structure:
for example, Chris and David or Tracy may not be treated as a single coordinate
structure with three conjuncts, but must be analyzed as nested coordination, either
Chris and [David or Tracy] or [Chris and David] or Tracy. The second line of the
lexical entry for both contains a constraining equation ensuring that both does
not appear with conjunctions other than and, allowing both tall and thin and
disallowing ⃰both tall or thin.
The rule in (21) permits the analysis of coordinated verbs with a preconjunction
(Chris both yawned and sneezed), and associates the information contributed by the
conjunction and any preconjunctions like both or either with the f-structure for the
coordinate phrase:
Given the lexical entries in (20) and the rule in (21), the c-structure and f-structure
for the phrase both selected and hired are:
Cnj[pre] V Cnj[main] V
↑=↓ ↓∈↑ ↑=↓ ↓∈↑
In (22), the preconj and conj features are attributes of the coordinate structure,
as required.
syntactic properties of coordinate structures 611
The coordination rule in (21) also permits more complex nested coordinate
structures like either [selected and hired] or [interviewed and rejected], which has
the following c-structure and f-structure:
Cnj[pre] V Cnj[main] V
↑ =↓ ↓∈↑ ↑ =↓ ↓∈↑
V Cnj[main] V V Cnj[main] V
either or
↓∈↑ ↑ =↓ ↓∈↑ ↓∈↑ ↑ =↓ ↓∈↑
V NP
↑=↓ (↑ ↓
Cnj[pre] V Cnj[main] V
N
↑=↓ ↓∈↑ ↑=↓ ↓∈↑
V Cnj[main] V V Cnj[main] V
either or David
↓∈↑ ↑=↓ ↓∈↑ ↓∈↑ ↑=↓ ↓∈↑
(25) David introduced [[Chris] [to Tracy]] and [[Matty] [to Ken]].
nonconstituent coordination 613
In (25), the sequence Chris to Tracy is not a phrase structure constituent; nev-
ertheless, it can be coordinated with the sequence Matty to Ken, which is not a
constituent either.
Maxwell and Manning (1996) propose a theory of nonconstituent coordination
that accounts for the grammaticality of examples like (25). Their account has its
basis in Wasow’s Generalization:
(27) VP −→ V NP PP
Maxwell and Manning propose that we can think of the right side of this rule as
being divided into two portions, which we will call VP-x for the first part and x-VP
for the second part:
To analyze (25), we assume that the first half of the VP rule analyzes V, and the
second half analyzes the sequence NP PP; for other constructions, a different split
of the rule might be necessary:
(29) VP-x −→ V
x-VP −→ NP PP
Crucially, rules of coordination can refer to the partial phrase structure con-
stituents that result from splitting rules in this way. This allows the following
c-structure analysis of (25):
614 coordination
NP I!
N VP
introduced NP PP and NP PP
On Maxwell and Manning’s theory, any phrase structure rule can be broken up
into two or more parts in this way. However, the only rules that can refer to
these partial phrase structure constituents are rules of coordination. Therefore, the
partial constituents that result from rule splitting play no other role in the grammar
besides their role in the analysis of nonconstituent coordination.
In analyzing more complex examples, a c-structure rule may be broken into
more than two pieces. For the purposes of this example, we assume this VP rule:
(31) VP −→ V NP NP CP
We then break the VP into three parts. The first part of the VP, VP-x, analyzes V;
the second part, x-VP-y, analyzes the sequence NP NP; and the third part, y-VP,
analyzes CP:
These rules allow an analysis of the following example, in which there is shared
material both before and after the coordinate structure:
(33) David bet Chris five dollars and Matty ten dollars that it would rain.
IP
NP I!
N VP
bet NP NP and NP NP
that it would rain
Chris five dollars Matty ten dollars
nonconstituent coordination 615
Other examples show greater degrees of complexity. Examples (31) and (33)
involve partial constituents that are all dominated by a single mother node;
Maxwell and Manning also discuss examples in which more than one rule is
involved in the split, treating these examples by the use of a stack of partial
constituents that must be combined in a compatible way.
In formal terms, Maxwell and Manning (1996) state their proposal by reference
to the state of the finite-state automaton that corresponds to the regular expression
on the right-hand side of a phrase structure rule.3 In constructions involving
nonconstituent coordination, the automaton can stop in a particular state in a
phrase structure rule and can then continue from that state to analyze each
conjunct of a coordinate phrase.4 In other words, each conjunct in a coordinate
structure must constitute a valid expansion of the mother category.
Maxwell and Manning note that their analysis allows a natural treatment of
cases where each conjunct contains a different number of constituents. The only
constituent structure requirement in an example like (34) is that each conjunct
must constitute a valid completion of the VP rule, and different numbers of phrases
are allowed:
(34) You can call me [directly] or [[after three p.m.] [through my secretary]].
As pointed out by Milward (1994), such cases are problematic for some other
approaches to coordination, particularly the “3-D” approaches of Goodall (1987)
and Moltmann (1992).
3 A finite-state automaton is an abstract “machine” that advances through a string, moving from
state to state as the string is traversed. If the string is a member of the language of the regular expression
corresponding to the automaton, the automaton is in a final state when the end of the string is reached.
An automaton corresponding to the right-hand side of an LFG phrase structure rule advances through
the daughter categories it encounters, moving from state to state as the daughter phrases are analyzed.
4 As Maxwell and Manning (1996) point out, another way of thinking of the theory is in terms of the
regular expression that appears on the right-hand side of a phrase structure rule; on this view, the partial
phrase structure constituents that are involved in nonconstituent coordination must be members of
the suffix language of the regular expression representing the right-hand side of a phrase structure rule,
where the prefix consists of the phrase structure categories that precede the coordinated subconstituent.
616 coordination
NP I!
(↑ subj) = ↓ ↑ =↓
VP
David ↑ =↓
VP-x x-VP
↑ =↓ ↑ =↓
V x-VP x-VP
Cnj[main]
↑ =↓ ↓∈↑ ↓∈↑
NP PP NP PP
introduced and
(↑ obj) = ↓ (↑ oblgoal ) = ↓ (↑ obj) = ↓ (↑ oblgoal ) = ↓
NP I
N VP
We VP-x x-VP
Adv
with great care
slowly
’
618 coordination
(40) a. [A plumber] and [making a fortune] though Bill may be, he’s not going to
be invited to my party. [NP and VP]
b. [Stupid] and [a liar] Paul undoubtedly is, but he is still my friend.
[AP and NP]
c. [In town] and [itching for a fight] is the scourge of the West, Zitty Zeke.
[PP and VP]
There are several reasons why these examples cannot be analyzed as noncon-
stituent coordination. First, the coordinate structures in these examples are not
partial constituents which are completed by other constituents in the immediate
context, and so they cannot be analyzed in terms of the rule-splitting analysis
proposed by Maxwell and Manning (1996). Second, it is not possible in general
for partial constituents in nonconstituent coordination to appear in contexts like
those in (40):
(41) a. David introduced [[Chris] [to Tracy]] and [[Matty] [to Ken]].
b. ⃰[[Chris] [to Tracy]] and [[Matty] [to Ken]] though David introduced, ...
Examples such as those in (40) show the need for a rule which allows constituents
of different categories to be coordinated to form a single coordinated constituent.
What is the c-structure category of a coordinate structure with conjuncts of
different categories? Valuable clues come from predicates whose arguments must
be phrases of a particular category. As discussed in §6.10.3, Pollard and Sag (1994)
show that the verb wax requires an adjectival complement, while the verb become
requires either a nominal or an adjectival complement:
5 See Patejuk (2015: Chapter 5) for discussion of multiclausal versus monoclausal analyses of lexico-
semantic coordination in Polish, which resembles these structures in certain respects.
unlike category coordination 619
The CAT predicate, defined in (157) of Chapter 6 (§6.10.3) allows for the imposi-
tion of these constraints; the xcomp complement of wax must be of category AP,
and the complement of become must be either NP or AP:
These and other examples show that each conjunct in an unlike category coordi-
nation structure must satisfy the category requirements imposed by the predicate.
To impose the appropriate constraints in coordinate structures, Kaplan and
Maxwell (1996) and Crouch et al. (2008) propose that the CAT predicate is
distributive, and must hold of each conjunct in a coordinate structure. Under this
assumption, the requirements in (44) produce the correct result for constructions
involving these verbs: the complement of wax must be a phrase of category AP
or a coordinated phrase in which all conjuncts are of category AP, while the
complement of become must be a phrase of category NP or AP, or a coordinated
phrase in which each conjunct is either NP or AP.
Treating CAT as distributive allows for checking of category requirements
imposed by predicates like wax and become, but still leaves open the question
of the c-structure category of an unlike category coordinate structure such as a
Republican and quite conservative in (45a). Patejuk (2015) proposes the rule in (46),
in which YP and ZP are metacategories ranging over full phrasal categories, while
620 coordination
UP is a category label standing for Unlike Phrase for all instances of unlike category
coordination:
On this view, all unlike category coordinate structures have the same c-structure
category: here, UP. This approach is problematic in that it complicates the grammar
by requiring introduction of the category UP wherever unlike category coordina-
tion structures are allowed, in addition to standard categories such as NP or AP.
Peterson (2004a: 652) proposes that the category of an unlike category coordi-
nate structure is the same as the category of the first conjunct:
In this rule, X and Y are metacategories ranging over all categories; X is used as the
category label of the mother node as well as the category label of the first conjunct.
As noted by Patejuk (2015), this leads to the expectation that the distribution of
a coordinate phrase mirrors the distribution of its first conjunct. This proposal
runs into problems originally noted by Sag et al. (1985: 141), where a coordinate
structure does not have the same distribution as its first conjunct. The coordinate
structure longwinded and a bully is acceptable as a copular complement:
However, this coordinate structure does not have the distribution of a normal AP,
and cannot appear in other places in which AP can appear:
The problem for each of these proposals is that the category of the coordinate
phrase does not reflect the categories of any of the conjuncts (Patejuk 2015)
or the categories of the non-initial conjuncts (Peterson 2004a). As discussed by
Dalrymple (2017), this makes it difficult to enforce category-function correlations
and to control the distribution of phrases of different categories. For example, if
we assume a standard annotated phrase structure rule that associates the comp
grammatical function with a CP verbal complement, the Patejuk and Peterson
proposals allow conjuncts of categories other than CP in an unlike category
coordination patterns crosslinguistically 621
coordinate structure bearing the grammatical function comp. We can prevent this
by introducing CAT constraints throughout the grammar to control the category
of the conjuncts in unlike category coordination, but the result is a considerable
complication of the grammar.
Dalrymple (2017) provides an analysis of unlike category coordination which
assumes that the CAT predicate is not distributive. Instead, the phrase structure
category of an unlike category coordinate structure reflects properties of the cate-
gories of each of the conjuncts, and the CAT predicate constrains the category of
the coordinate structure as a whole. Similar to the view of category labels advocated
by Sag et al. (1985), Marcotte (2014), Bresnan et al. (2016), and Lovestrand and
Lowe (2017), this proposal assumes that category labels are feature structures with
features encoding phrasal category as well as bar level, the status of the category as
functional or lexical, and whether the category is a projecting or a nonprojecting
category. The category label in a coordinate structure is determined on the basis
of the category features of each of the conjuncts: for example, a noun phrase bears
the feature n with value +; an adjective phrase bears the feature adj with value
+; and a coordinate structure containing a noun phrase conjunct and an adjective
phrase conjunct bears both features, n + and adj +. Predicates such as wax and
become place requirements on the feature composition of their complement by use
of the CAT predicate: wax requires an adjectival complement, with value − for all
features except for the feature adj, and become requires an adjectival or nominal
complement, with value − for all features except for the features adj and n. See
Dalrymple (2017) for further discussion and exemplification.
As observed by Bayer (1996), the successful treatment of category features in
unlike category coordination is very much dependent on the feature inventory
that is assumed: Bayer discusses difficulties in the analysis of predicates which
require either an NP or S complement under Sag et al.’s assumptions, where NP
has the features [+N, −V], and S has the features [−N, +V], with no features
in common. In fact, the same issues in the definition and representation of
f-structure features discussed in §2.5.1 arise for c-structure category features. For
example, Bresnan et al. (2016: chapter 6) propose a two-dimensional classification
of c-structure features as “predicative” (verbs and adjectives) and “non-predicative”
(prepositions and nouns), and as “transitive” (verbs and prepositions) and “non-
transitive” (nouns and adjectives). These features successfully capture the phrase-
internal properties of these categories: verbs and prepositions can take objects,
but nouns and adjectives generally do not, for example. However, they do not
straightforwardly allow for the statement of selection requirements when these
phrases appear in coordinate structures.
6 The terminology used by Haspelmath (2004b) for these four types is somewhat confusing: he
uses the term monosyndetic coordination for what we call “one fewer conjunction than the number
of conjuncts,” treating the one-conjunction type as monosyndetic coordination with ellipsis of all but
one conjunction, and he uses the term bisyndetic coordination for what we call “one conjunction for
each conjunct.”
7 Placement of the conjunction ’a in Chechen is governed by complex rules; see Jeschull (2004) for
a full discussion.
8 See §6.8 for definition and discussion of rule macros.
coordination patterns crosslinguistically 623
For English, the structures represented by the rule macros one-conj and one-
fewer-conj are both used; we thus provide a single macro english-coord which
allows either the “one conjunction in the coordinate structure” pattern or the “one
fewer conjunction than the number of conjuncts” pattern, which we will build on in
our discussion of the semantics of coordination. This macro requires a coordinate
structure to consist of an optional preconjunction,9 an initial conjunct, any number
of medial conjuncts, and a final conjunction and conjunct phrase. The medial
conjuncts must either all have conjunctions (as defined by the rule macro @cnj(_c)
or all lack conjunctions (the simple category _c):
This macro allows the statement of coordination rules for a range of c-structure
categories as in (52), exemplifying with V and IP:
Constraints lexically associated with the preconjunctions both and either prevent
their appearance with certain phrase structure categories. For instance, as shown
in (19), both cannot appear with coordinated nouns (⃰The both [sandwich] and
[soup] are cold), though it can appear with coordinated noun phrases (Both [the
man] and [the woman] sneezed); this constraint is reflected in the final line of this
lexical entry for both, which disallows coordinate structures with category N:
9 The preconjunction both requires the coordinate phrase to have only two conjuncts; in light of
noncoordinate examples involving the related determiner both in examples like both (the) girls, which
refers to two girls, we take this to be a semantic fact, and we do not encode this constraint in the
c-structure rule. Prescriptively, the preconjunction either is sometimes also claimed to require only
two conjuncts, but examples with three or more conjuncts are commonly found: I’ll either [call out] or
[bang on the door] or [blow my whistle] (Huddleston and Pullum 2002, 2005).
624 coordination
The rule macro in (52) is sufficient to account for the syntactic patterns which
we have examined so far. We will augment this rule macro in two ways in the
remainder of this chapter. First, in our discussion of the compositional semantics of
coordination, we designate one of the conjuncts as the “seed” and treat it specially
in semantic composition; this requires a separate macro encoding the special
semantic role of the “seed” conjunct. Second, in our discussion of feature resolution
patterns in nominal coordination, we propose additional constraints to ensure that
the appropriate resolved features are assigned to a nominal coordinate structure.
For the sentence Chris yawned and David sneezed to be true, it must be the
case both that Chris yawned and that David sneezed. We assume the following
f-structure and meaning constructor for this example:
To obtain this result, we adopt the compositional analysis of Asudeh and Crouch
(2002a),11 which can be intuitively described as follows. One of the conjuncts in a
10 Since we are not concerned with anaphoric relations in this chapter, we do not need the expressive
power of PCDRT, and we use predicate logic as our meaning language.
11 Dalrymple (2001) follows Kehler et al. (1995) in proposing an analysis that uses the of course
operator of linear logic, written as an exclamation point, to turn off resource accounting in the analysis
coordination and semantic composition 625
coordinate structure is the “seed,” or the initial resource contributed to the meaning
of a coordinate structure. A special meaning constructor assigns the meaning of
this conjunct to be the initial meaning of the coordinate structure as a whole.
The other conjuncts are associated with meaning constructors which add their
meanings to the “seed” one by one, with the result that the meanings of all of the
conjuncts are combined to produce the final, complete, and coherent meaning of
the coordinate structure.
We begin our discussion by inspecting the meaning constructor in (57), which
is associated with the seed conjunct in the phrase structure rule for clausal coor-
dination. Its effect is to consume the resource associated with the seed conjunct
daughter (↓𝜎 , of type t) and reassociate its meaning with the coordinate structure
as a whole (↑𝜎 , also of type t). Note that this meaning constructor is specified as
applying only to conjuncts of type t; we extend this treatment to coordination
of conjoinable types (types ‘ending in t’) in §16.7.2, and we discuss nominal
coordination and group formation in §16.8.2.
of coordinate structures. Asudeh and Crouch (2002a) observe that this analysis goes against the spirit
of the glue approach, which has a strong commitment to the preservation of resource sensitivity in
meaning composition, and so we do not adopt that analysis here.
12 This raises a resource issue. In §16.6, we noted the “one conjunction for each conjunct” pattern
of coordination, where multiple conjunctions can appear in a single coordinate structure. If each
conjunction contributes a meaning constructor to the same coordinate structure, but only one meaning
constructor is needed for interpretation of the coordinate structure, resource surplus results. We can
address this issue by providing a more complex lexical entry such as (a), which uses an instantiated
symbol (§5.2.2.1) as the value for the conj feature; recall that an instantiated symbol is like a semantic
form in that it takes on a unique value for each instance of use. The lexical entry in (a) contains a
disjunction: in the first case, the conj feature is contributed with an instantiated (unique) value, and
the meaning constructor is also contributed; in the second case, a constraining equation checks that
another instance of the conjunction has contributed a unique value for the conj feature. This ensures
that only one instance of the conjunction contributes a meaning constructor.
(a) and Cnj[main] { (↑ conj) = and
∧ : (↑𝜎 coord-rel)
| (↑ conj) =c and }
In the following discussion, for simplicity, we do not discuss the “one conjunction for each conjunct”
pattern, and we use the less complicated lexical entry in (58).
626 coordination
(58) Lexical entries for and and or, including meaning constructors
and Cnj[main] (↑ conj) = and
∧ : (↑𝜎 coord-rel)
or Cnj[main] (↑ conj) = or
∨ : (↑𝜎 coord-rel)
We also require meaning constructors which have the effect of adding the meaning
of the non-seed conjuncts to the meaning of the seed to produce the meaning
for the full coordinate structure.13 The meaning constructor for each non-seed
conjunct consumes the meaning resource of the conjunct (associated with ↓𝜎 ,
which is of type t) and combines the meaning of the conjunct with the meaning of
the coordinate structure derived so far (associated with ↑𝜎 ), using the meaning of
the conjunction (associated with (↑𝜎 coord-rel)) to combine them.14 This means
that we must also manage the meaning resource contributed by the conjunction,
which may need to be reused several times in a coordinate structure with more
than two conjuncts: for this reason, the coord-rel resource is both consumed and
produced by this meaning constructor, so that it can be consumed and produced
again by another conjunct:
To produce a complete and coherent meaning for the sentence, the meaning
constructor in (58) for the conjunction must be consumed in the course of
the derivation, like other meaning resources. According to Asudeh and Crouch
(2002a), this is accomplished by the meaning constructor for the seed conjunct,
which appears in its final form in (60). In fact, the meaning of the seed conjunct
consumes the resource (↑𝜎 coord-rel) contributed by the conjunction, but does
not incorporate the meaning C of the conjunction; the resulting meaning reflects
only the meaning P of the seed conjunct:
13 Winter (2007) provides an illuminating discussion of the complex issues raised by the distinction
between binary coordination (as assumed here and by Asudeh and Crouch 2002a), where a coordinate
structure with three conjuncts A, B, C is given a binary analysis and (A, and (B, C)), as opposed to a “flat”
treatment like and ({A, B, C}). Winter provides evidence indicating that the binary analysis is superior,
though it may be that some of the advantages he adduces for binary structure are better explained by
assuming syntactic ambiguity rather than ambiguity of semantic composition. There is no obstacle to
the formulation of a “flat” treatment of coordination in a glue setting, though we do not provide the
details here; indeed, a “flat” treatment would be simpler in some respects, in that it would not require
multiple copies of the conjunction in assembling the meaning of the coordinate structure.
14 The meaning constructor in (59) produces the meaning ∧(yawn (Chris ), sneeze (David )) for exam-
ple (1), with ∧ as a predicate taking the meanings of the two conjuncts as its arguments. This is a
notational variant of the expression yawn (Chris ) ∧ sneeze (David ), and we will use the two variants
interchangeably.
coordination and semantic composition 627
We now associate the [seed0] and [conj0] meaning constructors with the phrase
structure rule for coordination, exemplifying with a modified version of the rule
macro for English coordination given in (52). Following Asudeh and Crouch
(2002a), we treat the final conjunct as the seed, though nothing rests on this
choice; the first conjunct or any other conjunct could be chosen instead. Given
this choice, the final conjunct is associated with a different meaning constructor
from the nonfinal conjuncts, and so we define a new macro final-conj for the
final conjunct, associated with the [seed0] meaning constructor; the non-final
conjuncts are associated with the [conj0] meaning constructor:
We now go through the glue deduction for the clausal coordination Chris yawned
and David sneezed:
In this example, the subject Tracy of the conjoined verbs sang and danced is shared
across the conjuncts in the coordinate structure, as described in §16.2. Argument
sharing in subsentential coordination presents a special challenge for a resource-
based account of the syntax-semantics interface.
In (64), the subject Tracy is shared by the verbs sang and danced, and each
verb requires a meaning contribution from its subject. Our theory of meaning
assembly and the syntax-semantics interface relies crucially on the assumption that
the meaning constructor for Tracy makes a single, unique meaning contribution.
Clearly, however, the acceptability of (64) entails that this single meaning contri-
bution can satisfy the requirements imposed by each of the verbs in the coordinate
structure sang and danced. Therefore, in the analysis of examples like (64), we
require a theory of resource management in argument sharing that accounts for
the grammaticality of examples like Tracy sang and danced while maintaining the
desirable properties of our linear-logic-based glue approach.
In fact, reliance on a theory of resource management in the analysis of examples
like (64) is of paramount importance: the acceptability of (64) does not indicate
that resource accounting is completely abandoned for the shared argument Tracy.
If resource accounting were switched off entirely for shared dependents—for
example, by prefixing the meaning constructor for Tracy with the linear logic of
course operator, allowing it to be used any number of times—we would have no
way of accounting for the unacceptability of examples like:
must in fact be enforced for (65), as it is in all other cases. We require, then, a
complete and explicit theory of resource accounting, argument sharing, and their
interactions.
Particular care must be taken in the treatment of certain kinds of arguments
shared across conjunctions. As noted by Partee (1970), a sentence like Someone
sang and danced has the following meaning:
Here, the single quantifier someone scopes over the coordinate structure, and the
variable x bound by the quantifier appears as an argument of both sing and dance .
This meaning is not the same as the one for the sentence Someone sang and someone
danced:
In (67), different people might be involved in each activity, while (66) requires that
there is a single person who both sang and danced. This fact must also be captured
by our theory of argument sharing and semantic composition.
Kehler et al. (1995) were the first to propose a treatment of argument sharing
and resource management within the glue approach, and a similar approach
was adopted by Dalrymple (2001). That proposal appeals to the geometry
of f-structures in constructions involving argument sharing: intuitively, their
approach focuses on occurrences of f-structures, where an f-structure occurs more
than once if there is more than one path leading to it. Semantic resources are
associated with paths through the f-structure and thus with occurrences of f-
structures; in essence, Kehler et al. (1995) provide a means for making one copy
of a semantic resource for each f-structure path leading to its corresponding f-
structure.
For instance, in the analysis of (64), there are two paths leading to the f-structure
for Tracy, since the f-structure for Tracy appears as the value of two different
subj attributes. On Kehler et al.’s analysis, each verb requires a semantic resource
associated with the path leading to its subj. Since the f-structure for Tracy appears
at the end of each of these two paths, two copies of the meaning constructor
for Tracy are made available, and a semantically complete and coherent meaning
deduction results.
This approach successfully accounts for the acceptability of examples involving
argument sharing by allowing the creation of as many semantic resources as are
needed to satisfy the requirements of each predicate involved in the sharing of a
single argument. However, a major problem with this approach is that it also allows
resource duplication in cases where such duplication is unwarranted. For example,
constructions with raising verbs also exhibit argument sharing: the subject of a
verb like seem is also the subject of its xcomp (§15.1). As pointed out by Asudeh
(2000), we do not wish to enforce resource duplication in this case; as discussed
coordination and semantic composition 631
in §15.2.2, the derivation of the meaning seem (yawn (David )) of a sentence like
David seemed to yawn requires exactly one occurrence of the meaning resource
contributed by David, not two. Similarly, resource duplication is not warranted in
constructions in which an f-structure is involved in a long-distance dependency
and appears as a member of the dis set as well as bearing an argument function.
In these cases, relying simply on the geometry of the f-structure to license feature
duplication leads to the wrong result; a more constrained theory is needed.
Instead, we adopt a variant of the approach to resource management proposed
by Asudeh and Crouch (2002a). This approach provides special rules to combine
the semantic requirements imposed by each predicate that shares an argument into
a requirement for a single semantic resource.
In order to introduce the approach, we propose special-purpose templates
[conj1] and [seed1] for coordinated predicates which share one argument, as in
(68); these templates are special cases of the general approach to be presented later
in this section. These resemble the [seed0] and [conj0] templates in (59) and (60),
except that the meaning constructors for the conjoined phrases each require a subj
argument, and the coordinate structure as a whole requires a subj argument which
is shared across the conjunct daughters. The underlined portion of the glue side
of the meaning constructors in (68) is different from the meaning constructors in
(59) and (60) in that the resources involved are implicational, requiring a subj;
other aspects of their structure are the same. Thus, the meaning constructor in
(68a) combines two conjuncts which are each missing a subject, and produces
a resource for the coordinate structure which is still missing a subject. A single
resource satisfies the missing-subject requirement of the coordinate structure:
We now associate the components of the English coordination rule macro with
these templates. Below, we will provide a general definition that covers coordina-
tion at various clause levels.
final-cnj(_c) ≡ Cnj[main] _c
↑=↓ ↓∈↑
[seed1]
c. Alternative rule macro for English coordination, provisional:
english-coord(_c) ≡
2 3
Cnj[pre] _c _c⃰ @cnj(_c)⃰ @final-cnj
↑=↓ ↓ ∈↑ ↓∈↑
[conj1]
[conj1]
Using these modified templates and this rule, we can derive the desired meaning
for Tracy sang and danced, where the requirement for each verb to combine with a
subj is satisfied by a single resource associated with the subject Tracy:
What is needed, then, is a single rule schema which can be used for coordination
of conjuncts of any conjoinable type: clausal coordination with conjuncts of type t
producing a coordinate structure of type t, predicate coordination with conjuncts
634 coordination
of type 'e → t( producing a coordinate predicate of type 'e → t(, and so on. As
Asudeh and Crouch (2002a) point out, similar proposals have been made in many
theoretical settings (Gazdar 1980; Keenan and Faltz 1985; Emms 1990), though the
approach is most commonly associated with Categorial Grammar treatments of
coordination (Steedman 1985, 2001; Morrill 2011): Steedman (2001: 39) refers to
the “ancient intuition that coordination is an operation that maps two constituents
of like type onto a constituent of the same type.” Building on observations by
Asudeh and Crouch (2002a), we propose the following schematic definitions
of [conj] and [seed], which differ from [conj0]/[conj1] and [seed0]/[seed1]
in the underlined portion; these template schemas allow coordinate structures
of all conjoinable types (resources of type t and any resources that depend
on them):
!
In (73), the expression ' α(n represents a sequence of n (zero or more) dependents
!
that are shared across the conjuncts of the coordinate structure; x represents the
corresponding meanings. Importantly, the number of required dependents for the
seed conjunct is inherited by the coordinate structure as a whole, and the fact
that the [conj] macro consumes a resource requiring n dependents and produces
a resource which also requires n dependents ensures that valency requirements
for all conjuncts are shared. Thus, semantic completeness and coherence (§8.7.1)
ensures that requirements for each conjunct are met, with no surplus resources
remaining. For instance, the ungrammaticality of (65) (Chris selected and arrived
Tracy) is due to a mismatch of valency requirements. The seed conjunct arrived
requires only a subj argument, and the schema is instantiated so that the seed
conjunct shares the requirement for a subj argument (and no other arguments)
with the coordinate structure as a whole; the non-seed conjuncts must match this
requirement. The result is that the two verbs selected and arrived share a subj but
not an obj: the resource for Tracy is left unused, violating semantic coherence,
and the requirement imposed by selected for an obj resource is not met, violating
semantic completeness.
As before, these templates are used in the final definitions of the rule macros for
coordinate structures with conjoinable types:
noun phrase coordination 635
This macro allows any dependent that is shared across a coordinate structure to
satisfy the requirements of each conjunct.
Asudeh and Crouch (2002a) discuss a very similar general coordination schema
to the templates in (73), and present two objections to the resulting analysis. First,
they point out that not all coordination involves conjoinable types; in particular,
it has been argued that nominal coordination works differently, and cannot be
captured by this rule schema. We are in agreement with this observation, and in
§16.8.3 we propose a different set of templates for nominal conjunction; however,
this does not constitute motivation for abandoning the analysis of conjoinable
types which we have presented in this section. Second, they propose that more
control over the types involved in coordination is available if a coordination rule
is separately specified for each c-structure category. However, there is no one-to-
one relation between c-structure categories and semantic types: for example, in V
coordination, the type of the coordinate structure depends on the type of the verb
('e → t( for intransitive verbs, ''e → t( → t( for transitive verbs, and so on). This
means that a rule schema like (74) is needed even if separate rules are proposed for
each c-structure category, and we do not see any advantage to proposing a separate
schema for each c-structure category over the general schema in (74). Thus, we do
not consider either of these objections as counting against this approach.
Our theory of the representation of the values of the pers feature must account for
this pattern. As discussed in §2.5.7.2, Dalrymple and Kaplan (2000) and Vincent
and Börjars (2007) adopt a set-based representation of the pers feature, as shown
in (76). Sadler (2011) demonstrates that any set-based analysis of complex feature
values can be directly translated to an equivalent feature-based analysis, with a
positive value representing the presence of an element in the set, and a negative
value representing its absence:
noun phrase coordination 637
If we assume a set-based theory of the values of the pers feature, treating resolution
as set union of the values of the pers features of the conjuncts produces the same
pattern as in (75), as desired:
In the following, we show how to define resolution rules for the pers feature in
a feature-based setting, preserving the essential intuitions underlying the union
operation while retaining the advantages of a feature-based analysis, including
underspecification.
In (78), the resolved value in the third column has a + value for the attribute 1 or
2 whenever one or both of the conjuncts has the value + for that attribute.
In order to impose the proper constraints, we define the following basic tem-
plate, which we call plus:15
We can now define a template @pers-res which ensures the correct pers resolu-
tion patterns: if any conjunct has a positive value for the 1 or 2 attributes of the
pers feature, the set as a whole has a positive value for that attribute. Otherwise,
the set has a negative value for that attribute:
15 See §5.2.10 for discussion of implicational constraints using the ⇒ operator; see §6.3.1 for an
explanation of the symbol ∈ used as an attribute.
noun phrase coordination 639
the pers-res template with the final conjunct, but since the template refers only to
the f-structure ↑, it could appear on any of the daughters:
(82) Rule macro for English nominal coordination including person resolution
constraints:
english-ncoord(_c) ≡
2 3 4 5
Cnj[pre] _c _c⃰ @cnj(_c)⃰ @cnj(_c)
↑=↓ ↓∈↑ ↓∈↑ @pers-res(↑)
(84) Values for the gend feature in Icelandic according to Vincent and Börjars
(2007):
masculine: {m}
feminine: {f}
neuter: {}
Given these set values, gender resolution can be modeled as set intersection:
(85) Resolution of the gend feature as set intersection (Vincent and Börjars
2007):
{m} (masculine) ∩ {m} (masculine) = {m} (masculine)
{f} (feminine) ∩ {f} (feminine) = {f} (feminine)
{ } (neuter) ∩ { } (neuter) = { } (neuter)
{m} (masculine) ∩ {f} (feminine) = { } (neuter)
{m} (masculine) ∩ { } (neuter) = { } (neuter)
{f} (feminine) ∩ { } (neuter) = { } (neuter)
In this definition, _set is a set of f-structures. The template requires that the value
of index m for the set is + if all of the conjuncts have the value + for index m, and
− otherwise. The same pattern holds for index f.
The gender resolution template in (88) is associated with one of the conjuncts
in the Icelandic nominal coordination rule; because the template refers only to the
mother f-structure ↑, it can be associated with any of the conjuncts.
and the final conjunct if it follows the coordinate phrase), and so this agreement
pattern is sometimes called “closest conjunct agreement.”
Sadler (1999, 2003) examines agreement patterns in Welsh, a VSO language in
which only pronominals control agreement: the verb agrees with the pronominal
subject in (89a), and shows default (third person singular) agreement in (89b).
With coordinated subjects, verb agreement does not reflect the resolved features
of the coordinate phrase; instead, the verb agrees with the initial conjunct if it
is pronominal (examples 90a,b) and shows default agreement otherwise (exam-
ple 90c):
A potential (but incorrect) analysis of these patterns is that verb agreement does in
fact reflect the features of the coordinate structure as a whole, but that in examples
like (90), the coordinate structure is unexpectedly associated with the features of
just one of its conjuncts, rather than the resolved features. For example, on this
view the coordinate phrase i a Mair ‘I and Mair’ in (90b) would have the features
of a first person singular phrase, matching the features of the initial conjunct.
Sadler (2003) demonstrates convincingly that such an analysis is not viable,
showing that distinguished conjunct agreement patterns as in (90) can coexist in
the same clause with resolved agreement patterns. In (91), the passive auxiliary
chaffodd shows third person singular (distinguished conjunct) agreement with the
initial conjunct e ‘he’, while the nonfinite main verb is preceded by an anaphoric
form eu showing third person plural (resolved) agreement with the coordinate
phrase e a’i milwyr ‘he and his soldiers’:
Similarly, in (92) the verb gwelais shows first person singular (distinguished
conjunct) agreement with the initial conjunct i ‘I’, and the reflexive anaphor ein
hunain shows first person plural (resolved) agreement with the coordinate phrase
i a’m brawd ‘I and my brother’:
Thus, the coordinate phrase must be analyzed as having the expected resolved
features rather than the features of one of the conjuncts, since the resolved features
are referenced in auxiliary agreement in (91) and anaphoric agreement in (92).
Arnold et al. (2007) provide an in-depth discussion of Brazilian Portuguese data
that show even more complex patterns. Brazilian Portuguese allows resolved agree-
ment with the features of a coordinate structure as well as distinguished conjunct
agreement with the initial conjunct or the final conjunct. Arnold et al. provide
(93) to show that distinguished conjunct agreement with different conjuncts can
coexist in the same structure; a prenominal determiner can exhibit distinguished
conjunct agreement with the initial conjunct, while a postnominal adjective can
exhibit distinguished conjunct agreement with the final conjunct:
Thus, our theory must allow reference to the syntactic and semantic features of
a coordinated noun phrase (in resolved agreement), as well as the syntactic and
semantic features of the initial and final conjuncts (in distinguished conjunct
agreement).
Kuhn and Sadler (2007) review earlier proposals for the treatment of distin-
guished conjunct agreement, dividing them into representation-based approaches
and description-based approaches. Representation-based approaches augment the
f-structure of coordinate nominal phrases with additional features to encode
agreement information for the distinguished conjunct. Kuhn and Sadler (2007)
discuss several problems with representation-based approaches: these additional
features are purely for bookkeeping, and introduce unneeded complexity into
the f-structure representation; further, features for both the initial conjunct and
the final conjunct would be needed to handle examples like the Brazilian Por-
tuguese example in (93), where material on the left agrees with the initial con-
junct, and material on the right agrees with the final conjunct. Kuhn and Sadler
(2007) propose a description-based approach which introduces a new kind of
structure, a local f-structure sequence, in the representation of coordinate struc-
tures; they also propose an expanded classification of f-structure features as
left-peripheral, right-peripheral, or proximity-based, in addition to the standard
distributive/nondistributive classification. Dalrymple and Hristov (2010) build on
Kuhn and Sadler’s insights in a revised proposal that does not depend on an
noun phrase coordination 643
[everyone] λS.every (x, person (x), S(x)) : ∀H 't( .[e𝜎 'e( ⊸ H 't( ] ⊸ H 't(
[yawn] λz.yawn (z) : e𝜎 'e( ⊸ y𝜎 't(
16 Not all cases of nominal coordination are best treated as group-forming; a standard example which
must be treated as conjunction of predicates is the president and CEO, which refers to a single individual
who is both the president and the CEO (the (x.president (x) ∧ CEO (x))). For discussion of such examples,
see Bergmann (1982), Dowty (1988), King and Dalrymple (2004), and Heycock and Zamparelli (2005),
among others.
644 coordination
Notably, the semantic structure e𝜎 for the NP everyone corresponds to the e-type
variable bound by the quantifier: it is a syntactically accessible reflex of the e-type
component of a quantifier meaning. When a quantifier appears as a conjunct in
nominal coordination, it is the e-type component of its meaning that corresponds
to the syntactic structure involved in coordination. This means that our proposal
does not require the complicated machinery for meaning assembly in quantifier
coordination which is required by other analyses, including Hoeksema (1988).
We adopt a variant of Asudeh and Crouch’s (2002a) proposed semantics for
e-type coordination:
This sentence means that a group consisting of the individuals x and y met, where
x is a student and y is a professor. We can also coordinate a proper name with an
indefinite:
Here, the group that met consists of y, a professor, and the individual David .
This approach works well for these simple cases. As noted by Schwarzschild
(1994: 22), complications arise in the treatment of other cases; in particular, the
quantifier no is problematic:
As with our analysis of clausal coordination, we assume that one of the conjuncts is
the seed; again, we treat the final conjunct as the seed, though nothing depends on
this choice. The meanings of the non-seed conjuncts are then added to the meaning
of the coordinate structure one by one. Partee and Rooth (1983) note that it is
not generally assumed that there are “e-conjoinable” types (types “ending in e”)
other than e itself (in particular, Montague (1973) does not propose any additional
“e-conjoinable” types), so we provide only the definition for e-type conjunction
here, and do not allow for complex types “ending in e.”
The meaning for (99) has a structure similar to a generalized quantifier, with
constraints on the composition of X analogous to the restriction of the quantifier,
and the predication on X as its scope (§8.4.1.4). Thus, for nominal coordination
we introduce a semantic structure attribute coord-restr of type t, associated with
constraints on the composition of the group X (for example, that David and Tracy
are individual parts of X). There are two meaning constructors associated with the
seed, which we call [nseed1] and [nseed2]. The meaning constructor [nseed1]
requires the meaning of the seed conjunct (here, Tracy) to be a component of
646 coordination
Thus, we have the following deduction of the meaning of David and Tracy met,
using the f-structure labels introduced in (99):
The meaning constructors for this example combine as shown in (103). The
general form of the proof is very close to (101), and in fact the first three steps are
exactly the same, since in both examples the non-seed conjunct is David. We do not
show the details of the combination of the determiner a with the noun professor;
see §8.8.2 for discussion and illustration of how the meaning constructors for
determiners and nouns are combined:
The meaning deduction for an example such as (96), in which two quantifiers are
coordinated, proceeds similarly; we do not provide the proof here.
The challenge in the analysis of such structures is to ensure that each conjunct bears
the appropriate f-structure role. For in-depth discussion and analysis of lexico-
semantic coordination from an LFG perspective, see Gazdik (2011) and Patejuk
(2015). Patejuk (2015) goes beyond lexico-semantic coordination to provide a
more general discussion and analysis of examples in which the conjuncts differ in
some respect, including not only conjuncts with different grammatical functions,
but also conjuncts with the same grammatical function but different cases or
different c-structure categories; see also Przepiórkowski and Patejuk (2012) and
Patejuk and Przepiórkowski (2012).
The relation between apposition and coordination is explored in detail by Sadler
and Nordlinger (2006a), who illustrate their discussion with examples from a
range of Australian languages. Building on this foundation, Nordlinger and Sadler
(2008) and Sadler and Nordlinger (2010) explore the use of sets as an appropriate
representation not only of coordinate structures, but also of apposition, generic-
specific constructions, inclusory constructions, and other related constructions.
Some early LFG work on coordination is based on syntactic assumptions that
have since been abandoned. Kaplan and Maxwell (1988b) proposed a theory of
function application in coordination that does not rely on a distinction between
distributive and nondistributive features or on the definition presented in (11) for
application of a function involving a distributive feature to a set; instead, they pro-
pose that the properties of a set are defined as the generalization of the properties
of its elements (see §6.9.3 for definition and discussion of generalization). Many
of the predictions of the Kaplan and Maxwell (1988b) theory are indistinguishable
from the theory presented in this chapter. However, Kaplan and Maxwell’s theory
makes unwanted predictions involving constraining properties of sets, and also has
difficulty in cases where a set has a property that is different from its conjuncts; see
Dalrymple and Kaplan (2000) for more discussion.
17
Long-distance dependencies
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
syntax of long-distance dependencies 653
We do not include the features topic and focus at f-structure, instead representing
them at the separate level of i-structure (Chapter 10). Rather, we use the overlay
function dis to represent long-distance dependencies in f-structure in this book;
see §2.1.11. We therefore assume that the Extended Coherence Condition applies
to dis rather than to topic and focus: dis must be linked to the semantic predicate
argument structure of the sentence in which it occurs, either by functionally or by
anaphorically binding an argument.
In this section, we will examine the syntax of the topicalization construction in
English. Examination of this construction reveals c-structural constraints on the
permitted constituent structure categories of the fronted constituent as well as f-
structural constraints on the path relating the fronted constituent to its within-
clause grammatical function. Constraints on the topicalization construction in
other languages may differ: as we will see, other languages may allow a different
654 long-distance dependencies
set of phrasal categories to appear in the relevant position or may place different
constraints on the f-structural relation between the displaced constituent and its
within-clause role.
The sentences in (2) exemplify the permitted range of phrase structure categories
in this construction:
In this sentence, Chris is the obj of the verb like, but this phrase is displaced: it does
not appear in the usual immediately postverbal position associated with objects in
English. The f-structure for Chris is a member of the dis set, and there is also a path
through the f-structure defining its within-clause grammatical function; in (4), the
path is obj. In (5), the displaced phrase is the obj in the subordinate clause, comp,
so that the path leading to the within-clause function of the displaced phrase is
comp obj:
syntax of long-distance dependencies 655
Not all within-clause functions can be related to a member of the dis set, however.
As discussed in §6.6, for some speakers it is possible for a displaced phrase to be
related to a position within the comp of a so-called “bridge verb” like think, but
not to a position within the comp of a nonbridge verb like whisper. The distinction
between bridge and nonbridge verbs is not reflected in the grammatical function
of the clausal complement: in both cases, the path to the within-clause argument
is comp obj. Instead, this requirement is stated as an additional condition on the
f-structures in the extraction domain which applies in the case of those speakers
who identify a difference in grammaticality that is related to the bridge/nonbridge
verb distinction.
Where bridge and nonbridge verbs are judged to be distinct in this respect, we
use the f-structure attribute ldd (for “long-distance dependency”) to distinguish
between the two. The value − for the feature ldd is lexically specified by a
nonbridge verb like whisper as appearing in its comp. Such f-structures cannot
participate in a long-distance dependency. This captures the unacceptability of (6)
for the relevant set of speakers, since the comp f-structure contains the attribute-
value pair $ldd, −%:
−
656 long-distance dependencies
In (6), the path relating the f-structure for Chris to its within-clause function goes
through the comp f-structure of the verb whisper, which has the value − for the
attribute ldd; this is disallowed. We demonstrate later in this section how, for the
relevant dialects, this constraint can be captured in formal terms.
A number of other constraints on long-distance dependencies were originally
explored by Ross (1967) and have since been the subject of intense scrutiny and
debate. Ross identified a number of constraints involving “islands”: areas of a
sentence from which extraction of an element to form a long-distance depen-
dency is prohibited. Among these constraints is the Sentential Subject Constraint,
according to which a long-distance dependency cannot involve a position inside a
sentential subject—that is, subjects are islands for extraction.
This constraint is simply stated: the path to the within-clause function of dis
may not include subj. Other constraints on long-distance dependencies can be
characterized similarly, either as constraints on grammatical functions permitted
on the path or as constraints on attributes and values in f-structures through which
the path passes.
There is little consensus on the proper characterization of long-distance depen-
dencies involving modifying adjuncts. The situation appears to be similar to the
one we have reported for bridge and nonbridge verbs when it comes to variability
in speakers’ judgments. For example, Williams (1992: 18) claims that examples
like (8), which involve a relation between an initial question phrase and a position
inside an adverbial modifier, are “marginal though possible”:
However, Cinque (1990) and Hornstein and Weinberg (1995) count as ungram-
matical examples that are very similar to (8):
Other examples involving a dependency between a member of the dis set and a
position inside an adj are not acceptable. For example, the unacceptability of (11)
shows that extraction from a tensed sentential modifier is not permitted:
Other languages place different constraints on the topicalization path. For example,
Kroeger (1993: Chapter 7) shows that in Tagalog, the topicalization path must con-
sist exclusively of subjects: only a subj may be extracted, and the only function out
of which extraction is permitted is subj. This is true not only for the topicalization
construction but also for other long-distance dependencies in Tagalog; the path
in the constituent question construction and the relative clause construction must
also contain only subj attributes.1
If the displaced phrase Chris appeared in the specifier of CP, it would be expected
to appear before the complementizer that, not after it. Bresnan et al. (2016: 9) show
that topicalized phrases are adjoined to IP in English:
IP
NP IP
N NP I
Chris N VP
we V
like
Here we use the metacategory abbreviation TpczP for the phrase structure category
of the fronted phrase, where Tpcz stands for topicalization. We also use the
functional abbreviation TpczPath for the path through the f-structure to the
within-clause function of the fronted phrase.
660 long-distance dependencies
The set of phrasal categories that can participate in the topicalization construc-
tion in English is given in (3) of this chapter. On this basis, we define TpczP for
English in rule (15) as the following disjunction of categories:
We must also properly constrain the path through the f-structure that defines
the within-clause grammatical function of the displaced phrase. Formally, this
relation involves functional uncertainty, discussed in §6.1.2. In (12) above, a set of
constraints on the long-distance path for topicalization in English were outlined.
We can formally characterize these constraints as follows.
We start with the Sentential Subject constraint, exemplified in (7) and defined
as SSC:
The initial portion of the path defined in (17) may not include subj,2 though it
may end in subj or any other grammatical function gf. The possibility for deep
embedding is represented by the Kleene star operator ⃰ permitting any number of
attributes defined as members of the metacategory gf other than subj on the path.
The Bridge Verb Constraint, which seems to apply for a subset of speakers and
was exemplified in (6), is captured via the expression defined as BVC in (18). The
definition relies on an off-path constraint (§6.6) referring to a feature of any f-
structures on the path, (→ ldd) += −:
The off-path constraint in (18) permits any path relating the displaced phrase
to its within-clause grammatical function as long as it does not pass through an
f-structure bearing the attribute ldd with value −.
Finally, according to the Tensed Adjunct Constraint exemplified in (11), the path
defining the within-clause grammatical function of a displaced phrase may include
a nontensed adj function, but not a tensed adj. The following definition of the path
TAC accomplishes this:3
2 The relative-difference operator for regular expressions, written as a minus sign (−), is defined and
discussed in §6.1. The expression gf − subj represents any grammatical function (gf) except subj, and
[gf − subj]⃰ represents any sequence of grammatical functions that does not contain subj.
3 In §6.3.1, we discussed the use of expressions in which the set membership symbol ∈ is used as an
attribute to allow reference to some member of a set.
syntax of long-distance dependencies 661
The off-path constraint (→ tense) requires a member of the adj set to contain the
attribute tense. Thus, the path within the square brackets ([gf⃰ adj ∈ gf⃰])
(→ tense)
picks out all paths which pass through a tensed adjunct. The complement operator
for regular expressions, written as ¬, picks out all paths which do not obey these
constraints: in other words, paths which do not go through a tensed adjunct. (See
§6.1 for discussion of regular expressions and complementation.)
Together, the constraints in (17), (18), and (19) express restrictions which apply
to the path through the f-structure that defines the within-clause grammatical
function of the displaced phrase. Putting these three constraints together via
intersection (§6.1), we are able to define TpczPath in English thus:
This expression uses the Kleene plus operator + to indicate that the relevant path in
Tagalog consists of at least one occurrence of subj: only a subj may be involved in
establishing a long-distance dependency, and only a subj may contain a displaced
phrase. LFG research has not yet established a complete typology of possible paths
in long-distance dependencies: future research will no doubt reveal more about
the possible range of crosslinguistic and cross-constructional variation in long-
distance paths, as well as universal generalizations concerning constraints on long-
distance paths.
662 long-distance dependencies
In (23), the topicalized phrase is the obj of both hates and likes. The across-the-
board constraint falls out as a consequence of our theory of coordination and
long-distance dependencies. Since grammatical functions are distributive features
(§6.3.2), asserting that an f-structure is the obj of a set means that it is the
obj of each member of the set. That is, resolving the path to the within-clause
grammatical function of the topic correctly entails that the topicalized phrase must
bear a grammatical function within each conjunct of the coordinate phrase.
Kaplan and Maxwell (1988b) discuss examples like (24), in which the topicalized
phrase David is the obj of the first conjunct and the objtheme of the second
conjunct:
syntax of long-distance dependencies 663
These examples show that the fronted phrase can bear different grammatical
functions within each conjunct.4 To analyze these examples, Kaplan and Maxwell
(1988b) provide the definition of functional uncertainty given in (27) of §6.1.2,
repeated here:
(f a suff(a, η)) = v
for some symbol a, where suff(a, η) is the set of suffix strings s such that
as ∈ η.
4 Kaplan and Zaenen (1989b) discuss constraints on the long-distance path in coordinate structures
in Japanese, based on work by Saiki (1985): the grammatical function of the fronted constituent must
either be subj in all conjuncts, or a nonsubject function in all conjuncts. Similar constraints hold in
English.
(a) ⃰Who did Chris think [David met ] and [ saw Matty]?
Kaplan and Zaenen show how the relevant path can be defined so as to predict these facts. Falk (2000)
also discusses these patterns and provides an alternative analysis.
664 long-distance dependencies
5 Unlike right-dislocated full noun phrases, ProTags do not have a clarifying function. See Mycock
(2017) for details.
syntax of long-distance dependencies 665
also §4.2.3), we do not adopt this approach, but use instead the overlay function
dis in the analysis of long-distance dependencies and instances of dislocation. As
an overlay function, members of the dis set must be integrated into the structure
of a clause via an association with a primary, nonoverlay function, consistent with
the basic principle underlying the Extended Coherence Condition.
The second dependency involves the relative pronoun and its position, possibly
embedded, within the fronted phrase. We follow Kaplan and Bresnan (1982) in
representing this syntactic dependency at f-structure; the f-structure of the relative
pronoun appears as the value of the attribute relpro within the relative clause.
Similar representations have also been adopted by Butt et al. (1999) and Falk
(2001b).
The c-structure and f-structure for the phrase a man who Chris saw are:
’
NP
D N
a N CP
N NP C
man N IP
who NP I
N VP
Chris V
saw
In (30), the relative pronoun appears in initial position in the relative clause, and
its f-structure is both a member of the set that is the value of dis and the relpro
of the relative clause.
Example (31) shows that the relative pronoun can also appear as a subcon-
stituent of the initial phrase. Here the relative pronoun whose is a subconstituent
of the fronted phrase whose book:
syntax of long-distance dependencies 667
NP
D N
a N CP
N NP C
man D N IP
whose N NP I
book N VP
Chris V
read
In (31), the value of the dis attribute is a set with one member, the f-structure of the
fronted phrase whose book, and the value of the relpro attribute is the f-structure
of the relative pronoun whose. We examine syntactic constraints on both of these
dependencies below.
We propose the following phrase structure rules for the analysis of these
examples:
(32) N, −→ N, CP∗
↑ =↓ ↓ ∈ (↑ adj)
(33) CP −→ RelP C,
↓ ∈ (↑ dis) ↑=↓
(↑ RTopicPath) = ↓
(↑ relpro) = (↓ RelPath)
(↑ relpro prontype) =c rel
668 long-distance dependencies
The first two annotations on the RelP daughter in (33) are similar to the annota-
tions on the rule for English topicalization constructions in (15). The constraint
↓ ∈ (↑ dis) requires the f-structure corresponding to the RelP node to appear as
a member of the dis set in the f-structure. The constraint (↑ RTopicPath) = ↓
ensures that the f-structure of the initial phrase also fills a grammatical function
within the clause, constrained by the long-distance path RTopicPath; we define
RTopicPath below.
The third and fourth annotations constrain the value of the relpro attribute
in the relative clause f-structure. The constraint in the third line, (↑ relpro) =
(↓ RelPath), requires the relpro f-structure to appear at the end of the path
RelPath within the f-structure of the RelP. Below, we provide a definition of
RelPath that properly constrains the role of the relative pronoun within the
fronted phrase. Finally, the constraint (↑ relpro prontype) =c rel is a con-
straining equation (§5.2.8) requiring the value of the relpro attribute to have a
prontype attribute with value rel: the value of the relpro attribute must be a
relative pronoun.
Therefore, we define RelP for English in the rule in (33) as the following disjunction
of categories:
(38) Bridge Verb Constraint violation (ungrammatical for some but not all
speakers)
a. ⃰Chris, we whispered that David saw.
b. ⃰a man who you whispered that David saw
We therefore propose that the same constraints restrict both the English
RTopicPath and the long-distance path in topicalization constructions. The
expressions in (20) and (41) are exactly the same:
(42) [Reports [[the height of the lettering on the cover of which] the government
prescribes ]] should be abolished.
Ross (1967) originally used the term pied piping in the transformational analysis
of these constructions: in moving to the front of the sentence, the relative pronoun
lures some additional material along with it, like the Pied Piper of Hamelin lured
rats and children along with him as he left Hamelin.
670 long-distance dependencies
In all of these examples, the phrase structure category of the fronted phrase is one
of the categories defined by RelP, and no constraints on RTopicPath are violated.
Example (43a) shows that the relative pronoun can itself appear in fronted position;
in such a case, RelPath is the empty path. Examples (43b–c) indicate that the
relative pronoun can appear as a possessor phrase, filling the poss role in the dis
f-structure, or as the possessor of a possessor. It can also appear as the object of an
oblique argument, as in (43d–f), or embedded inside an oblique argument, as in
(43g), though it may not fill the poss role inside an oblique phrase (43h).7 It can
appear as the object of a fronted adjunct phrase (43i), though it may not appear as
an adjunct inside the fronted phrase (43j).
Given these facts, we propose the following definition of RelPath in English:
7 Louisa Sadler (p.c.) notes that some speakers judge constructions such as (43h) to be acceptable.
syntax of long-distance dependencies 671
is resumed by a phrase in the matrix clause. Butt et al. (2007) provide an analysis
of Urdu correlatives, illustrated in (45a), while Belyaev and Haug (2014) consider
correlatives in Ossetic (a language in which direct objects bear either nom or gen
case marking), illustrated in (45b):
The syntactic analysis of correlatives is fundamentally the same for these exam-
ples. The correlative clause appears initially8 and bears the dis function, while
simultaneously having the appropriate within-clause grammatical function. (At
i-structure, the correlative clause bears the discourse function topic.) Belyaev and
Haug (2014) provide the following f-structure analysis of the Ossetic correlative
construction in (45b):9
8 Butt et al. (2007) note that in Urdu, the subordinate clause containing the relative expression
can appear either before or after the matrix clause; following Srivastav (1991), they classify as true
correlatives only structures in which the subordinate clause appears initially.
9 Belyaev and Haug (2014) use the attribute stype with value correl to mark correlative clauses.
672 long-distance dependencies
’
’
NP C
N C IP
Who does NP I ’
N VP
David V
like
syntax of long-distance dependencies 673
To analyze constructions like (47), the following simplified rule was proposed in
§6.1.2:
(48) CP −→ XP C,
↓ ∈ (↑ dis) ↑=↓
∗
(↑ comp gf) = ↓
This rule ensures that the phrase in the specifier position of CP bears the dis
function and also fills a grammatical function within the utterance. We now refine
this rule to take into account constraints on the phrase structure category of
the fronted phrase and to give a more complete characterization of the path to
its within-clause function. We also introduce the q attribute, whose value is the
f-structure of the possibly embedded interrogative pronoun within the initial dis
phrase, analogous to the relpro attribute for relative clauses: for example, in the
analysis of the question Whose book did you read?, the f-structure for the fronted
phrase Whose book is a member of the dis set, and the question word whose
provides the value for the q attribute. See Kaplan and Bresnan (1982) for more
discussion of the q attribute.
We use the constituent structure metacategory QuesP and the functional abbre-
viations QDisPath and WhPath in the following reformulation of rule (48):
(49) CP −→ QuesP C,
↓ ∈ (↑ dis) ↑=↓
(↑ QDisPath) = ↓
(↑ q) = (↓ WhPath)
(↑ q prontype) =c wh
The annotations on the QuesP node in rule (49) are similar to those on the relative
clause rule in (33). The first two annotations require the f-structure corresponding
to the QuesP node to fill the dis role and also to bear some grammatical function
defined by the long-distance path QDisPath. The third annotation requires the
value of the q attribute to appear at the end of the long-distance path WhPath
within the dis f-structure; we discuss constraints on WhPath below. The fourth
annotation requires the prontype attribute of the q f-structure to bear the value
wh, ensuring that an interrogative pronoun plays the q role.
17.1.3.1 Category of the displaced phrase The first issue is the correct defi-
nition of QuesP: which phrasal categories can appear as a dis constituent in the
specifier of CP? These examples are wellformed:
Therefore, we can assume the same definition for QDisPath in English as we did
for TpczPath in (20):10
10 The path definitions assumed here are sufficient for present purposes, but Postal (1998) discusses
certain differences between these two types of path which may require further complications and
refinements.
syntax of long-distance dependencies 675
(61) a. ⃰[The height of the lettering on the cover of which report] does the govern-
ment prescribe?
b. the report [the height of the lettering on the cover of which] the government
prescribes
(66) Who did you say that, just a minute ago, sneezed?
One approach to the CAE constraint is proposed by Falk (2002b, 2006), framed
within his pivot/gf 4 approach to subjecthood (see §2.1.5). Recall that Falk (2006)
defines pivot as the grammatical function that connects its own clause to other
clauses. In (65a), for instance, the pivot is who. To account for the CAE constraint,
Falk proposes the following generalization:
(69) CAE constraint according to Falk (2006: 133), associated with the comple-
mentizer that:
𝜙−1 (↑ pivot) ⇒ ↑ →f (↑ pivot)
11 𝜙−1 is the relation that maps f-structures to c-structure nodes: see §6.10.3. 𝜙−1 (↑ pivot) is an
existential constraint that holds if there is some c-structure node corresponding to (↑ pivot) via the
𝜙−1 relation.
syntax of long-distance dependencies 677
intervening adverb does not change the f-structure relations which hold between
the pivot and the complementizer.
In §6.11.2, we presented Asudeh’s (2009) analysis of the CAE constraint: briefly,
Asudeh proposes a constraint, to be included in the lexical entry of the comple-
mentizer that, which rules out the ungrammatical sequences of words on the basis
of string precedence. Asudeh (2009) cites the Adverb Effect as a reason why the
CAE constraint must be treated as a precedence-based phenomenon rather than
a dominance-based one, as in Falk’s analysis. Asudeh’s (2009) own analysis of the
CAE constraint makes crucial reference to linear order relations and does account
for the Adverb Effect. His formulation of the CAE constraint is given in (179) of
§6.11.2, repeated here:
Asudeh’s CAE constraint rules out a situation in which the word following the
complementizer has a subject which is both phonologically realized and the head
of an unbounded dependency. When an adverb follows the complementizer that,
as in (66), the constraint in (70) is satisfied.
the question word bearing the grammatical function obj is displaced to initial
position, and hence also bears the function dis.12
(72) { [David] }
{ [who] }
{ [like] }
In some languages, such as Russian, the focus status of all question phrases is
signaled syntactically; that is, they occupy a syntactic position associated with the
discourse function focus and hence all appear initially, as in (71b). Of course,
focus status need not be marked by syntactic means only. Mycock (2006) argues
that the focus status of question phrases in Japanese, which appear in situ (71c), is
marked prosodically. Constituent questions in this language do not involve a long-
distance dependency and question phrases do not bear the function dis, but at
i-structure—like their Russian counterparts—all question words are members of
the focus set, according to Mycock (2006). What of the English-type formation,
which involves a single question phrase bearing the function dis appearing in ini-
tial position and other question phrases in situ, as in (71a)? Mycock claims that in-
situ question words in English are prosodically marked for focus status, like those
in Japanese. Therefore, while the c-structures for English, Russian, and Japanese
multiple constituent questions differ according to whether or not question phrases
bear the function dis, in each case the i-structures are fundamentally the same:
they contain a focus set which includes all of the question words in the sentence.
Whether all in-situ languages should be analyzed in the same way is an interest-
ing question. In his LFG analysis of Chinese constituent questions, Huang (1993)
shows that although these constructions do not involve long-distance dependen-
cies, the same sorts of scope phenomena found in English constituent questions
can also be found in Chinese. Huang uses inside-out functional uncertainty
(§6.1.3) in his analysis to capture the similarities between questions in English
and Chinese. Mycock (2006) considers an apparent island effect in Japanese, but
it emerges that this may be an issue of prosody rather than syntax (see Deguchi
and Kitagawa 2002). It remains to be determined beyond doubt whether in-situ
languages may be subject to exactly the same kinds of syntactic constraints as
languages which involve displacement of question phrases in multiple constituent
questions.
We assume that only question words that indisputably participate in a long-
distance dependency appear in the dis set at f-structure and provide the value of
the q attribute. This distinguishes displaced question words from those that appear
in situ under our approach. This difference is illustrated by the following analysis
of the English multiple constituent question in (71a), which includes one displaced
question word (what) and one in-situ question word (where):
CP
NP C ’
N C IP
What did NP I
N VP
Charlie V
V AdvP
i-structure:
put Adv
{ [Charlie] }
where [what]
[where]
[put]
The kind of “regular” constituent questions that we have considered so far are the
most well studied. LFG research has also investigated a variety of other construc-
tions that include question phrases. Mycock (2004, 2006) offers an LFG analysis
of the q-expletive (also referred to as the wh-expletive or wh-scope marking)
construction, in which a question phrase takes scope over a clause higher than the
one in which it appears. The question word’s extended scope is marked in some
languages, including Hungarian, by the presence of another element, represented
in the gloss as what:
Mycock adopts the position that only the final question word (in example (74), kit)
is the “true” question word; any higher ones are expletives associated with a “true”
question expression. The q-expletive is semantically empty, but it is syntactically
and/or prosodically prominent. For instance, in (74) the q-expletive (mit) occupies
immediately preverbal Focus position, just as the “true” question word kit does
in the most deeply embedded clause. The q-expletive’s presence thus serves to
focus “by proxy” the content of the phrase with which it is associated (which
680 long-distance dependencies
Mycock claims is the entire clause within which the “true” question words appear
in Hungarian). In this way, the q-expletive stands in for the relevant expression
in a higher clause. See Mycock (2006) for full details of an LFG analysis of scope-
marking constructions, including the q-expletive construction.
Other work on questions in LFG has considered the information structure status
of question words and phrases and how this interacts with syntactic structure.
We have seen that in “regular” information-seeking constituent questions there
is evidence that the question expression has focus status at i-structure, but there
are other possibilities. On the basis of question formation data from 15 languages,
Mycock (2013) proposes that question words can bear other discourse functions
as well (background, completive, topic), just like non-question words. For
instance, following Butt (2012), Mycock proposes to analyze question words in
echo questions as background rather than focus. This provides an explanation
for syntactic differences between these two types of constituent questions. Butt
(2014b) also deals with non-canonical constituent questions and the role that
information structure plays in the analysis of Hindi-Urdu. Butt draws on the work
of Mycock (2006) on questions and Krifka (2008) on information structure and the
concept of Common Ground Management to give an analysis of the immediately
postverbal position in Hindi-Urdu as a second focus position with particular
interpretational characteristics. Butt (2014b) also presents an analysis of polar
questions including the question marker kya in Hindi-Urdu, which provides a
similarly fine-grained approach to their distinct information structural properties.
As Dalrymple and King (2000) show, the missing object can be embedded arbi-
trarily deeply in the complement of a tough construction.
(76) This book is tough [to get her [to avoid [reading ]]].
This biclausal analysis is supported by, for instance, the interaction of the tough
predicate and a for-phrase, as shown in (79). Notice that the object in the for-phrase
can act as an overt controller of the comp subj:
Dalrymple and King (2000) also discuss syntactic constraints on the path between
filler and gap in tough constructions. Notably, it is not possible for the gap in this
682 long-distance dependencies
construction to bear any grammatical function other than obj. Dalrymple and
King (2000) show that this obj can be inside an adjunct or oblique phrase, for
instance:
Revising Dalrymple and King’s original rule slightly in order to bring it in line with
our assumptions about f-structure and the dis function, the relevant constraints
are captured in (81), which appears in the lexical entry of the tough predicate:
(82) %Mary is tough for me to believe that John would ever marry .
Note that the long-distance path specification in (81) is very different from
the path that is relevant for, for example, English topicalization (20); thus, tough
constructions neatly illustrate that long-distance dependencies and the constraints
upon them may vary not only across languages, but also within a single language.
Dalrymple and King (2000) observe that another set of predicates, including
need and require, also involve a “missing object,” but display distinct properties.
Their interaction with a for-phrase, for one, is different:
Further, the dependency involved is not arbitrarily long but bounded in the case
of need-type predicates:
In line with work by Barron (1999), Dalrymple and King (2000) analyze need-type
predicates as complex predicates that appear in a monoclausal structure; no long-
distance dependency is established. The f-structures of need-type and tough-type
predicates are therefore distinct; compare (78) with (85):
resumptive pronouns 683
(86) a. Kalle letar efter en bok som han inte vet hur den slutar.
Kalle looks for a book that he not knows how it ends
‘Kalle is looking for a book that he does not know how (it) ends.’
b. an ghirseach a-r ghoid na síogaí í
the girl comp-pst stole the fairies her
‘the girl that the fairies stole (her) away’
The challenge for any LFG analysis of resumptives is to deal with the apparent
violation of wellformedness which arises. If the resumptive pronoun is the same as
any other pronoun, then both it and the dis phrase that binds it contribute pred
values to the same f-structure, in violation of Consistency (§2.4.6.3).
Falk (2002b) proposes that a pronoun in a language with resumptive pronouns
makes one of two different f-structure contributions: in its use as a standard
pronoun it contributes a pred attribute with value ‘pro’ to the f-structure, while
in its resumptive use, it does not contribute a pred value, but its f-structure is
required to be identical with a member of the dis set. Under the second option,
the same kind of long-distance dependency is established as in the filler-gap
structures that we have examined so far. This dual nature of pronouns in languages
with resumptive pronouns accounts for the differences in syntactic constraints
restricting “ordinary” pronouns compared to those restricting resumptive pro-
nouns: according to Falk the type of dependencies involved are fundamentally
different, and therefore it is unsurprising that their properties are not necessarily
the same.
What this analysis fails to capture, Asudeh (2004, 2012) argues, is McCloskey’s
Generalization (McCloskey 2002: 192): resumptive pronouns are, in fact, ordinary
pronouns, and do not take on special forms in their use as resumptives. Asudeh
observes that under Falk’s analysis, the similarities between ordinary and resump-
tive pronouns are a matter of coincidence. Asudeh’s own analysis differs in treating
684 long-distance dependencies
This constraint ensures that the features of the dis f-structure other than pred
are identical with the features of the f-structure of its corresponding within-clause
function. At the same time, the optional off-path constraint in (87) permits, but
does not require, identification between the pred value of dis and its within-clause
function. If the optional off-path constraint applies, the effect of the restriction of
pred features in (87) is neutralized, resulting in simple identity of f-structures.
This is exactly what is required for long-distance dependencies with a gap. But
in a long-distance dependency involving a resumptive pronoun, the f-structure
of the within-clause grammatical function has its own pred, meaning that the
optional off-path constraint must not apply. Thus a single path specification, that in
(87), handles both ordinary long-distance dependencies with a gap, and those with
a SIR.
13 Following Asudeh (2012), we assume that the value of the dis attribute (which Asudeh calls udf)
in Swedish is a single f-structure, not a set.
resumptive pronouns 685
The f-structures for dis and comp subj are not identical, since they have distinct
preds; yet they share all other features, which is why the features for gend and
case, which are not contributed by vem ‘who’ but by the resumptive pronoun,
appear in the dis f-structure. Compare this with the f-structure for the equivalent
sentence without a subject resumptive, shown in (89): since there is no pronoun to
contribute a pred value for the comp subj, the optional off-path constraint given in
(87) applies, with the result that the f-structures for the dis and the comp subj are
identical; at the same time, the features gend and case are absent in (89), because
the only pronoun present does not supply these features.
lyzes the binding relationship between the SAR resumptive pronoun and the
dis f-structure as being established by the complementizer system in Hebrew.
Crosslinguistically, this is subject to variation: Asudeh argues that in Irish the
relationship is established by a particular complementizer, as we discuss in §17.3.2.
17.3.1 Kikuyu
As discussed by Zaenen (1983) and in more detail by Clements and Ford (1979),
in affirmative declarative Kikuyu sentences the verb is associated with a downstep
tone.15 This downstep tone affects the phonology of the words following the verb:
the downstep turns a sequence of low tones into high tones in the immediately
following phrasal category. In (91), downstep is associated with the two verbs ɛ́:! rírɛ́
‘tell’ and átɛ́mírɛ́ ‘cut’:
The downstep associated with ɛ́:! rírɛ́ ‘tell’ affects the words áté ‘that’ and Káriók! í
‘Kariũki’, whose citation forms are:
(92) a. ate
b. Kariokı̆
As Zaenen (1983) shows, the explanation for these differences can be ascribed to
the tonal shift imposed by the verb ɛ́:! rírɛ́ ‘tell’.
14 This analysis is based on unpublished joint work by Ron Kaplan, John Maxwell, Annie Zaenen,
and Mary Dalrymple.
15 In the following examples, high tones are marked with an acute accent, downstep tones are marked
with !, and low tones are unmarked.
688 long-distance dependencies
Interestingly, however, this tone change does not appear within the domain of a
long-distance dependency:
The question word nóo ‘who’ bears the dis function and also fills the role of the
subj of the verb otɛmírɛ́ ‘cut’ in the subordinate comp. The downstep that would
be expected to appear on both the matrix and subordinate clause verb does not
occur, and the effect that this downstep would have on the following words is
not present.16 As Clements and Ford (1979) and Zaenen (1983) show, the absence
of the downstep tone marks the domain of extraction of the fronted interrogative
pronoun. Here, the domain of extraction is the f-structure for the entire sentence.
We propose that every f-structure in the domain of extraction in Kikuyu is
marked with the attribute-value pair $ldd, +%:
(94) Each attribute on the path relating a member of the dis set to its within-
clause grammatical function must be an attribute of an f-structure that also
contains the pair $ldd, +%.
In (95), this constraint is satisfied, since the f-structures labeled t and c both contain
the attribute-value pair $ldd, +%:
’
+
c +
16 The high tone on the initial syllable of ate ‘that’ in (93) is spread from the final high tone of the
preceding word Ka:nákɛ́ ‘Kanake’ by an independent rule.
morphological marking of ldd paths 689
Within the domain marked by the attribute-value pair $ldd, +%, spreading of the
downstep tone does not occur.
To constrain the path defining a long-distance dependency in Kikuyu, ensuring
that the path passes only through f-structures that are appropriately marked with
the ldd feature, we use off-path constraints, defined and discussed in §6.6 and in
§17.1. We assume a simplified definition of QDisPath for Kikuyu, representing it
as comp⃰ gf:17
17.3.2 Irish
Asudeh (2004, 2012) explores the marking of long-distance dependencies in
Irish, in which morphophonology varies according to the precise type of long-
distance dependency construction: the form of the complementizer in the extrac-
tion domain is dependent on whether the path terminates in a gap or a resumptive
pronoun. In the non-past, two of the complementizers have the same form, but
they are distinct in triggering different phonological mutations on the following
word. The complementizer used with a filler-gap construction triggers lenition
mutation (and is therefore often glossed as aL), while the one that co-occurs with a
resumptive pronoun triggers nasalization mutation (and therefore is often glossed
as aN). Asudeh (2012) cites the following examples from McCloskey (1979); notice
the effect of lenition on the forms of the verbs glossed ‘praise’:
17 It is likely that grammatical functions other than comp are allowed in the extraction domain
in Kikuyu. In a full treatment of Kikuyu long-distance dependencies, the equation in (96) should be
enriched appropriately to characterize properly the full range of permitted relations between the initial
question word and its grammatical function within the clause, as was done for English in (58).
690 long-distance dependencies
Irish relative clauses lack relative pronouns. This means that the specifier of CP
position is unoccupied in an Irish relative clause, though not in Irish constituent
questions and cleft constructions. The phrase structure rule in (98) allows for the
analysis of relative clauses as well as constituent questions and clefts with a phrase
(XP in 98) filling the specifier position. When the specifier position is unfilled, the
CP is a relative clause: it must be a member of the adj set of the phrase in which it
is contained, and it has a dis attribute containing the f-structure we would expect
for a relative pronoun.18
(98) CP −→ { XP | 𝜖 } C,
↓ ∈ (↑ dis) (adj ∈ ↑) ↑=↓
%rel ∈ (↑ dis)
(%rel pred) = ‘pro’
(%rel prontype) = rel
In the rule in (98), a local name %rel is used. When the specifier of CP is not filled
by an XP at c-structure (hence 𝜖 for the empty string), the dis set includes a member
%rel whose feature pred has the value ‘pro’, and whose feature prontype has
value rel.
The two Irish long-distance dependency constructions that Asudeh analyses
differ in another significant respect. When a long-distance dependency involves
multiple clauses and a gap, aL appears in every clause on the path between filler and
gap. By contrast, when a long-distance dependency involves multiple clauses and
a resumptive pronoun, a second pattern emerges: aN appears in the “top” clause
in the dependency (the furthest clause from the resumptive pronoun, but the one
closest to its binder), and all intervening clauses contain a third complementizer
go instead.19 The full long-distance dependency path is therefore marked in Irish,
similar to the situation in Kikuyu examined earlier, but only in the case of a filler-
gap dependency and not a resumptive pronoun, as these data from McCloskey
(1979, 2002) demonstrate:
18 For a similar analysis of the pronoun-less relative clause construction in English, see §17.6.
19 McCloskey (2002) identifies three other multi-clause complementizer patterns which he refers to
as mixed chains; see McCloskey (2002) and Asudeh (2012) for discussion of these patterns.
morphological marking of ldd paths 691
the gap, the defining relation in a filler-gap dependency. Second, in its “marking”
role in higher clauses, it establishes a link between the dis set in its clause and
the one in its comp. The latter applies in each intervening clause, resulting in the
multiple marking that we see in (99a).
Asudeh (2012) captures these features of long-distance dependency construc-
tions in Irish by means of a lexical entry like the one in (100) for the complemen-
tizer aL, which he analyzes as a nonprojecting word:20
(100) aL C
4 (↑ compform) = aL
(↑ comp⃰ gf) ∈ (↑ dis)
(→ dis) = (↑ dis)
The first constraint in the lexical entry in (100) ensures that aL specifies the
value aL for the attribute compform. The second constraint establishes the long-
distance dependency between a member of the dis set and some within-clause
function gf. The off-path constraint annotated on the comp attribute ensures that
if the within-clause function gf is embedded in one or more comps, the dis value of
each intervening comp f-structure is the same as the top-level dis value. Asudeh’s
analysis of Irish long-distance dependency path-marking thus crucially involves a
comp-to-comp relation established by the complementizer aL. As Asudeh notes,
the analysis has commonalities with the successive cyclic movement analyses of
Irish proposed by McCloskey (1990, 2002). On Asudeh’s analysis, however, there
is no movement involved; rather, a successive cyclic dependency is introduced.21
Combined with the phrase structure rule in (98), the lexical entry in (100)
enables us to formulate an analysis of a simple filler-gap dependency including
the complementizer aL in Irish:
20 For discussion and motivation for the treatment of aL as a head-adjoined nonprojecting word C 4,
see Asudeh (2012: 157–63).
21 In fact, the analysis we present in this section differs from Asudeh’s analysis in assuming that the
value of the dis attribute in Irish is a set rather than a single f-structure: the constraint in (100) identifies
the entire top-level dis set with the dis set in the subordinate comps. This does not quite express the
intuition behind Asudeh’s analysis, which is stated in terms of a successive cyclic dependency involving
only one dis f-structure, not a set of f-structures. To express his intuition more faithfully, we could
instead propose the lexical entry in (a) for aL, which requires only that one member of the dis set is
shared between the top-level f-structure and the subordinate comps:
(a) aL C
4 (↑ compform) = aL
%d ∈ (↑ dis)
%d = (↑ comp⃰ gf)
%d ∈ (→ dis)
Since we are not concerned with cases in which the dis set has more than one member (and we do not
know of situations in which such a configuration would arise), we stick with the simpler formulation
in (100).
692 long-distance dependencies
DP
D0 NP
an NP CP
the
scríbhneoir IP
writer
I0 S
C I0
na mic léinn
a the students
aL mholann
praise
The specifier position of CP is unfilled in (101), and the dis set member pred ‘pro’
is introduced by the phrase structure rule in (98). The dependency established
between this member of dis and a within-clause function (in this case, obj) is
consistent with the lexical entry in (100).
The off-path constraint in (100) further requires the dis set in each intervening
comp to be identified with the dis set in the top-level f-structure.22 This takes care
of the “path-marking” function of aL in intervening clauses:
22 In order to ensure successive cyclic marking, the possibility of the complementizer go appearing
in any intervening comp has to be ruled out by including ¬ (↑ dis) in the lexical entry for go (Asudeh
2012: 173).
morphological marking of ldd paths 693
DP
D0 NP
an NP CP
the
t-úrscéal IP
novel
I0 S
C I0 DP VP
a mheas mé CP
aL thought I
IP
I0 S
C I0 DP
a thuig
mé
aL understood
I
694 long-distance dependencies
The final piece of the puzzle is semantic and concerns how the head is understood
to be modified by the relative clause(s). We discuss the semantic analysis of relative
clauses in §17.6, including the treatment of resumptive pronouns.
XP
↓ ∈ (↑ C
(↑ ↓
traces and empty categories 695
In fact, however, an outside-in expression annotated on the filler is not the only
possible way in which a filler-gap dependency can be licensed. While in this book
we consistently adopt an outside-in functional uncertainty analysis (in line with the
vast majority of work on long-distance dependencies in LFG), it is also possible
to use inside-out functional uncertainty (§6.1.3) in constraining the filler-gap
relationship.
One of the issues with adopting an inside-out functional uncertainty approach
to long-distance dependencies is the question of what precisely the inside-out
expression should be associated with. One possibility is to assume the existence
of an unpronounced c-structure element, a trace or gap (represented as t in (104)),
with which the inside-out constraint can be associated:
XP
C
↓ ∈ (↑
..
.
XP
(↑ ↓
t
↑∈ ↑
An alternative possibility, following analyses such as Ginzburg and Sag (2000) and
Bouma et al. (2001) in the HPSG framework, is to treat the constraint as an optional
lexical specification associated with all predicates, allowing any of the dependents
of the predicate to participate in a long-distance dependency. (Recall that gf ranges
over arguments and adjuncts of a predicate.)
This approach faces complications in languages in which more than one dependent
of the same predicate can participate in a long-distance dependency: Russian
examples such as (71b), repeated in (106), would require multiple inside-out
constraints associated with each predicate.
Falk (2006) argues against assuming a uniform approach to the type of functional
uncertainty involved in long-distance dependencies. Falk proposes that long-
distance dependencies involving subj as the within-clause grammatical function
of the dependency involve an outside-in path specification, while long-distance
dependencies involving non-subj functions require inside-out specifications.23 As
discussed in §17.4.2, Falk (2006) assumes that in the case of non-subj dependen-
cies, the inside-out specification is associated not with the predicate governing the
gap position, but with the gap itself, and he therefore assumes the existence of a
trace, or empty category, in the gap position at c-structure.
Most LFG analyses which use inside-out functional uncertainty to constrain
filler-gap dependencies assume the existence of traces. As discussed in §17.4.2, we
do not accept the existence of traces in this work, and adopt a uniform outside-in
analysis of long-distance dependencies.
The name crossover comes from early transformational analyses of examples like
(107), which assumed that a violation ensues when a question phrase like who
“crosses over” a coindexed pronoun in moving to the front of the sentence: in
(107), who crosses over the coindexed pronoun his. Bresnan (1995, 1998, 2001c)
and Bresnan et al. (2016) propose a theory of prominence for anaphoric binding
that accounts for examples like (107) by assuming the presence of a trace in the
object position of like.
As Bresnan demonstrates, languages vary in imposing different types of promi-
nence requirements on the binding relation between a question phrase and the
pronouns it binds. Three prominence dimensions are relevant: the functional
23 Falk’s (2006) proposal is framed within his theory of subjecthood, according to which the function
subj is to be understood as the conflation of two distinct functions, pivot and gf.
4 As such, Falk’s claims
are stated relative to pivot vs. non-pivot functions, rather than subj vs. non-subj functions. In other
work, as here, Falk refers to the distinction in terms of subj vs. non-subj, for the sake of simplicity.
traces and empty categories 697
hierarchy (on which, for example, subj outranks obj); the thematic hierarchy (on
which agent outranks theme); and linear order (f-preceding elements outrank
the elements they f-precede). Bresnan et al. (2016) propose that in English, the
relation between the question phrase and the pronoun is governed by either the
grammatical function hierarchy (Syntactic Prominence) or the linear precedence
hierarchy (Linear Prominence), so that satisfaction of either type of prominence
requirement is sufficient. The two relevant dimensions of prominence are summa-
rized by Dalrymple et al. (2007) as follows, where the “operator” is the displaced
question phrase:
In (107), the question phrase who fills the obj role, and the pronoun his is
contained in the subj. Thus, the pronoun outranks the question phrase in Syntactic
Prominence, violating the first prominence requirement in (108).
In examining whether the Linear Prominence condition in (108) is satisfied,
we must determine whether the pronoun his f-precedes the question phrase who.
Bresnan et al. assume that constituent question formation in English involves the
presence of traces, so that the f-structure for who in (107) corresponds to two c-
structure positions: the initial position in which who appears, and the gap position
following like. They rely on the definition of f-precedence given in (187) of §6.11.4,
repeated here:
In (107), the rightmost node corresponding to the f-structure for who is the trace,
which follows the pronoun; therefore, the pronoun f-precedes the question phrase
according to the definition given in (109), and the Linear Prominence requirement
is not satisfied. Since neither the Syntactic Prominence condition nor the Linear
Prominence condition is met, (107) is correctly predicted to be ungrammatical.
Bresnan (1995, 1998), Berman (2003), and Bresnan et al. (2016) examine
weak crossover violations in German, proposing that German imposes the same
prominence conditions on binding by question phrases as English, and ascribing
differences between the two languages to variation in the presence of traces in the
two languages. The German equivalent of (107) is fully grammatical:
Berman (2003) and Bresnan et al. (2016) argue that no trace is present in German
examples like (110), since the grammatical function of the fronted argument is
determined locally, by reference to its casemarking, rather than by its position.
Since no trace appears in this clause, the pronoun does not f-precede the question
phrase; the Linear Prominence condition is thereby satisfied, accounting for the
acceptability of the example.
In contrast, they argue that traces must be assumed in extractions from sub-
ordinate clauses in German, since case information is insufficient to identify the
grammatical function of the extracted element in the embedded clause. Bresnan
et al. (2016) do not explicitly represent the trace in (111), since its position is
uncertain, though it appears somewhere within the subordinate clause:
Here, the Linear Prominence condition is not satisfied because the f-structure for
wen ‘who’ also corresponds to the position of the trace in the subordinate clause.
The subordinate clause trace position is rightmost, and is therefore the one that
is relevant in evaluating f-precedence requirements. The matrix clause pronoun
f-precedes the trace in the subordinate clause, and so the Linear Prominence
condition is not met. Since the Syntactic Prominence condition is also not met, the
sentence is correctly predicted to be ungrammatical with the indicated indexing.
Thus, according to Bresnan’s proposal, weak crossover phenomena provide a new
view of the availability and distribution of empty c-structure categories in German,
Hindi, Malayalam, and Palauan.
However, not all accounts of weak crossover appeal to the existence of traces.
Dalrymple et al. (2001, 2007) explore a different definition of the Linear Promi-
nence condition, based on unpublished work by Sag (1998). In this approach, the
Linear Prominence condition is evaluated by considering the f-precedence relation
between the pronoun and an f-structure that contains the question phrase.
The approach proposed by Dalrymple et al. (2001) is defined in terms of co-
dependenthood; co-dependents are defined as in (112):24
(112) Co-Dependents:
the arguments and adjuncts of a single predicate
(113) Let CodepOp and CodepPro be co-dependents such that CodepOp con-
tains O and CodepPro contains P. Then:
24 Dalrymple et al. (2001) use the term “coargument” for this concept, but since the term ranges over
adjuncts as well as arguments, we use the term “co-dependent.”
traces and empty categories 699
According to these requirements, it is the head of the overt phrase containing the
gap (for instance, the preposition in the case of an oblique) which determines
Linear Prominence under Dalrymple et al.’s analysis. In (114), CodepOp (the
f-structure which contains the question phrase operator) is the comp f-structure,
as labeled:25
Pron
CodepPro
CodepOp
CP
NP C
N C VP
Wen meinte NP VP
said
D N CP
seine Mutter
his mother habe sie getrösted
has she consoled
25 For clarity, in these examples displaced f-structures are depicted as the value of their within-clause
grammatical function at f-structure rather than as a member of the dis set. See §2.4.3 for a discussion
of the representation of f-structures which are the value of more than one different attribute.
700 long-distance dependencies
CP
NP C Pron
CodepPro
N C VP
Wen mag NP
likes
D N CodepOp
seine Mutter
his mother
The CodepOp’s f-structure f-precedes the pronoun’s f-structure, and the Linear
Prominence condition is satisfied. The ungrammaticality of the corresponding
English example in (107) is accounted for by the Syntactic Prominence condition,
as described above; thus, this alternative account of weak crossover violations
differs from accounts presented by Berman (2003) and Bresnan et al. (2016)
in assuming that in English, though not in German, the Syntactic Prominence
condition must always be met. On this approach, weak crossover violations are
accounted for without assuming traces.
Nadathur (2013) proposes a further refinement of the definition of Linear
Prominence in a new LFG analysis of weak crossover, based on Pickering and
Barry’s (1991) Direct Association Hypothesis, which also eschews traces. Pickering
and Barry provide psycholinguistic evidence that the relevant link in a long-
distance dependency is not between the filler and the gap, but between the filler
and the predicate (verb, preposition, or other category) that selects for the gap,
which Nadathur refers to as its anchor. She provides the following definition of
Linear Prominence:
This differs from Dalrymple et al.’s (2001) approach in treating the element which
subcategorizes for the operator, rather than a phrase containing the operator, as
participating in the relevant structural relationship.
traces and empty categories 701
17.4.2.2 Nested dependencies In many but not all languages, multiple long-
distance dependencies must not cross, but rather must be nested (Kuno and
Robinson 1972: 77):
b. Crossed dependency:
⃰This is the salami that my knife is easy to cut with .
These facts about crossed dependencies have been taken as evidence for traces,
the argument being that it is the position of the traces that is responsible for the
attested ungrammaticality. However, non-trace analyses of this phenomenon are
also possible: Dalrymple and King (2013) offer a non-trace approach.
Dalrymple and King (2013) reject an f-precedence analysis on the basis that no
definition of f-precedence makes it possible to single out unacceptable crossing
dependencies from other types of configurations involving multiple long-distance
dependencies. Instead, they offer an analysis framed in terms of operator superi-
ority, where each filler is identified as an operator. Their definition of superiority
incorporates a particular version of f-command (§6.9.1), dis-command, which
holds between operators each of which is in a dis set:
Dalrymple and King provide an initial formulation of the constraint which disal-
lows crossed dependencies:
The rule in (121) appeals to the definition of dis-command given in (119). What
remains to be defined is the superiority relation that holds between gaps (T1, T2).
For this, Dalrymple and King (2013) rely on the Direct Association Hypothesis
(Pickering and Barry 1991; Pickering 1993) and the notion of anchor discussed in
the previous section. On this view, the relevant superiority relation is not between
traces T1 and T2, but between Anchor1 (the predicate which selects for T1) and
Anchor2 (the predicate which selects for T2). Dalrymple and King (2013) define
the final version of the constraint on nested dependencies as follows:
By contrast, (124) does not satisfy the constraint in (122) because Anchor1 f-
precedes Anchor2:
Dalrymple and King (2013) also review previous work which shows that, while
crossing dependencies result in ungrammaticality in English, for example, this is
not true crosslinguistically. Crossed dependencies are permitted in Norwegian and
merely dispreferred in Swedish (Maling and Zaenen 1982). Thus, the constraint in
(122) does not apply across all languages. Dalrymple and King (2013) consider how
this variation can be accounted for within the LFG framework without recourse to
traces.
(127) a. [Which articles]i did John file i without reading i /themi ? (multiple
gap construction)
b. These are [the articles]i that [John filed i ] and [everyone else read
i /⃰themi ]. (‘across-the-board’ construction)
(128) a. ⃰[What/Which articles]i did John file the book without reading i?
(129) a. [Who/Which friend]i did you tell i that you would visit your brother?
b. [Who/Which friend]i did you tell your brother that you would visit i?
For this reason, Falk advocates the term “multiple-gap construction” to refer to
both types of example, and we adopt Falk’s terminology here. Falk (2011) provides
the following f-structure for a multiple-gap construction:
multiple-gap constructions 705
Two main issues are the subject of continuing investigation in LFG analyses of
such multiple-gap constructions. First, there is the question of how the relation
is established between a single filler and multiple gaps. As depicted in (103),
it is often assumed that filler-gap dependencies are constrained via an outside-in
specification on the filler:
(132) ⃰Whoi did you say i convinced you [ i should pass the course]?
(cf. You said Davidi convinced you (that) [hei should pass the course].)
However, Engdahl (1983) points out that the c-command requirement does not
account for all of the data: there are sentences, in whose non-gap equivalent a
reflexive can appear, that do not support multiple gaps, regardless of the fact that
no c-command relation holds between the two gaps.
Based on these patterns, Falk (2011) proposes that the appropriate constraint
is best expressed in terms of f-structure relations: a non-subject gap cannot
participate in a multiple gap construction with an f-commanding subject.
(135) man
λx.man (x)
The meaning of a noun modified by a relative clause, like man who Chris saw,
is of the same type. It represents the set of individuals who are men (the meaning
contribution of man), who are people (the meaning contribution of who), and who
were seen by Chris (the meaning contribution of Chris saw):
The meaning contribution of the relative pronoun who is redundant here, since the
fact that an individual is a man entails that he is a person. In some cases, however,
relative clauses and semantic composition 707
v[ ]
m m
s r[ ]
’
c
As described in Chapter 13, noun modifiers such as the relative clause who Chris
saw combine with the meaning of the noun that they modify to produce a new,
modified meaning of the same type. Thus, the implicational meaning constructor
in (138) is similar to the meaning constructor for a common noun like man, but
with a modified meaning reflecting the meaning of the relative clause:
(139) man
v[ ]
m m
r[ ]
λx.man (x) : v ⊸ r
708 long-distance dependencies
In the analysis of (138), we assume the relative clause rule given in (33) of this
chapter, augmented with the meaning constructor labeled [rel]:
(140) CP −→ RelP C,
↓ ∈ (↑ dis)
↑ = ↓
(↑ RTopicPath) = ↓
[rel]
(↑ relpro) = (↓ RelPath)
(↑ relpro prontype) =c rel
As discussed in §8.6, there are cases in which meaning contributions are associated
with phrase structure configurations rather than lexical items. Relative clause
formation in English is one such case: the meaning constructor [rel], which
combines the relative clause meaning with the meaning of the modified noun, is
associated with the relative clause CP rule. The definition of [rel] is given in (141):
This meaning constructor consumes the meaning resource of the relative clause
w𝜎 ⊸ s𝜎 and the meaning resource of the modified noun v ⊸ r to produce a new
noun resource, also associated with v ⊸ r but reflecting a modified meaning.
We must also provide a meaning for the relative pronoun who. We propose that
who augments the relative clause meaning by providing the additional meaning
that the entity involved is a person. This is the lexical entry for who:
We begin by combining the premises labeled [Chris] and [see] to obtain the
meaning constructor labeled [Chris-see]:
This provides the meaning resource required by the premise [who]. Combining
[Chris-see] and [who], we have the meaning constructor labeled [who-Chris-see]:
Next, we combine the premises [who-Chris-see] and [rel], producing the premise
[who-Chris-see-rel]:
As discussed in Chapter 13, this meaning constructor has the characteristic form
of a modifier: it consumes the resource [v ⊸ r] and produces a new resource of the
same form but with a modified meaning. Combining [who-Chris-see-rel] with
[man], we have, as desired:
According to this rule, when no RelP phrase is present, the equations under 𝜖
must be satisfied. The annotations on the 𝜖 node provide a dis attribute whose
value contains the relpro f-structure; this f-structure also plays a role inside the
clause defined by RelPath, and its value for the attribute pred is ‘pro’. With these
assumptions, the phrase man Chris saw has this f-structure, semantic structure,
and meaning constructor:
v[ ]
m ’ m
s r[ ]
c
As above, we combine the premises labeled [Chris] and [see], obtaining the
premise labeled [Chris-see]:
Our final task is to examine relative clauses with possessive relative pronouns. The
following are the f-structure, semantic structures, and meaning constructor for
man whose pen Chris used:
w
p
m
u
’
c
mv[ ]
m
mr[ ]
λx.the (y, poss (x, y) ∧ pen (y), use (Chris , y)) ∧ man (x) : mv ⊸ mr
We assume this lexical entry for the possessive relative pronoun whose:
Besides a meaning for its restriction, a determiner like every requires a meaning
resource for its scope. When an appropriate scope meaning resource is found, it is
consumed and a new meaning resource for the scope is provided, incorporating
the semantics of the quantifier. Analogously, the possessive relative determiner
whose requires a meaning resource p𝜎 ⊸ u𝜎 , specifying its scope.26 When the
resource p𝜎 ⊸ u𝜎 is consumed, a meaning resource w𝜎 ⊸ u𝜎 for the relative clause
is provided.
In the derivation of the meaning of man whose pen Chris used, the meaning
constructors in (159) are relevant:
26 Note that we treat the scope of the relative determiner whose as syntactically fixed (p𝜎 ⊸ u𝜎 ),
rather than unspecified as for a quantificational determiner like every. For discussion of scope ambiguity
for quantifiers, see §8.8.1.1.
relative clauses and semantic composition 713
(161) [wh-pen-Chris-use]
λx.the (y, poss (x, y) ∧ pen (y), use (Chris , y)) : w𝜎 ⊸ u𝜎
(162) [wh-pen-Chris-see-rel]
λQ.λx.the (y, poss (x, y) ∧ pen (y), use (Chris , y)) ∧ Q(x) :
[mv ⊸ mr] ⊸ [mv ⊸ mr]
Combining this meaning with the meaning constructor [man], we have the desired
result for man whose pen Chris used:
Nonrestrictive relative clauses are also associated with “wide scope” effects. For
example, Arnold and Sadler note that (166a), with a restrictive relative clause,
entails that Kim has a belief only about linguists who use the IPA. In contrast,
(166b), with a nonrestrictive relative, entails that Kim has a belief about linguists
in general, and the claim that linguists use the IPA is not a part of Kim’s beliefs:
(166) a. Kim believes that linguists who use the IPA are clever. [restrictive]
b. Kim believes that linguists, who use the IPA, are clever. [nonrestrictive]
714 long-distance dependencies
Arnold and Sadler (2010, 2011) assume that restrictive and nonrestrictive relative
clauses are syntactically integrated in the same way: they are both CPs adjoined
at the single-bar level, forming a phrase with the head that they modify. The
difference between them is rooted in their semantics. We assume the following
f-structure and semantic structure for the phrase Kim, who Sam dislikes, which
includes an appositive relative clause (ARC):
v[ ]
k k
d r[ ]
’
s
Arnold and Sadler (2010, 2011) base their semantic analysis of ARCs on the
approach to the semantics of supplemental constructions developed by Potts
(2005). As well as truth-conditional or at-issue content, Potts proposes another
dimension comprising Gricean conventional implicatures, ci content. Thus, each
expression is associated with a pair of meanings and a pair of types: one of the usual
at-issue type, and the other of the additional ci type, distinguished notationally
by a superscript c on the type specification. Potts proposes that parenthetical
and supplemental/appositive constructions, including ARCs, contribute to the
ci dimension of meaning. He analyzes constructions like ARCs as involving a
function comma which moves semantic content from the at-issue dimension to
the ci-dimension: formally, it takes an argument of type $e, t% and returns a result
of type $e, t c %, where t c is a ci rather than an at-issue type. In the case of an
ARC, the argument whose type is changed to ci content by comma is the relative
clause.
Arnold and Sadler (2010, 2011) combine Potts’s approach with the analysis of
relative clause semantics discussed in §17.6.2; their analysis derives an at-issue
meaning for the relative clause, which is reclassified by the [comma] meaning
constructor as ci meaning if the relative clause is nonrestrictive. We can begin by
instantiating the metavariables in the [rel] meaning constructor (141) according
to the f-structure labels in (167), resulting in the premise [rel] given in (168):27
(168) [rel] λP.λQ.λx.P(x) ∧ Q(x) : [w𝜎 $e% ⊸ d𝜎 $t% ] ⊸ [[v$e% ⊸ r$t% ] ⊸ [v$e% ⊸ r$t% ]]
27 Arnold and Sadler (2010, 2011) provide a slightly different formulation of [rel] from that given
in (141), because they associate [rel] with the CP node, rather than C, . The result of instantiation is
identical, so we ignore this detail of their analysis here.
relative clauses and semantic composition 715
For clarity, we explicitly annotate semantic structures with their types, since the
distinction between at-issue and ci types is important. Note that all of the types in
the [rel] meaning constructor are at-issue types.
The meaning of the relative clause who Sam dislikes (whether used restrictively
or nonrestrictively) is derived from the following premises:
(169) Meaning constructors for the relative clause who Sam dislikes
[dislike] λx.λy.dislike (x, y) : s𝜎 $e% ⊸ [w𝜎 $e% ⊸ d𝜎 $t% ]
[Sam] Sam : s𝜎 $e%
[who] λS.λx.person (x) ∧ S(x) : [w𝜎 $e% ⊸ d𝜎 $t% ] ⊸ [w𝜎 $e% ⊸ d𝜎 $t% ]
[rel] λP.λQ.λx.P(x) ∧ Q(x) :
[w𝜎 $e% ⊸ d𝜎 $t% ] ⊸ [[v$e% ⊸ r$t% ] ⊸ [v$e% ⊸ r$t% ]]
Combining the premises for the relative clause in the usual way, we have the
meaning constructor in (171) for the relative clause:
This meaning constructor contains some elements we have not previously encoun-
tered. On the glue side, the meaning constructor requires a resource of the type
[v ⊸ r] ⊸ [v ⊸ r], which corresponds to the meaning of the relative clause who
Sam dislikes. Corresponding to this on the meaning side, the meaning P of the rela-
tive clause is applied to the vacuous $e → t%-type meaning λz.true. This satisfies the
variable Q in the meaning of the relative clause in (171). The result of combining
the meaning constructors [who-Sam-dislike] and [comma] is as follows:
(174) Kim × [person (Kim ) ∧ dislikes (Sam , Kim ) ∧ true ] : k𝜎 $e% ⊗ k𝜎 $tc %
This meaning constructor can be used in a larger phrase or clause in exactly the
same way as the meaning constructor [Kim]. How the ci meaning content interacts
with the at-issue content in larger phrases and clauses is discussed by Arnold and
Sadler (2011). Giorgolo and Asudeh (2011a, 2012a) present an alternative analysis
of supplemental, non-at-issue, content which appeals to monads, a construction
from category theory.
(176) F-structure and s-structure for Davidi thought that Chris had seen himi
according to Asudeh (2012), with the antecedent relation explicitly
represented at s-structure:
’
d
’
p d []
Following Dalrymple et al. (1997), Asudeh (2012) proposes that the meaning
constructor for a pronoun consumes the meaning of the pronoun’s antecedent and
produces it again, in the process assigning the same meaning to the pronoun. This
is accomplished by the following meaning constructor:
On this view, the instantiated meaning constructors for the pronoun him and its
antecedent David in (176) are:
The meaning constructor for the pronoun consumes the meaning of its antecedent
and produces a pair of resources, one for the pronoun and one for the antecedent,
with the same meaning assigned to both resources. Combining [David] and [him],
we have the meaning constructor in (179), where the pronoun p𝜎 and its antecedent
d𝜎 are both associated with the meaning David :
’
g
m p a []
t ’
l p
In (181), the resumptive pronoun is labeled p, and its antecedent is the dis f-
structure labeled a.
We now provide a more explicit formulation of the manager resource which
consumes the surplus pronominal meaning constructor. Notice that the position
of the resumptive pronoun is not syntactically fixed; Asudeh uses a local name %rp
relative clauses and semantic composition 719
Given this definition of %rp, the manager resource consuming the pronoun mean-
ing constructor can be formulated as in (183) (similar in form to the schematic
resource in (180)):
The lexical entry for the complementizer aN discussed in §17.3 is revised to include
these meaning constructors.
The full set of meaning constructors that are relevant in the analysis of (181)
is given in (187). We assume that the semantic structure corresponding to m has
a var attribute which we call v, and a restr attribute which we call r. Asudeh
(2012) represents the denotation of plural fir ‘men’ as λx.man⃰ (x), where the
720 long-distance dependencies
We begin by combining the premises labeled [loyalto] and [theKing] to obtain the
meaning constructor labeled [loyalto-theKing]:
Next, we combine the premises labeled [AG] and [think] to obtain the meaning
constructor labeled [AG-think]:
(190) [AG-think-loyalto-theKing]
λx.think (atty-gen , loyal-to (x, ιk.king (k)))) : p ⊸ t
This meaning is of the form required by the [relabel] meaning constructor, and so
we can combine [AG-think-loyalto-theKing] and [relabel], producing the mean-
ing constructor [relabel-AG-think-loyalto-theKing]. The resulting meaning is
unchanged, but the dependency on the glue side is now on a𝜎 rather than p𝜎 :
28 The iota operator ι is something like the definite determiner the: ιk.king (k) is the individual k
satisfying the description king (k). For discussion, see Gamut (1991a: Chapter 5).
29 This step involves assumption of a hypothetical resource p to produce a resource l from
[loyalto-theKing], combining the result with [AG-think], and subsequently discharging the assump-
tion p. We do not represent these steps explicitly; for discussion of assumption and discharge of
hypothetical resources, see §8.8.1.4.
constituent questions and semantic composition 721
(191) [relabel-AG-think-loyalto-theKing]
λx.think (atty-gen , loyal-to (x, ιk.king (k)))) : a𝜎 ⊸ t𝜎
We now use the manager resource [mr] to consume the meaning of the pronoun
[they]. The result of combining [mr] and [they] is the meaning constructor labeled
[mr-they], in which the surplus pronoun meaning has been consumed:
(193) [mr-they-relabel-AG-think-loyalto-theKing]
λx.think (atty-gen , loyal-to (x, ιk.king (k))) : a𝜎 ⊸ t𝜎
This meaning constructor is of the form required by [rel]. Combining the meaning
constructor in (193) with [rel], we obtain the meaning constructor in (194):
(194) [rel-mr-they-relabel-AG-think-loyalto-theKing]
λP.P(x) ∧ think (atty-gen , loyal-to (x, ιk.king (k))) : (v ⊸ r) ⊸ (v ⊸ r)
Belyaev and Haug (2014) provide an analysis of the syntax and semantics of
correlatives in Ossetic (discussed earlier in §17.1.2) which also deals with the issue
of resumption, though by its nature this differs from the kind of resumption with
which Asudeh (2012) is concerned. In a correlative, a subordinate clause contains a
relative pronoun or phrase that is in some sense resumed by a full nominal phrase
in the matrix clause. Belyaev and Haug (2014) analyze as an instance of pronominal
anaphora the relationship between the relative pronoun/phrase and its correlate
which appears in the matrix clause. Their LFG analysis of the semantics of Ossetic
correlatives is framed in terms of glue semantics as well as Partial Compositional
Discourse Representation Theory (see Chapter 14).
This simple treatment is compatible with the Hamblin postulates, since it identifies
the meaning of the question with the set of propositions that constitute its complete
answer.
Next, we examine the issue of semantic composition and identification of the
meaning constructors that are involved in the derivation of a question meaning.
The sentence Who does David like? has the following f-structure:
constituent questions and semantic composition 723
f
’
A major difference between the meaning of a declarative sentence like David likes
Chris and the meaning of the question Who does David like? is the type of the
expressions representing their meaning. The expression in (197) is of type $t → t%:
it represents a set of possible answer meanings such as like (David , Chris ), each of
which has the type t of a declarative sentence. Up to now, we have associated only
the basic types e and t with semantic structures, not more complex types like
$t → t%. We can continue to associate only basic types with semantic structures
if the meaning of a question is associated not with a single semantic structure
but with an implicational meaning constructor, one whose right-hand side has
the form A ⊸ B. However, we have argued in Chapter 8 that a semantically
complete and coherent glue derivation results in a meaning constructor that is not
implicational, so that some refinement to the definition of semantic completeness
and coherence would need to be provided.
An alternative approach is to permit higher types such as $t → t% to be associ-
ated with semantic structures. For example, the meaning representation in (197)
is of type $t → t%, and it would be associated with the semantic structure f𝜎 in a
full analysis of the question in (199). In that case, we must provide a theory that
accounts for the discrepancy between the question type $t → t% and the basic type
t that is associated with a declarative sentence whose head is the verb likes. The
possibility of associating higher types with semantic structures is argued for by
Lowe (2014b), for reasons independent of the semantics of questions.
Finally, a desirable characteristic of any analysis of constituent questions is that
it should extend unproblematically to multiple constituent questions: questions that
contain more than one interrogative pronoun. The meaning constructors that are
relevant in the analysis of the question Who does David like? should be reusable
straightforwardly and without augmentation to produce the following meaning
representation:
The most complete discussion and analysis of the semantics of constituent ques-
tions in an LFG setting is by Mycock (2006). Mycock adopts Ginzburg and Sag’s
(2000) propositional abstract approach to the semantics of questions, according
724 long-distance dependencies
This book has presented the theory of Lexical Functional Grammar. Part I pre-
sented the two syntactic structures of LFG, constituent structure and functional
structure, and showed how they are represented, described, and constrained.
Part II discussed the representation of semantic structure, information structure,
prosodic structure, and the morphological module within the modular architec-
ture of LFG, and showed how these different aspects of linguistic structure are
related to one another. Part III provided syntactic and semantic analyses of a variety
of linguistic phenomena. Much work remains to be done on the constructions we
examined, as well as on topics that have unfortunately gone unexamined in this
book. In this final chapter, we turn to a brief discussion of some LFG work that has
not been covered in earlier chapters.
The Oxford Reference Guide to Lexical Functional Grammar. First edition. Mary Dalrymple, John J. Lowe, and Louise
Mycock. © Mary Dalrymple, John J. Lowe, and Louise Mycock 2019. First published 2019 by Oxford University Press.
psychological reality: processing and acquisition 727
2015), Russian (Artoni and Magnani 2015), and Swedish (Glahn et al. 2001).
Pienemann (2005a) and Bettoni and Di Biase (2015a) are significant collections
of PT work. Pienemann and Keßler (2012) provides an overview of the theory;
Pienemann and Keßler (2011) is an introductory textbook on PT.
further back in time, and which itself developed out of yet another, chronologically
prior, synchronic grammar.1
The relations between different but diachronically related synchronic grammars
is the realm of historical or diachronic linguistics. Diachronic changes in grammar
are well studied from a descriptive perspective, but they raise certain questions
for formal grammatical theories, which are primarily oriented towards synchronic
analysis. The most basic question is the extent to which it is even possible to
provide a descriptively adequate account of diachronic syntactic developments
in a synchronically oriented formalism like LFG. Although the majority of work
undertaken in LFG is exclusively synchronic in its aims, a number of authors have
explored diachronic linguistic developments from an LFG perspective, and have
shown that LFG is indeed adequate for representing diachronic developments.
A set of important early papers are collected in Butt and King (2001b). The loss
of case marking and its effect on complementation and grammatical relations
in Germanic and Romance languages has been explored by Allen (1995, 2008)
and Vincent and Börjars (2010a); the development of verbal periphrasis, and
the grammaticalization of lexical verbs into auxiliaries, is discussed by Allen
(2001), Schwarze (2001b), Butt and Lahiri (2013), Lowe (2015c), and Börjars
and Vincent (2017); the shift from accusative to split-ergative system in Indo-
Aryan is discussed by Butt (2001a); Nikitina (2008) discusses the development of
deverbal nominalizations into “mixed,” i.e. partially verbal, nominalizations, and
into infinitives; Börjars and Vincent (2017) discuss the development of pronouns
into copulas, an instance of “lateral grammaticalization.”
A more significant question is whether or to what extent LFG can provide an
explanatory model of syntactic change. For example, in grammatical theories that
rely on the concept of “movement” of words and constituents, the (supposed)
synchronic process of movement from lexical to functional head is often invoked
as an explanation for the well-attested diachronic development of lexical items into
grammatical items. As discussed by Börjars (2013) and Börjars and Vincent (2017),
there is no equivalent process within the LFG framework which can be so obviously
invoked to explain grammatical change. However, some works have advocated the
value of LFG over other grammatical theories in advancing our understanding of
diachronic grammatical change, in particular Vincent (1999, 2001b), Vincent and
Börjars (2010b), and Börjars and Vincent (2017). For example, Vincent (1999),
Börjars et al. (2016), and Börjars and Vincent (2017) show that there is diachronic
evidence for the gradual evolution of c-structure categories: a category such as P or
D may develop in a language, but there may not be evidence for PP or DP; phrasal
structure such as complement positions and specifier positions may develop at
a later period. Such developments can be modeled in LFG, where c-structural
representation is relatively free and nonprojecting categories are admitted, but it
challenges theories with more fixed approaches to phrase structure, where if a
category P exists in a language then a full PP, including specifier and complement
1 For a detailed discussion of the nature of diachronic relations between languages at different stages,
as well as the relations between the grammars of those languages, see Hale (2007).
730 related research threads and new directions
positions, necessarily also exists. In a similar vein, Lowe (2015b) argues that the
theory of Lexical Sharing in LFG provides a means of modeling the gradual process
of development between morpheme and clitic, or vice versa, which is otherwise
problematic for theories of syntax that assume an absolute distinction between
word and morpheme.
A number of authors have shown that the loss of a pred feature is a key
characteristic of grammaticalization processes. Vincent (1999) shows that once a
pred feature has been lost, constraints on the synchronic system explain associated
diachronic developments. That is, synchronic constraints on linguistic systems
modeled in LFG by reference to the pred feature provide an explanatory model of
certain diachronic changes. The loss of pred features in grammaticalization is also
discussed by Bresnan and Mchombo (1987) and Börjars et al. (1997). Coppock
and Wechsler (2010) show that the loss of person and number features is also
implicated in such processes.
On some level, of course, the idea of a fixed and invariant synchronic grammar
is a convenient fiction. To quote from Ohala’s (1989) work on sound change,
grammatical change is (at least in part) “drawn from a pool of synchronic vari-
ation.” Some work in LFG has sought to model synchronic variation, which may
serve as the basis of diachronic change. Bresnan et al. (2007b) make use of OT-
LFG to analyze individual speaker variation and dialectal variation with respect
to auxiliary-negative contractions in English. Vincent (1999, 2001a) and Vincent
and Börjars (2010b) also investigate the value of combining OT with LFG to model
synchronic variation as well as variation over time.
18.3.1 Parsing
Much important work has been done on the theory of parsing with LFG grammars
as well as on efficient parsing algorithms. A summary of work up to the mid-1990s
is given by Dalrymple et al. (1995c). Significant breakthroughs have been made on
several fronts.
computational issues 731
2 An exponential increase can be thought of in terms of some constant number raised to the nth
power, where n is the number of constraints. To give a rough idea of the rate of increase in complexity
as the number of constraints grows, if solving one constraint takes 21 = 2 seconds, solving three
constraints could take 23 = 8 seconds, and solving fifteen constraints could take 215 = 32,768 seconds
or about 9 hours.
3 The formal properties of context-free languages and their grammars, context-free grammars,
are described in Partee et al. (1993: Chapter 16). The added expressivity gained by allowing regular
expressions on the right-hand side of phrase structure rules, including the Kleene star, does not take us
beyond the power of context-free languages, as demonstrated by Woods (1970).
4 Parsing in cubic time is a particular instance of a problem that can be solved in polynomial time. If
a problem can be solved in polynomial time, the time it takes to solve a problem of size n is n raised to
some constant power; in the case of cubic time parsing, that constant is 3. Thus, if parsing complexity
depends on the length of the sentence we are parsing and a sentence of one word can be parsed in 13 = 1
second, a sentence consisting of three words would be parsed in 33 = 27 seconds, and a sentence with
fifteen words would be parsed in 153 = 3,375 seconds, or about 56 minutes. Of course, actual parsing
times are much faster than this for both short and long sentences: these figures are merely intended to
illustrate the rate of growth of a polynomial problem as opposed to an exponential problem. Parsing
in cubic time is much quicker and more efficient than the exponential growth associated with solving
arbitrary sets of functional constraints.
732 related research threads and new directions
18.3.2 Generation
Work on generation in LFG generally assumes that the generation task is to
determine the strings of a language that correspond to a specified f-structure, given
a particular grammar. Based on these assumptions, several interesting theoretical
results have been attained; Wedekind (2006) provides a useful overview, and
Butt et al. (2006: Part I) includes several papers describing LFG-based work on
generation and translation.
Wedekind (1995, 1999, 2014) addresses the issue of the decidability of genera-
tion from f-structures: the problem of determining whether there is any sentence
that corresponds to a given f-structure according to a given grammar. Wedekind
(1995) demonstrates that the problem is decidable for fully specified, acyclic5
f-structures: if we assume that the f-structure we are given is complete, has
no cycles, and no additional features can be added, we can always determine
whether or not there is a sentence that corresponds to that f-structure. Wedekind
5 An acyclic f-structure is one in which no f-structure contains a path leading back to itself. Kaplan
and Bresnan (1982) suggest that acyclic structures are the only f-structures that are motivated for
linguistic analysis.
computational issues 733
(1999) shows that the corresponding problem for underspecified f-structures is not
decidable: it is not always possible to determine whether there is a sentence that
corresponds to a given f-structure if we are allowed to add additional attributes and
values to the f-structure. Wedekind (2014) demonstrates that the corresponding
problem for cyclic f-structures is not decidable, even when the f-structure is fully
specified: it is not always possible to determine whether there is a sentence that
corresponds to a given f-structure if it contains one or more cycles.
In further work on the formal properties of generation from f-structures, Kaplan
and Wedekind (2000) show that if we are given an LFG grammar and an acyclic
f-structure, the set of strings that corresponds to that f-structure according to
the grammar is a context-free language. Building on and extending that work,
Wedekind and Kaplan (2012) provide a method for constructing the context-free
grammar for that set of strings by a process of specialization of the full grammar
that we are given. This result leads to a new way of thinking about generation,
opens the way to new and more efficient generation algorithms, and clarifies a
number of formal and mathematical issues relating to LFG parsing and generation.
Wedekind and Kaplan (1996) explore issues in ambiguity-preserving generation,
where a set of f-structures rather than a single f-structure is considered, and the
sentences of interest are those that correspond to all of the f-structures under
consideration; Shemtov (1997) also explores issues in ambiguity management
and ambiguity preservation in generation from sets of f-structures. The potential
practical advantages of ambiguity-preserving generation are clear: consider, for
example, a scenario involving translation from English to German. We first parse
the input English sentence, producing several f-structures if the English sentence is
ambiguous. For instance, the English sentence Hans saw the man with the telescope
is ambiguous: it means either that the man had the telescope or that Hans used the
telescope to see the man. The best translation for this sentence would be a German
sentence that is ambiguous in exactly the same way as the English sentence, if such
a German sentence exists; in the case at hand, we would like to produce the German
sentence Hans sah den Mann mit dem Fernrohr, which has exactly the same two
meanings as the English input. To do this, we map the English f-structures for the
input sentence to the set of corresponding German f-structures; our goal is then
to generate the German sentence Hans sah den Mann mit dem Fernrohr, which
expresses each of these f-structures. Though this approach is appealing, Wedekind
and Kaplan (1996) show that determining whether there is a single sentence that
corresponds to each member of a set of f-structures is in general undecidable for
an arbitrary (possibly linguistically unreasonable) LFG grammar: there are LFG
grammars and sets of f-structures for which it is impossible to determine whether
there is any sentence that corresponds to those f-structures. This result is important
in understanding the limits of ambiguity-preserving generation.
LFG grammars. One of the earliest LFG implementations was the Grammar
Writer’s Workbench (Kaplan and Maxwell 1996), originally implemented in the
early 1980s in the Xerox Interlisp-D environment, and under development through
the mid-1990s at the Xerox Palo Alto Research Center (PARC). In 1993, the PARC
team undertook a new implementation in C; this became the Xerox Linguistic
Environment (XLE: Crouch et al. 2008). The XLE underpins a number of large-
scale grammar implementations within the PARGRAM project (§18.3.4). More
recently, Lionel Clément and his colleagues have developed xLFG, a web-based
LFG parsing platform (Clément and Kinyon 2001; Clément 2016), and Damir
Cavar is leading a team to produce a freely available grammar development plat-
form for LFG and related grammar formalisms, the Free Linguistic Environment
(https://gorilla.linguistlist.org/fle).
The availability of computational implementations of LFG, and in particular the
XLE, has enabled research on combining LFG grammars with other computational
tools to increase parsing efficiency, improve parsing results, or produce more useful
language-based applications. Kaplan and King (2003) experiment with tools which
augment the input to the parsing process with various kinds of additional informa-
tion: using a finite-state parser to bracket the input string to indicate the presence of
phrasal boundaries, using a tagger to add part-of-speech annotations to the words
in the input, and using a gazetteer or onomasticon to mark word sequences as the
name of a person, place, or company. For example, the unannotated string in (1a)
might be annotated as in (1b) for input to XLE processing:
translation system, but has also been used to rewrite f-structures into semantic
structures, as described by Crouch and King (2006). Theorem provers for glue
semantics have been developed by Giorgolo (2017), including a standalone prover
as well as the “glue-xle” system, which can be used together with XLE.
Andrews, Avery D., III (2018). ‘Sets, Heads and Spreading in LFG.’ Journal of Lan-
guage Modelling 6(1). URL http://jlm.ipipan.waw.pl/index.php/JLM/article/
view/175.
Andrews, Avery D., III and Christopher D. Manning (1993). ‘Information Spread-
ing and Levels of Representation in LFG.’ Technical Report CSLI-93-176, CSLI
Publications, Stanford.
Andrews, Avery D., III and Christopher D. Manning (1999). Complex Predicates
and Information Spreading in LFG. Stanford: CSLI Publications.
Aranovich, Raúl (2013). ‘Mismatched Spanish Unaccusativity Tests.’ In King and
de Paiva (2013), 33–45.
Arka, I Wayan (1998). From Morphosyntax to Pragmatics in Balinese. Ph.D. thesis,
University of Sydney.
Arka, I Wayan (2003). Balinese Morphosyntax: A Lexical-Functional Approach.
Canberra: Pacific Linguistics.
Arka, I Wayan (2005). ‘The Core-Oblique Distinction and Core Index in
Some Austronesian Languages of Indonesia.’ URL http://www.academia.edu/
6659164/Arka_I_Wayan._2005._The_core-oblique_distinction_and_core_
index_in_some_Austronesian_languages_of_Indonesia._Keynote_paper_
presented_at_International_ALT_VI_Association_of_Linguistic_Typology_
conference_Padang_Indonesia_July_2005. Presented at International
Association of Linguistic Typology conference, Padang, Indonesia, July 2005.
Arka, I Wayan (2011). ‘Constructive Number Systems in Marori and Beyond.’ In
Butt and King (2011). URL http://csli-publications.stanford.edu/LFG/16/lfg11.
html.
Arka, I Wayan (2012). ‘Verbal Number, Argument Number and Plural Events in
Marori.’ In Butt and King (2012). URL http://csli-publications.stanford.edu/
LFG/17/lfg12.html.
Arka, I Wayan (2013). ‘Nominal Aspect in Marori.’ In Butt and King (2013). URL
http://csli-publications.stanford.edu/LFG/18/lfg13.html.
Arka, I Wayan (2014). ‘Double and Backward Control in Indonesian: An LFG
Analysis.’ In Butt and King (2014). URL http://csli-publications.stanford.edu/
LFG/19/lfg14.html.
Arka, I Wayan, Mary Dalrymple, Meladel Mistica, Suriel Mofu, Avery D. Andrews,
III, and Jane Simpson (2009). ‘A Computational Morphosyntactic Analysis for
the Applicative -I in Indonesian.’ In Butt and King (2009). URL http://csli-
publications.stanford.edu/LFG/14/lfg09.html.
Arka, I Wayan and Jane Simpson (1998). ‘Control and Complex Arguments in
Balinese.’ In Butt and King (1998b). URL http://csli-publications.stanford.edu/
LFG/LFG3-1998/.
Arka, I Wayan and Jane Simpson (2008). ‘Objective Voice and Control Into Subject
Clauses in Balinese.’ In Austin and Musgrave (2008), 83–114.
Arka, I Wayan and Stephen Wechsler (1996). ‘Argument Structure and Lin-
ear Order in Balinese Binding.’ In Butt and King (1996a). URL http://csli-
publications.stanford.edu/LFG/LFG1-1996/.
Arnold, Doug, Miriam Butt, Berthold Crysmann, Tracy Holloway King, and Stefan
Müller (eds.) (2017). Proceedings of the Joint 2016 Conference on Head-driven
references 741
Asudeh, Ash (2005). ‘Control and Semantic Resource Sensitivity.’ Journal of Lin-
guistics 41(3): 465–511.
Asudeh, Ash (2006). ‘Direct Compositionality and the Architecture of LFG.’ In Butt
et al. (2006), 363–387.
Asudeh, Ash (2007). ‘Some Notes on Pseudo Noun Incorporation in Niuean.’ URL
http://users.ox.ac.uk/~cpgl0036/pdf/niuean.pdf. Unpublished manuscript,
Carleton University.
Asudeh, Ash (2009). ‘Adjacency and Locality: A Constraint-Based Analysis of
Complementizer-Adjacent Extraction.’ In Butt and King (2009). URL http://
csli-publications.stanford.edu/LFG/14/lfg09.html.
Asudeh, Ash (2011). ‘Towards a Unified Theory of Resumption.’ In Alain Rouveret
(ed.), Resumptive Pronouns at the Interfaces, Amsterdam: John Benjamins,
121–188.
Asudeh, Ash (2012). The Logic of Pronominal Resumption. Oxford: Oxford Univer-
sity Press.
Asudeh, Ash and Douglas Ball (2005). ‘Niuean Incorporated Nominals as
Non-Projecting Nouns.’ URL http://users.ox.ac.uk/~cpgl0036/handouts/lsa05-
niuean-ho.pdf. Presented at the 2005 Annual Meeting of the Linguistic Society
of America, Oakland, CA.
Asudeh, Ash and Richard Crouch (2002a). ‘Coordination and Parallelism in Glue
Semantics: Integrating Discourse Cohesion and the Element Constraint.’ In Butt
and King (2002). URL http://csli-publications.stanford.edu/LFG/7/lfg02.html.
Asudeh, Ash and Richard Crouch (2002b). ‘Derivational Parallelism and Ellipsis
Parallelism.’ In WCCFL 21: Proceedings of the 21st West Coast Conference on
Formal Linguistics. Medford, MA: Cascadilla Press.
Asudeh, Ash, Mary Dalrymple, and Ida Toivonen (2008). ‘Constructions With
Lexical Integrity: Templates as the Lexicon-Syntax Interface.’ In Butt and King
(2008). URL http://csli-publications.stanford.edu/LFG/13/lfg08.html.
Asudeh, Ash, Mary Dalrymple, and Ida Toivonen (2013). ‘Constructions With
Lexical Integrity.’ Journal of Language Modelling 1(1): 77–130. URL http://jlm.
ipipan.waw.pl/index.php/JLM/article/view/56.
Asudeh, Ash and Gianluca Giorgolo (2012). ‘Flexible Composition for Optional
and Derived Arguments.’ In Butt and King (2012), 64–84. URL http://csli-
publications.stanford.edu/LFG/17/lfg12.html.
Asudeh, Ash, Gianluca Giorgolo, and Ida Toivonen (2014). ‘Meaning and Valency.’
In Butt and King (2014). URL http://csli-publications.stanford.edu/LFG/19/
lfg14.html.
Asudeh, Ash and Ida Toivonen (2006). ‘Expletives and the Syntax and Semantics of
Copy Raising.’ In Butt and King (2006b). URL http://csli-publications.stanford.
edu/LFG/11/lfg06.html.
Asudeh, Ash and Ida Toivonen (2007). ‘Copy Raising and Its Consequences for
Perception Reports.’ In Zaenen et al. (2007), 49–67.
Asudeh, Ash and Ida Toivonen (2012). ‘Copy Raising and Perception.’ Natural
Language and Linguistic Theory 30(2): 321–380.
Asudeh, Ash and Ida Toivonen (2015). ‘Lexical-Functional Grammar.’ In Heine
and Narrog (2015).
references 743
Baltin, Mark (1982). ‘A Landing Site Theory of Movement Rules.’ Linguistic Inquiry
13(1): 1–38.
Banfield, Ann (1973). ‘Narrative Style and the Grammar of Direct and Indirect
Speech.’ Foundations of Language 10: 1–39.
Barlow, Michael and Charles A. Ferguson (eds.) (1988). Agreement in Natural
Language: Approaches, Theories, Descriptions. Stanford: CSLI Publications.
Barron, Julia (1997). ‘LFG and the History of Raising Verbs.’ In Butt and King
(1997). URL http://csli-publications.stanford.edu/LFG/LFG2-1997/.
Barron, Julia (1999). Perception, Volition, and Reduced Clausal Complementation.
Ph.D. thesis, University of Manchester.
Barron, Julia (2001). ‘Perception and Raising Verbs: Synchronic and Diachronic
Relationships.’ In Butt and King (2001b), 73–103.
Barwise, Jon (1979). ‘On Branching Quantifiers in English.’ Journal of Philosophical
Logic 8: 47–80.
Barwise, Jon and Robin Cooper (1981). ‘Generalized Quantifiers and Natural
Language.’ Linguistics and Philosophy 4(2): 159–219.
Barwise, Jon and John Perry (1983). Situations and Attitudes. Cambridge, MA: The
MIT Press.
Bary, Corien and Dag T. T. Haug (2011). ‘Temporal Anaphora Across and Inside
Sentences: The Function of Participles.’ Semantics and Pragmatics 4, article 8:
1–56.
Bäuerle, Rainer, Christoph Schwarze, and Arnim von Stechow (eds.) (1983).
Meaning, Use, and Interpretation of Language. Berlin: Walter de Gruyter.
Bayer, Samuel (1996). ‘The Coordination of Unlike Categories.’ Language 72(3):
579–616.
Beard, Robert (1988). ‘On the Separation of Derivation from Morphology: Toward
a Lexeme-Morpheme Based Morphology.’ Quaderni di Semantica 9: 3–59.
Beard, Robert (1995). Lexeme-Morpheme Base Morphology: A General Theory of
Inflection and Word Formation. Stony Brook: SUNY Press.
Beaver, David I. (1992). ‘The Kinematics of Presupposition.’ In Paul Dekker,
and Martin Stokhof (eds.), Proceedings of the Eighth Amsterdam Colloquium.
Amsterdam: Institute for Logic, Language, and Information, University of
Amsterdam, 17–36.
Beckman, Mary E., Julia Hirschberg, and Stefanie Shattuck-Hufnagel (2005). ‘The
Original ToBI System and the Evolution of the ToBI Framework.’ In Jun (2005),
9–54.
Beesley, Kenneth R. and Lauri Karttunen (2003). Finite-State Morphology.
Stanford: CSLI Publications.
Bell, Sara (1983). ‘Advancements and Ascensions in Cebuano.’ In Perlmutter
(1983), 143–218.
Belyaev, Oleg I. (2013). ‘Optimal Agreement at M-Structure: Person in Dargwa.’ In
Butt and King (2013), 90–110. URL http://csli-publications.stanford.edu/LFG/
18/lfg13.html.
Belyaev, Oleg I. (2014). ‘Osetinskij kak jazyk s dvuhpadežnoj sistemoj: gruppovaja
fleksija i drugie paradoksy padežnogo markirovanija [Ossetic as a Two-Case
Language: Suspended Affixation and Other Case Marking Paradoxes].’ Voprosy
Jazykoznanija 6: 74–108.
references 745
Bodomo, Adams B. (1996). ‘Complex Verbal Predicates: The Case of Serial Verbs
in Dagaare and Akan.’ In Butt and King (1996a). URL http://csli-publications.
stanford.edu/LFG/LFG1-1996/.
Bodomo, Adams B. (1997). Paths and Pathfinders: Exploring the Syntax and Seman-
tics of Complex Verbal Predicates in Dagaare and Other Languages. Ph.D. thesis,
Norwegian University of Science and Technology.
Bögel, Tina (2010). ‘Pashto (Endo-)Clitics in a Parallel Architecture.’ In Butt and
King (2010), 85–105. URL http://csli-publications.stanford.edu/LFG/15/lfg10.
html.
Bögel, Tina (2012). ‘The P-Diagram – A Syllable-Based Approach to P-Structure.’
In Butt and King (2012), 99–117. URL http://csli-publications.stanford.edu/
LFG/17/lfg12.html.
Bögel, Tina (2013). ‘A Prosodic Resolution of German Case Ambiguities.’ In Butt
and King (2013), 131–151. URL http://csli-publications.stanford.edu/LFG/18/
lfg13.html.
Bögel, Tina (2014). ‘Degema and the String Interface.’ In Butt and King (2014).
URL http://csli-publications.stanford.edu/LFG/19/lfg14.html.
Bögel, Tina (2015). The Syntax-Prosody Interface in Lexical Functional Grammar.
Ph.D. thesis, Universität Konstanz. URL http://ling.uni-konstanz.de/pages/
home/boegel/Dateien/Boegel_thesis2015.pdf.
Bögel, Tina and Miriam Butt (2013). ‘Possessive Clitics and Ezafe in Urdu.’ In Kersti
Börjars, David Denison, and Alan Scott (eds.), Morphosyntactic Categories and
the Expression of Possession, Amsterdam: John Benjamins, 291–322.
Bögel, Tina, Miriam Butt, Ronald M. Kaplan, Tracy Holloway King, and John T.
Maxwell, III (2009). ‘Prosodic Phonology in LFG: A New Proposal.’ In Butt and
King (2009). URL http://csli-publications.stanford.edu/LFG/14/lfg09.html.
Bögel, Tina, Miriam Butt, Ronald M. Kaplan, Tracy Holloway King, and John T.
Maxwell, III (2010). ‘Second Position and the Prosody-Syntax Interface.’ In
Butt and King (2010). URL http://csli-publications.stanford.edu/LFG/15/lfg10.
html.
Bögel, Tina, Miriam Butt, and Sebastian Sulger (2008). ‘Urdu Ezafe and the
Morphology-Syntax Interface.’ In Butt and King (2008). URL http://csli-
publications.stanford.edu/LFG/13/lfg08.html.
Bolinger, Dwight L. (1965). Forms of English. Cambridge, MA: Harvard University
Press.
Bonami, Olivier and Berthold Crysmann (2016). ‘The Role of Morphology in
Constraint-Based Lexicalist Grammars.’ In Andrew Hippisley, and Gregory T.
Stump (eds.), Cambridge Handbook of Morphology, Cambridge, UK: Cambridge
University Press.
Bond, Oliver (2016). ‘Negation Through Reduplication and Tone: Implications for
the Lexical Functional Grammar/Paradigm Function Morphology Interface.’
Journal of Linguistics 52(2): 277–310. doi:10.1017/S0022226715000134.
Borer, Hagit (2013). Structuring Sense, Volume III: Taking Form. Oxford: Oxford
University Press.
Börjars, Kersti (1998). ‘Clitics, Affixes, and Parallel Correspondences.’ In Butt and
King (1998b). URL http://csli-publications.stanford.edu/LFG/LFG3-1998/.
references 747
Butt, Miriam and Tracy Holloway King (eds.) (1996a). On-Line Proceedings of the
LFG96 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/LFG1-1996/.
Butt, Miriam and Tracy Holloway King (1996b). ‘Structural Topic and Focus
Without Movement.’ In Butt and King (1996a). URL http://csli-publications.
stanford.edu/LFG/LFG1-1996/.
Butt, Miriam and Tracy Holloway King (eds.) (1997). On-Line Proceedings of the
LFG97 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/LFG2-1997/.
Butt, Miriam and Tracy Holloway King (1998a). ‘Interfacing Phonology With
LFG.’ In Butt and King (1998b). URL http://csli-publications.stanford.edu/
LFG/LFG3-1998/.
Butt, Miriam and Tracy Holloway King (eds.) (1998b). On-Line Proceedings of the
LFG98 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/LFG3-1998/.
Butt, Miriam and Tracy Holloway King (eds.) (1999). On-Line Proceedings of the
LFG99 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/LFG4-1999/.
Butt, Miriam and Tracy Holloway King (eds.) (2000a). Argument Realization.
Stanford: CSLI Publications.
Butt, Miriam and Tracy Holloway King (2000b). ‘Null Elements in Discourse
Structure.’ In K. V. Subbarao (ed.), Papers From the NULLS Seminar. Delhi:
Motilal Banarsidass.
Butt, Miriam and Tracy Holloway King (eds.) (2000c). On-Line Proceedings of the
LFG2000 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/LFG5-2000/.
Butt, Miriam and Tracy Holloway King (eds.) (2001a). On-Line Proceedings of the
LFG2001 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/6/lfg01.html.
Butt, Miriam and Tracy Holloway King (eds.) (2001b). Time Over Matter:
Diachronic Perspectives on Morphosyntax. Stanford: CSLI Publications.
Butt, Miriam and Tracy Holloway King (eds.) (2002). On-Line Proceedings of the
LFG2002 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/7/lfg02.html.
Butt, Miriam and Tracy Holloway King (eds.) (2003a). Nominals: Inside and Out.
Stanford: CSLI Publications.
Butt, Miriam and Tracy Holloway King (eds.) (2003b). On-Line Proceedings of the
LFG2003 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/8/lfg03.html.
Butt, Miriam and Tracy Holloway King (eds.) (2004a). On-Line Proceedings of the
LFG2004 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/9/lfg04.html.
Butt, Miriam and Tracy Holloway King (eds.) (2004b). On-Line Proceedings of the
LFG2004 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/9/lfg04.html.
references 753
Butt, Miriam and Tracy Holloway King (2004c). ‘The Status of Case.’ In Veneeta
Dayal, and Anoop Mahajan (eds.), Clause Structure in South Asian Languages,
Dordrecht: Springer, 153–198.
Butt, Miriam and Tracy Holloway King (eds.) (2005). On-Line Proceedings of the
LFG2005 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/10/lfg05.html.
Butt, Miriam and Tracy Holloway King (eds.) (2006a). Lexical Semantics in LFG.
Stanford: CSLI Publications.
Butt, Miriam and Tracy Holloway King (eds.) (2006b). On-Line Proceedings of the
LFG2006 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/11/lfg06.html.
Butt, Miriam and Tracy Holloway King (2006c). ‘Restriction for Morphological
Valency Alternations: The Urdu Causative.’ In Butt et al. (2006), 235–358.
Butt, Miriam and Tracy Holloway King (eds.) (2007). On-Line Proceedings of the
LFG2007 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/12/lfg07.html.
Butt, Miriam and Tracy Holloway King (eds.) (2008). On-Line Proceedings of the
LFG2008 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/13/lfg08.html.
Butt, Miriam and Tracy Holloway King (eds.) (2009). On-Line Proceedings of the
LFG2009 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/14/lfg09.html.
Butt, Miriam and Tracy Holloway King (eds.) (2010). On-Line Proceedings of the
LFG2010 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/15/lfg10.html.
Butt, Miriam and Tracy Holloway King (eds.) (2011). On-Line Proceedings of the
LFG2011 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/16/lfg11.html.
Butt, Miriam and Tracy Holloway King (eds.) (2012). On-Line Proceedings of the
LFG2012 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/17/lfg12.html.
Butt, Miriam and Tracy Holloway King (eds.) (2013). On-Line Proceedings of the
LFG2013 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/18/lfg13.html.
Butt, Miriam and Tracy Holloway King (eds.) (2014). On-Line Proceedings of the
LFG2014 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/19/lfg14.html.
Butt, Miriam and Tracy Holloway King (2015a). ‘Lexical-Functional Grammar.’ In
Kiss and Alexiadou (2015), 839–874.
Butt, Miriam and Tracy Holloway King (eds.) (2015b). On-Line Proceedings of the
LFG2015 Conference. Stanford: CSLI Publications. URL http://csli-publications.
stanford.edu/LFG/20/lfg15.html.
Butt, Miriam and Tracy Holloway King (eds.) (2017). On-Line Proceedings of the
LFG2017 Conference. Stanford: CSLI Publications. URL http://web.stanford.
edu/group/cslipublications/cslipublications/LFG/LFG-2017/index.shtml.
754 references
Butt, Miriam, Tracy Holloway King, Hiroshi Masuichi, and Christian Rohrer
(2002b). ‘The Parallel Grammar Project.’ In J. Carroll, N.Oostijk, and R. Sutcliffe
(eds.), Proceedings of the Workshop on Grammar Engineering and Evaluation.
1–7. COLING02.
Butt, Miriam, Tracy Holloway King, and John T. Maxwell, III (2003). ‘Com-
plex Predicates via Restriction.’ In Butt and King (2003b). URL http://csli-
publications.stanford.edu/LFG/8/lfg03.html.
Butt, Miriam, Tracy Holloway King, María-Eugenia Niño, and Frédérique Segond
(1999). A Grammar Writer’s Cookbook. Stanford: CSLI Publications.
Butt, Miriam, Tracy Holloway King, and Gillian Ramchand (2010). ‘Complex
Predication: How Did the Child Pinch the Elephant?’ In Uyechi and Wee (2010),
231–256.
Butt, Miriam, Tracy Holloway King, and Sebastian Roth (2007). ‘Urdu Correl-
atives: Theoretical and Implementational Issues.’ In Butt and King (2007),
107–127. URL http://csli-publications.stanford.edu/LFG/12/lfg07.html.
Butt, Miriam and Aditi Lahiri (2013). ‘Diachronic Pertinacity of Light Verbs.’
Lingua 135: 7–29.
Butt, Miriam, María-Eugenia Niño, and Frédérique Segond (1996b). ‘Multilingual
Processing of Auxiliaries Within LFG.’ In Dafydd Gibbon (ed.), KONVENS
’96: Natural Language Processing and Speech Technology: Results of the 3rd
KONVENS Conference. Mouton de Gruyter, 111–122. Reprinted in Sadler and
Spencer (2004, 11–22).
Butt, Miriam and Gillian Ramchand (2005). ‘Complex Aspectual Structure in
Hindi/Urdu.’ In Nomi Erteschik-Shir, and Tova Rapoport (eds.), The Syntax of
Aspect, Oxford: Oxford University Press, 117–153.
Butt, Miriam and Alexandros Tantos (2004). ‘Verbal Semantics via Petri Nets.’ In
Butt and King (2004b). URL http://csli-publications.stanford.edu/LFG/9/lfg04.
html.
Camilleri, Maris and Louisa Sadler (2011). ‘Restrictive Relative Clauses in Maltese.’
In Butt and King (2011). URL http://csli-publications.stanford.edu/LFG/16/
lfg11.html.
Carnie, Andrew (2012a). ‘Lexical-Functional Grammar.’ In Carnie (2012b). URL
http://bcs.wiley.com/he-bcs/Books?action=resource&bcsId=7378&itemId=
0470655313&resourceId=28724.
Carnie, Andrew (2012b). Syntax: A Generative Introduction. Oxford: Blackwell,
third edition.
Cavar, Damir and Melanie Seiss (2011). ‘Clitic Placement, Syntactic Discontinuity
and Information Structure.’ In Butt and King (2011), 131–151. URL http://csli-
publications.stanford.edu/LFG/16/lfg11.html.
Çetinoğlu, Özlem and Miriam Butt (2008). ‘Turkish Non-Canonical Objects.’ In
Butt and King (2008). URL http://csli-publications.stanford.edu/LFG/13/lfg08.
html.
Çetinoğlu, Özlem, Miriam Butt, and Kemal Oflazer (2009). ‘Mono/Bi-Clausality
of Turkish Causative.’ In Essays on Turkish Linguistics: Proceedings of the 14th
International Conference on Turkish Linguistics, August 6–8, 2008. Wiesbaden:
Harrassowitz, 43–52.
references 755
Champollion, Lucas (2016). ‘Ten Men and Women Got Married Today: Noun
Coordination and the Intersective Theory of Conjunction.’ Journal of Semantics
33(3): 561–622.
Chen, Matthew Y. (1987). ‘The Syntax of Xiamen Tone Sandhi.’ Phonology Year-
book 4: 109–149.
Chierchia, Gennaro (1984a). ‘Anaphoric Properties of Infinitives and Gerunds.’
In Mark Cobler, Susannah MacKaye, and Michael Wescoat (eds.), WCCFL 3:
Proceedings of the Third West Coast Conference on Formal Linguistics. Stanford:
CSLI Publications, 28–39.
Chierchia, Gennaro (1984b). Topics in the Semantics of Infinitives and Gerunds.
Ph.D. thesis, University of Massachusetts, Amherst.
Chierchia, Gennaro (1985). ‘Formal Semantics and the Grammar of Predication.’
Linguistic Inquiry 16(3): 417–443.
Chierchia, Gennaro and Pauline Jacobson (1985). ‘Local and Long Distance
Control.’ In Stephen Berman, Jae-Woong Choe, and Joyce McDonough (eds.),
Proceedings of NELS 16. Graduate Linguistic Student Association, University of
Massachusetts, Amherst.
Chisarik, Erika (2002). ‘Partitive Noun Phrases in Hungarian.’ In Butt and King
(2002). URL http://csli-publications.stanford.edu/LFG/7/lfg02.html.
Chisarik, Erika and John Payne (2003). ‘Modelling Possessor Constructions in
LFG: English and Hungarian.’ In Butt and King (2003a), 181–199.
Choi, Hye-Won (1999). Optimizing Structure in Context: Scrambling and Informa-
tion Structure. Dissertations in Linguistics. Stanford: CSLI Publications. Revised
version of 1996 Stanford University dissertation.
Choi, Hye-Won (2001). ‘Phrase Structure, Information Structure, and Resolution
of Mismatch.’ In Sells (2001a).
Chomsky, Noam (1955). The Logical Structure of Linguistic Theory. New York:
Plenum Press. Published in 1975.
Chomsky, Noam (1957). Syntactic Structures. Mouton de Gruyter.
Chomsky, Noam (1965). Aspects of the Theory of Syntax. Cambridge, MA: The MIT
Press.
Chomsky, Noam (1970). ‘Remarks on Nominalization.’ In Roderick A. Jacobs,
and Peter S. Rosenbaum (eds.), Readings in English Transformational Grammar,
Waltham, MA: Ginn, 184–221.
Chomsky, Noam (1971). ‘Deep Structure, Surface Structure, and Semantic Inter-
pretation.’ In Danny D. Steinberg, and Leon A. Jakobovits (eds.), Semantics: An
Interdisciplinary Reader In Philosophy, Linguistics and Psychology, Cambridge,
UK: Cambridge University Press, 183–216.
Chomsky, Noam (1981). Lectures on Government and Binding. Dordrecht: Foris.
Chomsky, Noam (1986). Barriers. Cambridge, MA: The MIT Press.
Chomsky, Noam (1993). ‘A Minimalist Program for Linguistic Theory.’ In Hale and
Keyser (1993), 1–52. Reprinted as chapter 3 of Chomsky (1995).
Chomsky, Noam (1995). The Minimalist Program. Cambridge, MA: The MIT
Press.
Chomsky, Noam (2001). ‘Beyond Explanatory Adequacy.’ Technical report,
Department of Linguistics and Philosophy, MIT, Cambridge, MA. MIT Occa-
sional Papers in Linguistics 20.
756 references
Chomsky, Noam and Morris Halle (1968). The Sound Pattern of English. New York:
Harper & Row.
Chung, Sandra (1982). ‘Unbounded Dependencies in Chamorro Grammar.’ Lin-
guistic Inquiry 13(1): 39–78.
Cinque, Guglielmo (1990). Types of Ā-Dependencies. Linguistic Inquiry Mono-
graphs. Cambridge, MA: The MIT Press.
Cinque, Guglielmo (1999). Adverbs and Functional Heads: A Cross-Linguistic
Perspective. Oxford: Oxford University Press.
Cinque, Guglielmo (2003). ‘The Interaction of Passive, Causative and ‘Restruc-
turing’ in Romance.’ In Christina Tortora (ed.), The Syntax of Italian Dialects,
Oxford: Oxford University Press, 50–66.
Cinque, Guglielmo (2004). ‘Restructuring and Functional Structure.’ In Adri-
ana Belletti (ed.), Structures and Beyond, Oxford: Oxford University Press,
132–191.
Clark, Brady Zack (2000). ‘Some Things are Not Susceptible to Thinking
About: The Historical Development of Tough Complementation.’ Unpublished
manuscript, Department of Linguistics, Stanford University.
Clément, Lionel (2016). ‘XLFG Documentation.’ URL https://hal.inria.fr/hal-
01277648/document.
Clément, Lionel and Alexandra Kinyon (2001). ‘XLFG – An LFG Parsing Scheme
for French.’ In Butt and King (2001a). URL http://csli-publications.stanford.
edu/LFG/6/lfg01.html.
Clements, George N. and Kevin C. Ford (1979). ‘Kikuyu Tone Shift and its
Synchronic Consequences.’ Linguistic Inquiry 10(2): 179–210.
Colban, Erik (1987). ‘Prepositional Phrases in Situation Schemata.’ In Fenstad et al.
(1987).
Comrie, Bernard (1974). ‘Causatives and Universal Grammar.’ Transactions of the
Philological Society 73(1): 1–32.
Comrie, Bernard (1985). Tense. Cambridge Textbooks in Linguistics. Cambridge,
UK: Cambridge University Press.
Comrie, Bernard (1987). ‘Russian.’ In Timothy Shopen (ed.), Languages and their
Status, Philadelphia: University of Pennsylvania, 91–151.
Cook, Phillipa and John Payne (2006). ‘Information Structure and Scope in
German.’ In Butt and King (2006b). URL http://csli-publications.stanford.edu/
LFG/11/lfg06.html.
Coppock, Elizabeth and Stephen Wechsler (2010). ‘Less-Travelled Paths From
Pronoun to Agreement: The Case of the Uralic Objective Conjugations.’ In
Butt and King (2010). URL http://csli-publications.stanford.edu/LFG/15/lfg10.
html.
Coppock, Elizabeth and Stephen Wechsler (2012). ‘The Objective Conjugation in
Hungarian: Agreement Without Phi-Features.’ Natural Language and Linguistic
Theory 30(3): 699–740.
Corbett, Greville G. (1983). ‘Resolution Rules: Agreement in Person, Number, and
Gender.’ In Gerald Gazdar, Ewan Klein, and Geoffrey K. Pullum (eds.), Order,
Concord and Constituency, Dordrecht: Foris, 175–206.
Corbett, Greville G. (2000). Number. Cambridge, UK: Cambridge University Press.
Corblin, Francis (1996). ‘Multiple Negation Processing in Natural Language.’
Theoria 62(3): 214–259.
references 757
Dahl, Östen (1969). Topic and Focus: A Study in Russian and General Transforma-
tional Grammar. Göteborg: Elandres Botryckeri.
Dahlstrom, Amy (1986a). ‘Nominal Arguments and Pronominal Inflection in Fox.’
Presented at the 1986 Annual Meeting of the Linguistic Society of America,
New York.
Dahlstrom, Amy (1986b). Plains Cree Morphosyntax. Ph.D. thesis, University of
California, Berkeley. Reprinted by Garland Press, New York, 1991.
Dahlstrom, Amy (1987). ‘Discontinuous Constituents in Fox.’ In Robert E. Moore,
and Paul D. Kroeber (eds.), Native American Languages and Grammatical
Typology, Bloomington: Indiana University Linguistics Club, 53–73.
Dahlstrom, Amy (2003). ‘Focus Constructions in Meskwaki (Fox).’ In Butt and
King (2003b). URL http://csli-publications.stanford.edu/LFG/8/lfg03.html.
Dahlstrom, Amy (2009). ‘OBJ𝜃 Without OBJ: A Typology of Meskwaki Objects.’ In
Butt and King (2009). URL http://csli-publications.stanford.edu/LFG/14/lfg09.
html.
Dahlstrom, Amy (2013). ‘Argument Structure of Quirky Algonquian Verbs.’ In
King and de Paiva (2013), 61–71.
Dalrymple, Mary (1993). The Syntax of Anaphoric Binding. Number 36 in CSLI
Lecture Notes. Stanford: CSLI Publications.
Dalrymple, Mary (ed.) (1999). Semantics and Syntax in Lexical Functional Gram-
mar: The Resource Logic Approach. Cambridge, MA: The MIT Press.
Dalrymple, Mary (2001). Lexical Functional Grammar, volume 34 of Syntax and
Semantics. New York: Academic Press.
Dalrymple, Mary (2006). ‘Lexical Functional Grammar.’ In Brown (2006), 82–94.
URL http://users.ox.ac.uk/~cpgl0015/lfg.pdf.
Dalrymple, Mary (2010). ‘Constraints on Operator Binding in Malayalam.’ In
Uyechi and Wee (2010), 257–270.
Dalrymple, Mary (2012). ‘Number Marking: An LFG Overview.’ In Butt and King
(2012). URL http://csli-publications.stanford.edu/LFG/17/lfg12.html.
Dalrymple, Mary (2015). ‘Obligatory Nonlocal Binding: An Exclusively Long
Distance Anaphor in Ya̧g Dii.’ Natural Language and Linguistic Theory 33(4):
1089–1120.
Dalrymple, Mary (2017). ‘Unlike Phrase Structure Category Coordination.’ In The
Very Model of a Modern Linguist – In Honor of Helge Dyvik, Bergen: Bergen
Language and Linguistics Studies (BeLLS), volume 8. URL http://dx.doi.org/
10.15845/bells.v8i1.
Dalrymple, Mary, Helge Dyvik, and Tracy Holloway King (2004a). ‘Copular
Complements: Closed or Open?’ In Butt and King (2004b). URL http://csli-
publications.stanford.edu/LFG/9/lfg04.html.
Dalrymple, Mary and Jamie Y. Findlay (2019). ‘Lexical Functional Grammar.’ In
András Kertész, Edith Moravcsik, and Csilla Rákosi (eds.), Current Approaches
to Syntax: A Comparative Handbook, De Gruyter Mouton.
Dalrymple, Mary, Vineet Gupta, John Lamping, and Vijay A. Saraswat (1999a).
‘Relating Resource-Based Semantics to Categorial Semantics.’ In Dalrymple
(1999).
references 759
Dalrymple, Mary, Dag T. T. Haug, and John J. Lowe (2018). ‘Integrating LFG’s
Binding Theory with PCDRT.’ Journal of Language Modelling 6(1). URL http://
jlm.ipipan.waw.pl/index.php/JLM/article/view/204.
Dalrymple, Mary, Angie Hinrichs, John Lamping, and Vijay A. Saraswat (1993).
‘The Resource Logic of Complex Predicate Interpretation.’ In Keh-Jiann Chen,
and Chu-Ren Huang (eds.), ROCLING VI: Proceedings of the Computational
Linguistics Conference. Hsitou National Park, Taiwan: Computational Linguis-
tics Society of Republic of China, 3–21. Also published as Xerox Technical
Report ISTL-NLTT-1993-08-03.
Dalrymple, Mary and Bozhil Hristov (2010). ‘Agreement Patterns and Coordina-
tion in Lexical Functional Grammar.’ In Butt and King (2010). URL http://csli-
publications.stanford.edu/LFG/15/lfg10.html.
Dalrymple, Mary, Makoto Kanazawa, Yookyung Kim, Sam A. Mchombo, and
Stanley Peters (1998). ‘Reciprocal Expressions and the Concept of Reciprocity.’
Linguistics and Philosophy 21(2): 159–210.
Dalrymple, Mary and Ronald M. Kaplan (2000). ‘Feature Indeterminacy and
Feature Resolution.’ Language 76(4): 759–798.
Dalrymple, Mary, Ronald M. Kaplan, and Tracy Holloway King (2001). ‘Weak
Crossover and the Absence of Traces.’ In Butt and King (2001a). URL http://
csli-publications.stanford.edu/LFG/6/lfg01.html.
Dalrymple, Mary, Ronald M. Kaplan, and Tracy Holloway King (2004b). ‘Linguis-
tic Generalizations Over Descriptions.’ In Butt and King (2004b). URL http://
csli-publications.stanford.edu/LFG/9/lfg04.html.
Dalrymple, Mary, Ronald M. Kaplan, and Tracy Holloway King (2007). ‘The
Absence of Traces: Evidence From Weak Crossover.’ In Zaenen et al. (2007).
Dalrymple, Mary, Ronald M. Kaplan, and Tracy Holloway King (2015). ‘Economy
of Expression as a Principle of Syntax.’ Journal of Language Modelling 2(3):
377–412. URL http://jlm.ipipan.waw.pl/index.php/JLM/article/view/82.
Dalrymple, Mary, Ronald M. Kaplan, John T. Maxwell, III, and Annie Zaenen
(1995a). ‘Formal Architecture.’ In Dalrymple et al. (1995b), 1–5.
Dalrymple, Mary, Ronald M. Kaplan, John T. Maxwell, III, and Annie Zaenen
(eds.) (1995b). Formal Issues in Lexical-Functional Grammar. Stanford: CSLI
Publications.
Dalrymple, Mary, Ronald M. Kaplan, John T. Maxwell, III, and Annie Zaenen
(1995c). ‘Mathematical and Computational Issues.’ In Dalrymple et al. (1995b),
331–338.
Dalrymple, Mary, Ronald M. Kaplan, John T. Maxwell, III, and Annie Zaenen
(1995d). ‘Nonlocal Dependencies.’ In Dalrymple et al. (1995b), 131–135.
Dalrymple, Mary and Tracy Holloway King (2000). ‘Missing-Object Construc-
tions: Lexical and Constructional Variation.’ In Butt and King (2000c). URL
http://csli-publications.stanford.edu/LFG/LFG5-2000/.
Dalrymple, Mary and Tracy Holloway King (2013). ‘Nested and Crossed Depen-
dencies and the Existence of Traces.’ In King and de Paiva (2013), 139–152.
Dalrymple, Mary, Tracy Holloway King, and Louisa Sadler (2009). ‘Indeterminacy
by Underspecification.’ Journal of Linguistics 45: 31–68.
760 references
Falk, Yehuda N. (2002b). ‘Resumptive Pronouns in LFG.’ In Butt and King (2002).
URL http://csli-publications.stanford.edu/LFG/7/lfg02.html.
Falk, Yehuda N. (2003). ‘The English Auxiliary System Revisited.’ In Butt
and King (2003b), 184–204. URL http://csli-publications.stanford.edu/LFG/8/
lfg03.html.
Falk, Yehuda N. (2004). ‘The Hebrew Present Tense Copula as a Mixed Category.’
In Butt and King (2004b). URL http://csli-publications.stanford.edu/LFG/9/
lfg04.html.
Falk, Yehuda N. (2005). ‘Open Argument Functions.’ In Butt and King (2005). URL
http://csli-publications.stanford.edu/LFG/10/lfg05.html.
Falk, Yehuda N. (2006). Subjects and Universal Grammar: An Explanatory Theory.
Cambridge, UK: Cambridge University Press.
Falk, Yehuda N. (2007). ‘Do We Wanna (Or Hafta) Have Empty Categories?’ In
Butt and King (2007). URL http://csli-publications.stanford.edu/LFG/12/lfg07.
html.
Falk, Yehuda N. (2008). ‘Functional Relations in the English Auxiliary System.’
Linguistics 46: 861–889.
Falk, Yehuda N. (2009). ‘Islands: A Mixed Analysis.’ In Butt and King (2009),
261–281. URL http://csli-publications.stanford.edu/LFG/14/lfg09.html.
Falk, Yehuda N. (2010). ‘An Unmediated Analysis of Relative Clauses.’ In Butt
and King (2010), 207–227. URL http://csli-publications.stanford.edu/LFG/15/
lfg10.html.
Falk, Yehuda N. (2011). ‘Multiple-Gap Constructions.’ In Butt and King (2011).
URL http://csli-publications.stanford.edu/LFG/16/lfg11.html.
Faltz, Leonard M. (1977). Reflexivization: A Study in Universal Syntax. Ph.D. thesis,
University of California, Berkeley. Reprinted by Garland Press, New York,
1985.
Fassi-Fehri, Abdulkader (1981). Complémentation et Anaphore en Arabe Moderne:
Une Approche Lexicale Fonctionelle. Ph.D. thesis, Université Paris 3.
Fassi-Fehri, Abdulkader (1988). ‘Agreement in Arabic, Binding, and Coherence.’
In Barlow and Ferguson (1988), 107–158.
Featherston, Sam and Wolfgang Sternefeld (eds.) (2007). Roots: Linguistics in
Search of its Evidential Base. Berlin: Mouton de Gruyter.
Fenstad, Jens Erik, Per-Kristian Halvorsen, Tore Langholm, and Johan van Ben-
them (eds.) (1987). Situations, Language and Logic. Dordrecht: D. Reidel.
Féry, Caroline and Manfred Krifka (2008). ‘Information Structure: Notional Dis-
tinctions, Ways of Expression.’ In Piet van Sterkenburg (ed.), Unity and Diversity
of Languages, Amsterdam: John Benjamins, 123–136.
Fillmore, Charles J. (1968). ‘The Case for Case.’ In Emmon Bach, and Robert T.
Harms (eds.), Universals in Linguistic Theory, New York: Holt, Rinehart &
Winston, 1–210.
Findlay, Jamie Y. (2014). The Prepositional Passive: a Lexical Functional Account.
M.Phil. dissertation, University of Oxford.
Findlay, Jamie Y. (2016). ‘Mapping Theory Without Argument Structure.’ Journal
of Language Modelling 4(2): 293–338. URL http://jlm.ipipan.waw.pl/index.php/
JLM/article/view/171/0.
764 references
Frank, Anette and Annie Zaenen (2002). ‘Tense in LFG: Syntax and Morphology.’
In Hans Kamp, and Uwe Reyle (eds.), How We Say WHEN It Happens: Con-
tributions to the Theory of Temporal Reference in Natural Language, Tübingen:
Niemeyer, 17–51. Reprinted in Sadler and Spencer (2004, 23–66).
Franks, Steven and Tracy Holloway King (2000). A Handbook of Slavic Clitics.
Oxford: Oxford University Press.
Frantz, Donald G. (1981). Grammatical Relations in Universal Grammar. Bloom-
ington: Indiana University Linguistics Club.
Frey, Werner and Uwe Reyle (1983). ‘A PROLOG Implementation of Lexical-
Functional Grammar.’ In Proceedings of the Eighth International Joint Conference
on Artificial Intelligence. Karlsruhe, 727–729.
Frota, Sónia (2012). ‘Prosodic Structure, Constituents, and Their Implementation.’
In Abigail C. Cohn, Cécile Fougeron, and Marie K. Huffman (eds.), The Oxford
Handbook of Laboratory Phonology, Oxford: Oxford University Press, 255–265.
Fry, John (1999a). ‘Proof Nets and Negative Polarity Licensing.’ In Dalrymple
(1999), 91–116.
Fry, John (1999b). ‘Resource-Logical Event Semantics for LFG.’ URL http://
citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.84.6503. Presented at
LFG99, University of Manchester.
Gabbay, Dov M. (1996). Labelled Deductive Systems. Oxford: Clarendon Press.
Gamut, L. T. F. (1991a). Logic, Language, and Meaning, volume 1: Introduction to
Logic. Chicago: University of Chicago Press.
Gamut, L. T. F. (1991b). Logic, Language, and Meaning, volume 2: Intensional Logic
and Logical Grammar. Chicago: University of Chicago Press.
Gawron, Jean Mark and Stanley Peters (1990). Anaphora and Quantification in
Situation Semantics. Number 19 in CSLI Lecture Notes. Stanford: CSLI Pub-
lications.
Gazdar, Gerald (1980). ‘A Cross-Categorial Semantics for Coordination.’ Linguis-
tics and Philosophy 3: 407–409.
Gazdar, Gerald, Ewan Klein, Geoffrey K. Pullum, and Ivan A. Sag (1985). Gener-
alized Phrase Structure Grammar. Cambridge, MA: Harvard University Press.
Gazdar, Gerald and Geoffrey K. Pullum (1981). ‘Subcategorization, Constituent
Order, and the Notion ‘Head’.’ In Michael Moortgat, Harry van der Hulst, and
Teun Hoekstra (eds.), The Scope of Lexical Rules, Dordrecht: Foris, 107–123.
Gazdik, Anna (2011). Multiple Questions in French and Hungarian. A Lexical-
Functional Analysis with Special Emphasis on the Syntax-Discourse Interface.
Ph.D. thesis, Université Paris Diderot.
Georgopoulos, Carol (1985). ‘Variables in Palauan Syntax.’ Natural Language and
Linguistic Theory 3(1): 59–94.
Ginzburg, Jonathan (1996). ‘Interrogatives: Questions, Facts, and Dialogue.’ In
Lappin (1996), 385–422.
Ginzburg, Jonathan (2001). Questions, Queries, and Facts: A Semantics and Prag-
matics for Interrogatives. Dissertations in Linguistics. Stanford: CSLI Publica-
tions. Reprinted version of 1992 Stanford dissertation.
Ginzburg, Jonathan (2011). ‘Questions: Logic and Interactions.’ In van Benthem
and ter Meulen (2011), 1133–1146.
766 references
Ginzburg, Jonathan and Ivan A. Sag (2000). Interrogative Investigations: The Form,
Meaning and Use of English Interrogatives. Stanford: CSLI Publications.
Giorgolo, Gianluca (2017). ‘Glue Repository.’ URL https://github.com/
gianlucagiorgolo. GitHub Repositories.
Giorgolo, Gianluca and Ash Asudeh (2011a). ‘Multidimensional Semantics With
Unidimensional Glue Logic.’ In Butt and King (2011), 236–256. URL http://csli-
publications.stanford.edu/LFG/16/lfg11.html.
Giorgolo, Gianluca and Ash Asudeh (2011b). ‘Multimodal Communication in
LFG: Gestures and the Correspondence Architecture.’ In Butt and King (2011),
257–277. URL http://csli-publications.stanford.edu/LFG/16/lfg11.html.
Giorgolo, Gianluca and Ash Asudeh (2012a). ‘!M, η, ⃰" Monads for Conven-
tional Implicatures.’ In Ana Aguilar-Guevara, Anna Chernilovskaya, and Rick
Nouwen (eds.), Sinn und Bedeutung 16. MIT Working Papers in Linguistics.
Giorgolo, Gianluca and Ash Asudeh (2012b). ‘Missing Resources in a Resource-
Sensitive Semantics.’ In Butt and King (2012), 219–239. URL http://csli-
publications.stanford.edu/LFG/17/lfg12.html.
Giorgolo, Gianluca and Ash Asudeh (2014). ‘Monads as a Solution for Generalized
Opacity.’ In Proceedings of the EACL 2014 Workshop on Type Theory and Natural
Language Semantics (TTNLS). 19–27. URL https://www.aclweb.org/anthology/
W/W14/W14-1400.pdf.
Girard, Jean-Yves (1987). ‘Linear Logic.’ Theoretical Computer Science 50(1):
1–102.
Givón, Talmy (1997). ‘Grammatical Relations: An Introduction.’ In Talmy Givón
(ed.), Grammatical Relations: A Functional Perspective, Amsterdam: John
Benjamins, 1–84.
Glahn, Esther, Gisela Håkansson, Björn Hammarberg, Anne Holmen, Anne
Hvenekilde, and Karen Lund (2001). ‘Processability in Scandinavian Second
Language Acquisition.’ Studies in Second Language Acquisition 23: 389–416.
Glasbey, Sheila (2001). ‘Tense, Aspect and the Temporal Structure of Discourse:
Towards an LFG Account.’ In Butt and King (2001a). URL http://csli-
publications.stanford.edu/LFG/6/lfg01.html.
Goodall, Grant (1987). Parallel Structures in Syntax. Cambridge, UK: Cambridge
University Press.
Grano, Thomas (2015). Control and Restructuring. Oxford: Oxford University
Press.
Greenberg, Joseph H. (1966a). Language Universals: With Special Reference to
Feature Hierarchies. The Hague: Mouton de Gruyter.
Greenberg, Joseph H. (1966b). ‘Some Universals of Grammar With Particular Ref-
erence to the Order of Meaningful Elements.’ In Joseph H. Greenberg (ed.), Uni-
versals of Language, Cambridge, MA: The MIT Press, 73–113. Second edition.
Grimshaw, Jane (1982a). ‘On the Lexical Representation of Romance Reflexive
Clitics.’ In Bresnan (1982b), 87–148.
Grimshaw, Jane (1982b). ‘Subcategorization and Grammatical Relations.’ In
Zaenen (1982), 35–56.
Grimshaw, Jane (1990). Argument Structure. Cambridge, MA: The MIT Press.
Groenendijk, Jeroen and Martin Stokhof (1984). Studies on the Semantics of Ques-
tions and the Pragmatics of Answers. Ph.D. thesis, University of Amsterdam.
references 767
Hayes, Bruce and Aditi Lahiri (1991). ‘Bengali Intonational Phonology.’ Natural
Language and Linguistic Theory 9: 47–96.
Hearne, Mary (2005). Data-Oriented Models of Parsing and Translation. Ph.D.
thesis, Dublin City University, School of Computing.
Hearne, Mary and Andy Way (2006). ‘Disambiguation Strategies for Data-
Oriented Translation.’ In Proceedings of the 11th Conference of the European
Association for Machine Translation. Oslo, 59–68.
Heim, Irene (1982). The Semantics of Definite and Indefinite Noun Phrases. Ph.D.
thesis, University of Massachusetts, Amherst.
Heine, Bernd and Heiko Narrog (eds.) (2015). The Oxford Handbook of Linguistic
Analysis. Oxford: Oxford University Press, second edition.
Hellan, Lars (1988). Anaphora in Norwegian and the Theory of Grammar.
Dordrecht: Foris.
Henadeerage, Kumara (1998). ‘Anaphoric Binding in Colloquial Sinhala.’ In
Butt and King (1998b). URL http://csli-publications.stanford.edu/LFG/LFG3-
1998/.
Heny, Frank (1973). ‘Sentence and Predicate Modifiers in English.’ In John P.
Kimball (ed.), Syntax and Semantics, New York: Academic Press, volume 4 of
Syntax and Semantics, 217–245.
Her, One-Soon (1998). ‘Lexical Mapping in Chinese Inversion Constructions.’ In
Butt and King (1998b). URL http://csli-publications.stanford.edu/LFG/LFG3-
1998/.
Her, One-Soon (1999). ‘Interaction of Thematic Structure and Syntactic Struc-
tures: On Mandarin Dative Alternations.’ In Chinese Languages and Linguistics
V: Interaction of Form and Function, Taipei: Academia Sinica, 373–412.
Her, One-Soon (2003). ‘Chinese Inversion Constructions Within a Simplified
LMT.’ Journal of Chinese Linguistics: Lexical-Functional Grammar Analysis of
Chinese 1–31. Edited by Adams Bodomo and Kang Kwong Luke.
Her, One-Soon (2009). ‘Apparent Subject-Object Inversion in Chinese.’ Linguistics
47(5): 1143–1181.
Her, One-Soon (2010). Interaction and Variation in the Chinese VO Construction.
Taipei: Crane Publishing. Revised edition.
Her, One-Soon (2013). ‘Lexical Mapping Theory Revisited.’ In King and de Paiva
(2013), 47–59.
Her, One-Soon and Dun-Hong Deng (2012). ‘Lexical Mapping in Yami Verbs.’
In Butt and King (2012). URL http://csli-publications.stanford.edu/LFG/17/
lfg12.html.
Heycock, Caroline and Roberto Zamparelli (2005). ‘Friends and Colleagues:
Plurality, Coordination, and the Structure of DP.’ Natural Language Semantics
13: 201–270.
Higginbotham, James (1983). ‘Logical Form, Binding, and Nominals.’ Linguistic
Inquiry 14(3): 395–420.
Higginbotham, James (1992). ‘Reference and Control.’ In Larson et al. (1992),
79–108.
Higginbotham, James (1996). ‘The Semantics of Questions.’ In Lappin (1996),
361–383.
Hinds, John (1986). Japanese. London: Croom Helm.
770 references
Inkelas, Sharon and Draga Zec (eds.) (1990). The Phonology-Syntax Connection.
Chicago & Stanford: University of Chicago Press and CSLI.
Ishikawa, Akira (1985). Complex Predicates and Lexical Operations in Japanese.
Ph.D. thesis, Stanford University.
Itô, Junko and Armin Mester (2003). ‘Weak Layering and Word Binarity.’ In
Takeru Honma, Masao Okazaki, Toshiyuki Tabata, and Shin-ichi Tanaka (eds.),
A New Century of Phonology and Phonological Theory. A Festschrift for Pro-
fessor Shosuke Haraguchi on the Occasion of His Sixtieth Birthday, Tokyo:
Kaitakusha, 26–65. Slightly revised version of unpublished working paper, Lin-
guistic Research Center, LRC-92-09, University of California, Santa Cruz, 1992.
Izvorski, Roumyana (1993). ‘On WH-Movement and Focus Movement in
Bulgarian.’ In Regine Eckhardt, and Veerle van Geenhoven (eds.), Proceedings
of ConSoLE II. Amsterdam: Sole Publications.
Jackendoff, Ray (1972). Semantic Interpretation in Generative Grammar.
Cambridge, MA: The MIT Press.
Jackendoff, Ray (1977). X̄ Syntax: A Study of Phrase Structure. Cambridge, MA:
The MIT Press.
Jackendoff, Ray (1983). Semantics and Cognition. Cambridge, MA: The MIT Press.
Jackendoff, Ray (1990). Semantic Structures. Cambridge, MA: The MIT Press.
Jackendoff, Ray (2002). Foundations of Language: Brain, Meaning, Grammar,
Evolution. Oxford: Oxford University Press.
Jackendoff, Ray (2010). Meaning and the Lexicon: The Parallel Architecture
1975–2010. Oxford: Oxford University Press.
Jacobs, Joachim (ed.) (1992). Informationsstruktur und Grammatik. Opladen:
Westdeutscher Verlag.
Jacobson, Pauline (1990). ‘Raising as Function Composition.’ Linguistics and
Philosophy 13: 423–475.
Jacobson, Pauline (1992a). ‘The Lexical Entailment Theory of Control and the
Tough Construction.’ In Sag and Szabolcsi (1992), 269–297.
Jacobson, Pauline (1992b). ‘Raising Without Movement.’ In Larson et al. (1992),
149–194.
Jacobson, Pauline and Geoffrey K. Pullum (eds.) (1982). The Nature of Syntactic
Representation. Dordrecht: D. Reidel.
Jaeger, T. Florian and Veronica A. Gerassimova (2002). ‘Bulgarian Word Order
and the Role of the Direct Object Clitic in LFG.’ In Butt and King (2002). URL
http://csli-publications.stanford.edu/LFG/7/lfg02.html.
Jaeggli, Osvaldo (1980). On Some Phonologically Null Elements in Syntax. Ph.D.
thesis, Massachusetts Institute of Technology.
Jakobson, Roman (1932). ‘Structure of the Russian Verb.’ In Linda R. Waugh, and
Morris Halle (eds.), Roman Jakobson: Russian and Slavic Grammar. Studies
1931–1981, Berlin: Mouton, 1–14. Published in 1984.
Jam, R. (1990). ‘A Conservative Extension to the Functional Uncertainty
Notation.’ Unpublished manuscript, Xerox PARC, with authors Ron Kaplan,
John Maxwell, Annie Zaenen, and Mary Dalrymple.
Jansen, Louise and Bruno Di Biase (2015). ‘Acquiring V2 in Declarative Sentences
and Constituent Questions in German as a Second Language.’ In Bettoni and
Di Biase (2015a), 259–273.
772 references
Kameyama, Megumi (1985). Zero Anaphora: The Case of Japanese. Ph.D. thesis,
Stanford University.
Kameyama, Megumi (1989). ‘Functional Precedence Conditions on Overt and
Zero Pronominals.’ Technical report, MCC.
Kamp, Hans (1975). ‘Two Theories About Adjectives.’ In Edward L. Keenan (ed.),
Formal Semantics of Natural Language, Cambridge, UK: Cambridge University
Press, 123–155.
Kamp, Hans (1981). ‘A Theory of Truth and Semantic Representation.’ In Jeroen
Groenendijk, Theo M. V. Janssen, and Martin Stokhof (eds.), Formal Methods
in the Study of Language, Amsterdam: Mathematische Centrum, volume I,
277–321. Reprinted in Portner and Partee (2002, 189–222).
Kamp, Hans and Barbara H. Partee (1995). ‘Prototype Theory and
Compositionality.’ Cognition 57(2): 129–191.
Kamp, Hans and Uwe Reyle (1993). From Discourse to Logic: An Introduction
to Modeltheoretic Semantics of Natural Language, Formal Logic and Discourse
Representation Theory. Dordrecht: Kluwer.
Kaplan, Ronald M. (1975a). ‘On Process Models for Sentence Analysis.’ In
Donald A. Norman, and David E. Rumelhart (eds.), Explorations in Cognition,
San Francisco: W. H. Freeman, 117–135.
Kaplan, Ronald M. (1975b). Transient Processing Load in Relative Clauses. Ph.D.
thesis, Harvard University.
Kaplan, Ronald M. (1976). ‘Models of Comprehension Based on Augmented
Transition Networks.’ Presented at the ATT-MIT Convocation on
Communications, March 1976.
Kaplan, Ronald M. (1987). ‘Three Seductions of Computational Psycholinguistics.’
In Peter Whitelock, Mary McGee Wood, Harold L. Somers, Rod Johnson, and
Paul Bennett (eds.), Linguistic Theory and Computer Applications, London:
Academic Press, 149–188. Also published as CCL/UMIST Report No. 86.2:
Alvey/ICL Workshop on Linguistic Theory and Computer Applications: Tran-
scripts of Presentations and Discussions. Center for Computational Linguistics,
University of Manchester. Reprinted in Dalrymple et al. (1995b, 337–367).
Kaplan, Ronald M. (1988). ‘Correspondences and Their Inverses.’ Presented
at the Titisee Workshop on Unification Formalisms: Syntax, Semantics and
Implementation, Titisee, Germany.
Kaplan, Ronald M. (1995). ‘The Formal Architecture of Lexical-Functional
Grammar.’ In Dalrymple et al. (1995b), 7–27. Previously published in
Proceedings of ROCLING II, ed. C.-R. Huang and K.-J. Chen, Tapei 1989,
pp. 1–18, and in Journal of Information Science and Engineering 5 (1989),
pp. 305–322.
Kaplan, Ronald M. (2017). ‘Formal Aspects of Underspecified Features.’ In Cleo
Condoravdi, and Tracy Holloway King (eds.), Tokens of Meaning: Papers in
Honor of Lauri Karttunen, Stanford: CSLI Publications.
Kaplan, Ronald M. and Joan Bresnan (1982). ‘Lexical-Functional Grammar: A
Formal System for Grammatical Representation.’ In Bresnan (1982b), 173–281.
Reprinted in Dalrymple et al. (1995b, 29–130).
Kaplan, Ronald M. and Miriam Butt (2002). ‘The Morphology-Syntax Interface
in LFG.’ URL http://cslipublications.stanford.edu/LFG/7/lfg02.html. Abstract
774 references
in On-Line Proceedings of the LFG2002 Conference, ed. Miriam Butt and Tracy
Holloway King.
Kaplan, Ronald M. and Martin Kay (1994). ‘Regular Models of Phonological Rule
Systems.’ Computational Linguistics 20(3): 331–378.
Kaplan, Ronald M. and Tracy Holloway King (2003). ‘Low-Level Markup and
Large-Scale LFG Grammar Processing.’ In Butt and King (2003b). URL http://
csli-publications.stanford.edu/LFG/8/lfg03.html.
Kaplan, Ronald M. and John T. Maxwell, III (1988a). ‘An Algorithm for Functional
Uncertainty.’ In Proceedings of the International Conference on Computational
Linguistics (COLING88). International Committee on Computational
Linguistics, 297–302. Reprinted in Dalrymple et al. (1995b, 177–197).
Kaplan, Ronald M. and John T. Maxwell, III (1988b). ‘Constituent Coordination in
Lexical-Functional Grammar.’ In Proceedings of the International Conference on
Computational Linguistics (COLING88). International Committee on Compu-
tational Linguistics, 303–305. Reprinted in Dalrymple et al. (1995b, 199–210).
Kaplan, Ronald M. and John T. Maxwell, III (1996). ‘LFG Grammar Writer’s
Workbench.’ Technical report, Xerox Palo Alto Research Center, Palo Alto, CA.
URL http://www2.parc.com/isl/groups/nltt/medley/.
Kaplan, Ronald M., John T. Maxwell, III, Tracy Holloway King, and Richard
Crouch (2004). ‘Integrating Finite-State Technology with Deep LFG
Grammars.’ In Proceedings of the ESSLLI Workshop ‘Combining Shallow
and Deep Processing for NLP’. ESSLLI, 168–173.
Kaplan, Ronald M., John T. Maxwell, III, and Annie Zaenen (1987). ‘Functional
Uncertainty.’ CSLI Monthly Newsletter Stanford University.
Kaplan, Ronald M. and Jürgen Wedekind (1993). ‘Restriction and
Correspondence-Based Translation.’ In Proceedings of the 6th Meeting of the
EACL. Utrecht: European Association for Computational Linguistics, 193–202.
Kaplan, Ronald M. and Jürgen Wedekind (2000). ‘LFG Generation Produces
Context-Free Languages.’ In Proceedings of the International Conference
on Computational Linguistics (COLING2000). International Committee on
Computational Linguistics, 425–431.
Kaplan, Ronald M. and Annie Zaenen (1989a). ‘Functional Precedence and
Constituent Structure.’ In Chu-Ren Huang, and Keh-Jiann Chen (eds.),
ROCLING II: Proceedings of the Computational Linguistics Conference. Taipei,
19–40.
Kaplan, Ronald M. and Annie Zaenen (1989b). ‘Long-Distance Dependencies,
Constituent Structure, and Functional Uncertainty.’ In Mark R. Baltin, and
Anthony S. Kroch (eds.), Alternative Conceptions of Phrase Structure, Chicago:
University of Chicago Press, 17–42. Reprinted in Dalrymple et al. (1995b,
137–165).
Karttunen, Lauri (1976). ‘Discourse Referents.’ In McCawley (1976), 363–386.
Karttunen, Lauri (1977). ‘The Syntax and Semantics of Questions.’ Linguistics and
Philosophy 1: 3–44.
Kasper, Robert (1997). ‘The Semantics of Recursive Modification.’ URL http://
www.essex.ac.uk/linguistics/external/clmt/papers/hpsg/modification.ps.
Unpublished manuscript, The Ohio State University.
references 775
Kibort, Anna (2006). ‘On Three Different Types of Subjectlessness and How to
Model Them in LFG.’ In Butt and King (2006b). URL http://csli-publications.
stanford.edu/LFG/11/lfg06.html.
Kibort, Anna (2007). ‘Extending the Applicability of Lexical Mapping Theory.’
In Butt and King (2007), 250–270. URL http://csli-publications.stanford.edu/
LFG/12/lfg07.html.
Kibort, Anna (2008). ‘On the Syntax of Ditransitive Constructions.’ In Butt and
King (2008). URL http://csli-publications.stanford.edu/LFG/13/lfg08.html.
Kibort, Anna (2012). ‘Participles, Adjectives, and the Role of Argument Structure.’
In Butt and King (2012). URL http://csli-publications.stanford.edu/LFG/17/
lfg12.html.
Kibort, Anna and Joan Maling (2015). ‘Modelling the Syntactic Ambiguity of the
Active vs. Passive Impersonal in LFG.’ In Butt and King (2015b). URL http://
csli-publications.stanford.edu/LFG/20/lfg15.html.
Kim, Boomee (1995). ‘Predication in Tough Constructions.’ In José Camacho,
Lina Choueri, and Maki Watanabe (eds.), WCCFL 14: Proceedings of the 14th
West Coast Conference on Formal Linguistics. Stanford: CSLI Publications,
271–285.
King, Tracy Holloway (1995). Configuring Topic and Focus in Russian. Disser-
tations in Linguistics. Stanford: CSLI Publications. Revised version of 1993
Stanford University dissertation.
King, Tracy Holloway (1997). ‘Focus Domains and Information-Structure.’ In Butt
and King (1997). URL http://csli-publications.stanford.edu/LFG/LFG2-1997/.
King, Tracy Holloway (2005). ‘Cliticizing LFG.’ In Butt and King (2005). URL
http://csli-publications.stanford.edu/LFG/10/lfg05.html.
King, Tracy Holloway and Mary Dalrymple (2004). ‘Determiner Agreement and
Noun Conjunction.’ Journal of Linguistics 40(1): 69–104.
King, Tracy Holloway and Valeria de Paiva (eds.) (2013). From Quirky Case to Rep-
resenting Space: Papers in Honor of Annie Zaenen. Stanford: CSLI Publications.
King, Tracy Holloway, Martin Forst, Jonas Kuhn, and Miriam Butt (2005).
‘The Feature Space in Parallel Grammar Writing.’ Research on Language and
Computation 3(2-3): 139–163.
King, Tracy Holloway and Annie Zaenen (2004). ‘F-Structures, Information
Structure, and Discourse Structure.’ URL http://cslipublications.stanford.edu/
LFG/9/abstracts/lfg04kingzaenen-abs.html. Presented at the Winter School in
LFG and Computational Linguistics, University of Canterbury, Christchurch,
New Zealand.
Kiparsky, Paul (1982). ‘Word-Formation and the Lexicon.’ In F. Ingemann (ed.),
Proceedings of the 1982 Mid-America Linguistics Conference.
Kiparsky, Paul (1998). ‘Partitive Case and Aspect.’ In Miriam Butt, and Wilhelm
Geuder (eds.), The Projection of Arguments: Lexical and Compositional Factors,
Stanford: CSLI Publications, 265–307.
Kiparsky, Paul and Carol Kiparsky (1970). ‘Fact.’ In Manfred Bierwisch, and
Karl E. Heidolph (eds.), Progress in Linguistics, The Hague: Mouton de Gruyter,
143–173.
Kiss, Tibor (1995). ‘Lexical-Functional Grammar.’ In Joachim Jacobs, Arnim
von Stechow, Wolfgang Sternefeld, and Theo Vennemann (eds.), Syntax:
references 777
Maxwell, John T., III and Ronald M Kaplan (1996). ‘Unification-Based Parsers
That Automatically Take Advantage of Context Freeness.’ In Butt and King
(1996a). URL http://csli-publications.stanford.edu/LFG/LFG1-1996/.
Maxwell, John T., III and Christopher D. Manning (1996). ‘A Theory of Non-
Constituent Coordination Based on Finite-State Rules.’ In Butt and King
(1996a). URL http://csli-publications.stanford.edu/LFG/LFG1-1996/. Also
presented at the Symposium on Coordination, 1996 European Summer School
on Logic, Language, and Information, Prague.
May, Robert (1977). The Grammar of Quantification. Ph.D. thesis, Massachusetts
Institute of Technology.
Mayer, Elisabeth (2006). ‘Optional Direct Object Clitic Doubling in Limeño
Spanish.’ In Butt and King (2006b). URL http://csli-publications.stanford.edu/
LFG/11/lfg06.html.
Mayer, Elisabeth (2008). ‘Clitics on the Move: From Dependent Marking to Split
Marking.’ In Butt and King (2008). URL http://csli-publications.stanford.edu/
LFG/13/lfg08.html.
McCawley, James D. (1968). ‘Concerning the Base Component of a
Transformational Grammar.’ Foundations of Language 4: 243–269. Reprinted
in McCawley (1973).
McCawley, James D. (1973). Grammar and Meaning: Papers on Syntactic and
Semantic Topics. Tokyo: Taishukan.
McCawley, James D. (ed.) (1976). Notes From the Linguistic Underground, volume 7
of Syntax and Semantics. New York: Academic Press.
McCawley, James D. (1982). ‘Parentheticals and Discontinuous Constituent
Structure.’ Linguistic Inquiry 13(1): 91–106.
McCawley, James D. (1988). The Syntactic Phenomena of English. Chicago:
University of Chicago Press.
McCloskey, James (1979). Transformational Syntax and Model Theoretic Semantics:
A Case Study in Modern Irish. Dordrecht: D. Reidel.
McCloskey, James (1990). ‘Resumptive Pronouns, A# Binding, and Levels of
Representation in Irish.’ In Randall Hendrick (ed.), The Syntax of the Modern
Celtic Languages, New York: Academic Press, volume 23 of Syntax and
Semantics, 199–256.
McCloskey, James (2002). ‘Resumption, Successive Cyclicity, and the Locality of
Operations.’ In Samuel David Epstein, and T. Daniel Seely (eds.), Derivation
and Explanation in the Minimalist Program, Malden, MA: Blackwell,
184–226.
McConnell-Ginet, Sally (1979). ‘On the Deep (And Surface) Adjective ‘Good’.’ In
Frans Van Coetsem, and Linda R. Waugh (eds.), Contributions to Grammatical
Studies: Semantics and Syntax, Leiden: Brill, 132–150.
McGregor, R. S. (1972). Outline of Hindi Grammar. Oxford & Delhi: Oxford
University Press.
Mel’čuk, Igor A. (1988). Dependency Syntax: Theory and Practice. Albany: State
University of New York Press.
Miliĉević, Maja (2009). ‘On the Status of Clitic Reflexives and Reciprocals in
Italian and Serbian.’ In Butt and King (2009), 441–458. URL http://csli-
publications.stanford.edu/LFG/14/lfg09.html.
references 785
Nordlinger, Rachel and Louisa Sadler (2003). ‘The Syntax and Semantics of
Tensed Nominals.’ In Butt and King (2003b). URL http://csli-publications.
stanford.edu/LFG/8/lfg03.html.
Nordlinger, Rachel and Louisa Sadler (2004a). ‘Nominal Tense in Crosslinguistic
Perspective.’ Language 80(4): 776–806.
Nordlinger, Rachel and Louisa Sadler (2004b). ‘Tense Beyond the Verb: Encoding
Clausal Tense/Aspect/Mood on Nominal Dependents.’ Natural Language and
Linguistic Theory 22(3): 597–641.
Nordlinger, Rachel and Louisa Sadler (2007). ‘Verbless Clauses: Revealing the
Structure Within.’ In Zaenen et al. (2007), 139–160.
Nordlinger, Rachel and Louisa Sadler (2008). ‘From Juxtaposition to Incorpora-
tion: An Approach to Generic-Specific Constructions.’ In Butt and King (2008),
394–412. URL http://csli-publications.stanford.edu/LFG/13/lfg08.html.
Nunberg, Geoffrey, Ivan A. Sag, and Thomas Wasow (1994). ‘Idioms.’ Language
70(3): 491–538.
O’Connor, Mary Catherine (1987). Topics in Northern Pomo Grammar. Ph.D.
thesis, University of California, Berkeley. Reprinted by Garland Press, New
York, 1992.
O’Connor, Rob (2002a). ‘Clitics and Phrasal Affixation in Constructive
Morphology.’ In Butt and King (2002). URL http://csli-publications.stanford.
edu/LFG/7/lfg02.html.
O’Connor, Rob (2002b). ‘The Placement of Enclitics in Bosnian, Croatian and
Serbian.’ URL http://roa.rutgers.edu/files/521-0502/521-0502-OCONNOR-0-
0.PDF.gz. Unpublished manuscript, University of Manchester.
O’Connor, Rob (2005a). ‘Clitics in LFG: Prosodic Structure and Phrasal Affixation.’
URL http://csli-publications.stanford.edu/LFG/10/lfg05.html. Abstract in On-
Line Proceedings of the LFG2005 Conference, ed. Miriam Butt and Tracy
Holloway King.
O’Connor, Rob (2005b). ‘Information Structure in Lexical-Functional Grammar:
The Discourse-Prosody Correspondence in English and Serbo-Croatian.’ URL
http://csli-publications.stanford.edu/LFG/10/lfg05.html. Abstract in On-Line
Proceedings of the LFG2005 Conference, ed. Miriam Butt and Tracy Holloway
King.
O’Connor, Robert (2006). Information Structure in Lexical Functional Grammar:
The Discourse-Prosody Correspondence. Ph.D. thesis, University of Manchester.
Odden, David (1995). ‘Phonology At The Phrasal Level In Bantu.’ In Francis
Katamba (ed.), Bantu Phonology and Morphology, Munich: LINCOM, 40–68.
Oehrle, Dick (1999). ‘LFG as Labeled Deduction.’ In Dalrymple (1999).
Oehrle, Richard T., Emmon Bach, and Deirdre Wheeler (eds.) (1988). Categorial
Grammars and Natural Language Structures. Dordrecht: D. Reidel.
Ohala, John J. (1989). ‘Sound Change Is Drawn From A Pool Of Synchronic
Variation.’ In Leiv Egil Breivik, and Ernst Håkon Jahr (eds.), Language Change:
Contributions to the Study of its Causes, Berlin: Mouton de Gruyter, 173–198.
Ostler, Nicholas D. M. (1979). Case-Linking: A Theory of Case and Verb
Diathesis Applied to Classical Sanskrit. Ph.D. thesis, Massachusetts Institute of
Technology.
references 789
Patejuk, Agnieszka and Adam Przepiórkowski (2014). ‘In Favour of the Raising
Analysis of Passivisation.’ In Butt and King (2014). URL http://csli-publications.
stanford.edu/LFG/19/lfg14.html.
Patejuk, Agnieszka and Adam Przepiórkowski (2015). ‘An LFG Analysis of the
So-Called Reflexive Marker in Polish.’ In Butt and King (2015b). URL http://
csli-publications.stanford.edu/LFG/20/lfg15.html.
Patejuk, Agnieszka and Adam Przepiórkowski (2017). ‘Reducing Grammatical
Functions in LFG.’ In Arnold et al. (2017). URL http://cslipublications.stanford.
edu/HPSG/2016/.
Payne, John (2009). ‘The English Genitive and Double Case.’ Transactions of the
Philological Society 107(3): 322–357.
Payne, John and Kersti Börjars (2015). ‘Features and Selection in LFG: The
English VP.’ In Butt and King (2015b). URL http://csli-publications.stanford.
edu/LFG/20/lfg15.html.
Pereira, Fernando C. N. and David H. D. Warren (1983). ‘Parsing as Deduction.’ In
Proceedings of the 21st Annual Meeting of the ACL. Cambridge, MA: Association
for Computational Linguistics, 137–144.
Perlmutter, David (ed.) (1983). Studies in Relational Grammar 1. Chicago:
University of Chicago Press.
Perlmutter, David M. (1971). Deep and Surface Structure Constraints in Syntax.
New York: Holt, Rinehart & Winston.
Perlmutter, David M. and Paul M. Postal (1977). ‘Toward a Universal
Characterization of Passivization.’ In Proceedings of the Third Annual Meeting
of the Berkeley Linguistics Society. Berkeley: Berkeley Linguistics Society.
Reprinted in Perlmutter (1983, 3–29).
Perlmutter, David M. and Paul M. Postal (1983). ‘Some Proposed Laws of Basic
Clause Structure.’ In Perlmutter (1983), 81–128.
Perlmutter, David M. and Paul M. Postal (1984). ‘The 1-Advancement
Exclusiveness Law.’ In David Perlmutter, and Carol Rosen (eds.), Studies
in Relational Grammar 2, Chicago: University of Chicago Press.
Peters, Stanley and Dag Westerståhl (2006). Quantifiers in Language and Logic.
Oxford: Oxford University Press.
Peterson, John L. (1981). Petri Net Theory and the Modeling of Systems. London:
Prentice Hall.
Peterson, Peter G. (1982). ‘Conjunction in LFG.’ Unpublished manuscript,
Department of Linguistics, MIT and University of Newcastle, New South
Wales, Australia.
Peterson, Peter G. (2004a). ‘Coordination: Consequences of a Lexical-Functional
Account.’ Natural Language and Linguistic Theory 22(3): 643–679.
Peterson, Peter G. (2004b). ‘Introduction to the Workshop on Coordination and
Agreement.’ In Butt and King (2004b). URL http://csli-publications.stanford.
edu/LFG/9/lfg04.html.
Pica, Pierre (1987). ‘On the Nature of the Reflexivization Cycle.’ In Joyce
McDonough, and Bernadette Plunkett (eds.), Proceedings of NELS 17. Graduate
Linguistic Student Association, University of Massachusetts, Amherst,
483–500.
references 791
Reape, Mike (1994). ‘Domain Union and Word Order Variation in German.’ In
John Nerbonne, Klaus Netter, and Carl Pollard (eds.), German in Head-Driven
Phrase Structure Grammar, Stanford: CSLI Publications, number 46 in CSLI
Lecture Notes, 151–197.
Reichenbach, Hans (1947). Elements of Symbolic Logic. London: Macmillan.
Reinhart, Tanya (1976). The Syntactic Domain of Anaphora. Ph.D. thesis,
Massachusetts Institute of Technology.
Reinhart, Tanya (1981). ‘Pragmatics and Linguistics: An Analysis of Sentence
Topics.’ Philosophica 27: 53–94.
Reinhart, Tanya (2002). ‘The Theta System — An Overview.’ Theoretical Linguistics
28: 229–290.
Reinhart, Tanya and Tal Siloni (2005). ‘The Lexicon-Syntax Parameter:
Reflexivization and Other Arity Operations.’ Linguistic Inquiry 36(3): 389–436.
Reyle, Uwe (1988). ‘Compositional Semantics for LFG.’ In Reyle and Rohrer
(1988), 448–474.
Reyle, Uwe (1993). ‘Dealing With Ambiguities by Underspecification: Construc-
tion, Representation and Deduction.’ Journal of Semantics 10: 123–179.
Reyle, Uwe and Christian Rohrer (eds.) (1988). Natural Language Parsing and
Linguistic Theories. Dordrecht: D. Reidel.
Rhodes, Richard (1990). ‘Ojibwa Secondary Objects.’ In Dziwirek et al. (1990),
401–414.
Riezler, Stefan, Tracy Holloway King, Ronald M. Kaplan, Richard Crouch, John T.
Maxwell, III, and Mark Johnson (2002). ‘Parsing the Wall Street Journal Using
a Lexical-Functional Grammar and Discriminative Estimation Techniques.’ In
Proceedings of the 40th Annual Meeting of the ACL. Philadelphia: Association
for Computational Linguistics.
Rizzi, Luigi (1990). Relativized Minimality. Cambridge, MA: The MIT Press.
Rizzi, Luigi (1997). ‘The Fine Structure of the Left Periphery.’ In Liliane M. V.
Haegeman (ed.), Elements of Grammar: Handbook of Generative Syntax,
Dordrecht & Boston: Kluwer, 281–337.
Robins, R. H. (1967). A Short History of Linguistics. Bloomington: Indiana
University Press.
Rögnvaldsson, Eiríkur (1984). ‘Af Lýsingarorðsviðurlögum.’ Íslenskt mál 6: 57–80.
Rosen, Carol (1990). ‘Rethinking Southern Tiwa: The Geometry of a Triple-
Agreement Language.’ Language 66(4): 669–713.
Rosén, Victoria (1996). ‘The LFG Architecture and “Verbless” Syntactic
Constructions.’ In Butt and King (1996a). URL http://csli-publications.
stanford.edu/LFG/LFG1-1996/.
Ross, John Robert (1967). Constraints on Variables in Syntax. Ph.D. thesis,
Massachusetts Institute of Technology.
Rounds, William C. (1988). ‘Set Values for Unification-Based Grammar
Formalisms and Logic Programming.’ Technical Report CSLI-88-129,
CSLI Publications, Stanford.
Rounds, William C. (1997). ‘Feature Logics.’ In van Benthem and ter Meulen
(1997), 475–533. Co-published with Elsevier Science B.V., Amsterdam.
Rubio Vallejo, David (2011). Impersonal Constructions in Spanish: An LFG
Approach Based On Gradience. M.Phil. dissertation, University of Oxford.
references 795
Schwarze, Christoph (2001a). ‘On the Representation of French and Italian Clitics.’
In Butt and King (2001a), 280–304. URL http://csli-publications.stanford.edu/
LFG/6/lfg01.html.
Schwarze, Christoph (2001b). ‘Representation and Variation: On the Development
of Romance Auxiliary Syntax.’ In Butt and King (2001b), 143–172.
Schwarze, Christoph and Leonel Figueiredo de Alencar (2016). Lexikalisch-
funktionale Grammatik : Eine Einführung am Beispiel des Französischen mit
computerlinguistischer Implementierung. Tübingen: Stauffenburg. URL http://
nbn-resolving.de/urn:nbn:de:bsz:352-0-318774.
Schwarzschild, Roger (1994). ‘Plurals, Presuppositions and the Sources of
Distributivity.’ Natural Language Semantics 2(3): 201–248.
Seiss, Melanie (2008). ‘The English -ing Form.’ In Butt and King (2008), 454–472.
URL http://csli-publications.stanford.edu/LFG/13/lfg08.html.
Seiss, Melanie (2009). ‘On the Difference Between Auxiliaries, Serial Verbs and
Light Verbs.’ In Butt and King (2009), 501–519. URL http://csli-publications.
stanford.edu/LFG/14/lfg09.html.
Seiss, Melanie and Rachel Nordlinger (2010). ‘Applicativizing Complex Predicates:
A Case Study From Murrinh-Patha.’ In Butt and King (2010), 416–436. URL
http://csli-publications.stanford.edu/LFG/15/lfg10.html.
Seki, Hiroyuki, Takashi Matsumura, Mamoru Fujii, and Tadao Kasami (1991).
‘On Multiple Context-Free Grammars.’ Theoretical Computer Science 88(2):
191–229.
Selkirk, Elisabeth O. (1978). ‘On Prosodic Structure and Its Relation to Syntactic
Structure.’ In T. Fretheim (ed.), Nordic Prosody II, Trondheim: TAPIR.
Selkirk, Elisabeth O. (1980a). ‘Prosodic Domains in Phonology: Sanskrit
Revisited.’ In Mark Aronoff, and Mary-Louise Kean (eds.), Juncture, Saratoga,
CA: Anma Libri.
Selkirk, Elisabeth O. (1980b). ‘The Role of Prosodic Categories in English Word
Stress.’ Linguistic Inquiry 11: 563–605.
Selkirk, Elisabeth O. (1981). ‘On the Nature of Phonological Representation.’
In Terry Myers, John Laver, and John Anderson (eds.), The Cognitive
Representation of Speech, Amsterdam: North Holland, 379–388.
Selkirk, Elisabeth O. (1984). Phonology and Syntax: The Relation Between Sound
and Structure. Cambridge, MA: The MIT Press.
Selkirk, Elisabeth O. (1986). ‘On Derived Domains in Sentence Phonology.’
Phonology Yearbook 3: 371–405.
Selkirk, Elisabeth O. (1995). ‘The Prosodic Structure of Function Words.’ In
Jill Beckman, Laura Walsh Dickey, and Suzanne Urbanczyk (eds.), Papers in
Optimality Theory. Amherst, MA: Graduate Linguistic Student Association,
University of Massachusetts, Amherst, volume 18 of University of Massachusetts
Occasional Papers in Linguistics, 439–469.
Selkirk, Elisabeth O. (2011). ‘The Syntax-Phonology Interface.’ In John A.
Goldsmith, Jason Riggle, and Alan C. L. Yu (eds.), The Handbook of Phonological
Theory, Oxford: Wiley-Blackwell, 435–484. Second edition.
Selkirk, Elisabeth O. and Tong Shen (1990). ‘Prosodic Domains in Shanghai
Chinese.’ In Inkelas and Zec (1990), 313–337.
798 references
Zaenen, Annie and Elisabet Engdahl (1994). ‘Descriptive and Theoretical Syntax in
the Lexicon.’ In Beryl T. S. Atkins, and Antonio Zampolli (eds.), Computational
Approaches to the Lexicon, Oxford: Oxford University Press, 181–212.
Zaenen, Annie and Ronald M. Kaplan (1995). ‘Formal Devices for Linguistic
Generalizations: West Germanic Word Order in LFG.’ In Jennifer S. Cole, Geor-
gia M. Green, and Jerry L. Morgan (eds.), Linguistics and Computation, Stan-
ford: CSLI Publications, 3–27. Reprinted in Dalrymple et al. (1995b, 215–239).
Zaenen, Annie and Ronald M. Kaplan (2002). ‘Subsumption and Equality:
German Partial Fronting in LFG.’ In Butt and King (2002). URL http://csli-
publications.stanford.edu/LFG/7/lfg02.html.
Zaenen, Annie and Lauri Karttunen (1984). ‘Morphological Non-Distinctiveness
and Coordination.’ In Mark Cobler, Susannah MacKaye, and Michael T.
Wescoat (eds.), Proceedings of the Eastern States Conference on Linguistics
(ESCOL ’84). 309–320.
Zaenen, Annie and Joan Maling (1983). ‘Passive and Oblique Case.’ In Levin et al.
(1983), 159–191.
Zaenen, Annie and Joan Maling (1990). ‘Unaccusative, Passive, and Quirky Case.’
In Maling and Zaenen (1990), 187–234.
Zaenen, Annie, Joan Maling, and Höskuldur Þráinsson (1985). ‘Case and
Grammatical Functions: The Icelandic Passive.’ Natural Language and Linguistic
Theory 3(4): 441–483. Reprinted in Maling and Zaenen (1990, 95–164).
Zaenen, Annie, Jane Simpson, Tracy Holloway King, Jane Grimshaw, Joan Maling,
and Christopher Manning (eds.) (2007). Architectures, Rules, and Preferences:
Variations on Themes by Joan W. Bresnan. Stanford: CSLI Publications.
Zec, Draga (1987). ‘On Obligatory Control in Clausal Complements.’ In Iida et al.
(1987), 139–168.
Zhang, Yanyin (2005). ‘Processing and Formal Instruction in the L2 Acquisition
of Five Chinese Grammatical Morphemes.’ In Pienemann (2005a), 155–177.
Zoller, Claus Peter (1983). Die Sprache der Rang Pas von Garhwal (Ran Pɔ Bhāsa):
Grammatik, Texte, Wörterbuch [The Language of the Rang Pas of Garhwal:
Grammar, Texts, Dictionary]. Wiesbaden: Harrassowitz.
Zwicky, Arnold M. (1990). ‘Syntactic Representations and Phonological Shapes.’
In Inkelas and Zec (1990), 379–397.
Zwicky, Arnold M. (1992). ‘Some Choices in the Theory of Morphology.’ In
Robert D. Levine (ed.), Formal Grammar: Theory and Implementation, Oxford:
Oxford University Press, 327–371.
Zybatow, Gerhild and Grit Mehlhorn (2000). ‘Experimental Evidence for Focus
Structure in Russian.’ In Tracy Holloway King, and Irina Sekerina (eds.),
Formal Approaches To Slavic Linguistics: The Philadelphia Meeting 1999. FASL,
Ann Arbor: Michigan Slavic Publications, 414–434.
Zymla, Mark-Matthias, Maike Müller, and Miriam Butt (2015). ‘Modelling the
Common Ground for Discourse Particles.’ In Butt and King (2015b). URL
http://csli-publications.stanford.edu/LFG/20/lfg15.html.
Author Index
Borer, Hagit 135, 445 Chisarik, Erika 36, 37, 99, 364, 725
Börjars, Kersti 6, 26, 28, 59, 63, 71, 73, 75, Choi, Hye-Won 371–3, 376, 377
76, 99, 101, 116, 136, 221, 222, 247, 354, Chomsky, Noam 1, 21, 35, 43, 90, 91, 96,
364, 453, 498, 636, 639, 729, 730 97, 99, 107, 135, 247, 330, 334, 370, 398,
Borschev, Vladimir 707 447, 706
Börschinger, Benjamin 352 Chung, Sandra 687
Borsley, Robert D. 603 Cinque, Guglielmo 99, 656
Bos, Johan 287 Clark, Brady Zack 681
Bošković, Željko 398 Clément, Lionel 734
Bouchard, Denis 703 Clements, George N. 687, 688
Bouma, Gosse 695 Colban, Erik 498
Brame, Michael 100 Comrie, Bernard xix, 10, 11, 20, 63,
Bresnan, Joan 2, 6, 11, 14, 15, 17, 21, 26, 366, 816
30, 34–7, 41–4, 50, 52, 53, 61, 64, 69, 88, Cook, Phillipa 93
89, 97, 99, 101, 104, 108, 112, 115, 116, Cooper, Robin 268, 284
118–21, 128, 130–2, 134–8, 140, 147, Coppock, Elizabeth 502, 730
149, 150, 156, 161, 163, 164, 166, 168, Corbett, Greville G. 55, 74, 451, 639
172, 176, 177, 181, 183, 184, 194, 197, Corblin, Francis 522
211, 212, 225, 231, 235, 238, 245–7, 253, Cormons, Boris 727
259, 264, 278, 297, 322, 324, 329–38, Covington, Michael A. 9
341, 342, 345–7, 353, 355, 356, 358–60, Crouch, Richard 106, 198, 225, 237, 249,
364, 374, 375, 386, 452, 470, 471, 500–5, 250, 287, 293, 305, 319, 321, 359, 517,
510, 512–14, 543–5, 547, 550, 552, 561, 619, 624–7, 631, 634, 635, 644, 724,
565, 576, 577, 585, 586, 589, 593–5, 602, 734, 735
605, 621, 650, 652, 659, 665, 666, 669, Crysmann, Berthold 470
673, 676, 680, 682, 687, 694, 696–8, 700, Cueni, Anna 355
703, 725–8, 730, 732 Culicover, Peter W. 3, 273
Broadwell, George Aaron 6, 104, 115, 136, Culy, Christopher D. 509, 510, 544
324, 352, 495, 725, 728 Curnow, Timothy Jowan 190
Brown, Dunstan 55, 436, 451
Brown, Lea 650 Dahl, Östen 370
Brun, Caroline 244, 613 Dahlstrom, Amy 25, 28, 138, 184, 365, 394
Bürckert, Hans-Jürgen 209 Dale, Robert 352
Butt, Miriam 6, 21, 28, 32–4, 60, 61, 63–5, Dalrymple, Mary 2, 6, 14, 19, 28, 30–3, 35,
67, 79, 88, 96, 102, 103, 115, 120, 138, 37, 69, 71–5, 78, 81, 82, 96, 106, 114–16,
195, 237, 314, 324, 326, 327, 329, 335, 121, 132, 136, 138, 189, 192, 193, 196–8,
346–52, 356, 360, 361, 363, 365, 372, 209, 219, 220, 225, 232, 233, 237, 249,
373, 376–81, 383, 394, 400–5, 417, 418, 250, 264, 278, 280, 284, 287, 301, 302,
436, 438, 443–5, 447–50, 457, 458, 492, 305, 313, 319, 320, 326, 327, 352, 355,
498, 502, 601, 666, 671, 680, 725, 360, 361, 365, 368, 370, 373, 374, 379,
729, 735 381–3, 385, 386, 393, 406–9, 412,
416–18, 435, 439, 503, 505, 507,
Camilleri, Maris 684 510–12, 515, 517–19, 523, 539, 543,
Carnie, Andrew 6, 115, 754 544, 570, 571, 605, 619–21, 624, 630,
Cavar, Damir 116, 734 636, 639, 642, 643, 651, 652, 680–2, 687,
Çetinoğlu, Özlem 28, 352, 365 697–703, 716, 717, 730, 734
Champollion, Lucas 643, 645 Daneš, František 370
Chapman, Carol 453, 730 Davidson, Donald 314
Chen, Matthew Y. 398 Davis, Anthony 364
Chierchia, Gennaro 569, 570 Davison, Alice 102
author index 809
Hajičová, Eva 370 Jackendoff, Ray 3, 96, 97, 247, 273, 324,
Håkansson, Gisela 728 326, 370, 377
Hale, Kenneth 40, 41, 100 Jacobson, Pauline 549, 556, 564, 570, 600,
Hale, Mark 729 681
Halle, Morris 135, 398, 445 Jaeger, T. Florian 114
Halliday, M. A. K. 370 Jaeggli, Osvaldo 131
Halpern, Aaron 115 Jakobson, Roman 57
Halvorsen, Per-Kristian 266–9, 278–80, Jam, R. 225
286, 297, 494, 503–5, 556, 569 Jansen, Louise 727
Hamblin, C. L. 721, 722 Janssen, Theo M. V. 276
Hammarberg, Björn 728 Jelinek, Eloise 166, 184, 502
Harries, Pauline 729 Jeschull, Liane 622
Haspelmath, Martin 6, 63, 621, 622, 816 Johannessen, Janne Bondi 603
Haug, Dag T. T. 69, 179, 197, 287, 314, 316, Johnson, David E. 3, 37
320, 500, 518–21, 523, 534, 544, 556, Johnson, Mark 138, 164, 246, 728, 734
583, 584, 587, 590, 601, 671, 721, 725 Jones, Stephen M. 78
Hautli, Annette 348 Joshi, Smita 338, 348
Hay, Jennifer 355 Junghanns, Uwe 95
Hayes, Bruce 396, 401, 407, 418–20
Hearne, Mary 727 Kaisse, Ellen M. 397
Heim, Irene 518 Kallas, Krystyna 650, 651
Hellan, Lars 14, 505–7, 509, 511, 515 Kameyama, Megumi 118, 257, 259, 514
Henadeerage, Kumara 503 Kamp, Hans 281, 287, 315, 476–9, 518,
Heny, Frank 494 519, 522
Her, One-Soon 335, 337, 338, 354 Kanazawa, Makoto 571
Heycock, Caroline 643 Kanerva, Jonni M. 21, 26, 247, 331–4, 338,
Higginbotham, James 526, 570, 571, 722 345–7
Hinds, John 396 Kaplan, Ronald M. 2, 6, 14, 30, 50, 52, 53,
Hinrichs, Angie 278 59, 61, 72–5, 81, 93, 94, 96, 106–8, 114,
Hippisley, Andrew 436 115, 131, 132, 138, 140–2, 147, 149, 150,
Hirschberg, Julia 403 163, 164, 171, 198, 201, 203–5, 208, 209,
Hoeksema, Jacob 643, 644 211, 212, 219, 220, 222, 223, 225, 232,
Holmen, Anne 728 237, 238, 241–7, 249, 250, 252, 257, 259,
Hong, Ki-Sun 394 264, 266, 267, 273, 278, 280, 286, 297,
Hornstein, Norbert 656 324, 329, 330, 351, 404–6, 410, 436, 438,
Horrack, Kate 352 443–5, 457, 458, 471, 514, 553, 554, 564,
Hristov, Bozhil 69, 71, 78, 80, 218, 603, 602, 605, 619, 636, 639, 650–2, 662, 663,
636, 642, 664 666, 673, 680, 682, 687, 694, 696–701,
Huang, C.-T. James 15 726, 727, 730–4
Huang, Chu-Ren 138, 678, 725 Karttunen, Lauri 56, 80, 436, 518, 522,
Huddleston, Rodney 63, 609, 623 722, 734
Hudson, Richard A. 9, 25 Kasami, Tadao 732
Hurst, Peter 544 Kasper, Robert 473, 484, 485, 489
Hvenekilde, Anne 728 Kathol, Andreas 69
Kawaguchi, Satomi 727
Iida, Masayo 364 Kay, Martin 149, 204, 436, 734
Inkelas, Sharon 400 Kay, Paul 2, 3, 127, 272
Ishikawa, Akira 351 Kazana, Despina 287
Itô, Junko 400 Keenan, Edward L. 10, 11, 20, 284, 634
Izvorski, Roumyana 125 Kehler, Andrew 624, 630
author index 811
Saraswat, Vijay A. 164, 278, 280, 287, Sulger, Sebastian 33, 35–7, 103, 195, 348,
301–3, 305, 319, 517, 539, 624, 630, 380, 394
716, 717 Sung, Won-Kyung 503
Scedrov, Andre 319 Svartvik, Jan 29, 60, 158
Schachter, Paul 21 Szűcs, Peter 665
Scheer, Tobias 397, 399
Schein, Barry 314 Tang, Chih-Chen Jane 507
Schwarze, Christoph 6, 96, 729 Tantos, Alexandros 350
Schwarzschild, Roger 644, 645 ter Meulen, Alice 5, 47, 199, 256, 271, 281,
Segond, Frédérique 120, 436, 282, 284, 302, 731
447–50 Tesnière, Lucien 9
Seiss, Melanie 116, 352, 452 Thomas, Victoria 88
Seki, Hiroyuki 732 Þráinsson, Höskuldur 21, 25, 94, 331
Selkirk, Elisabeth O. 6, 398–400, 412 Togo, Patrice 544
Sells, Peter 6, 89, 94, 96, 100, 104, 116, 138, Toivonen, Ida 2, 6, 11, 14, 15, 30, 35–7, 43,
515, 544, 553–5, 728 44, 62, 69, 88, 89, 96, 97, 99, 104, 105,
Sgall, Petr 370 108, 112, 116, 119–21, 126, 127, 130,
Sharma, Devyani 730 131, 134–8, 168, 172, 176, 177, 181, 184,
Shattuck-Hufnagel, Stefanie 397, 403 194, 197, 225, 231, 233, 237, 247, 259,
Shemtov, Hadar 733 320, 327, 329, 331, 335–7, 353, 356,
Shen, Tong 398 359–62, 364, 386, 470, 471, 474, 502,
Shieber, Stuart M. 47, 149, 284, 570 503, 505, 510, 512–14, 543, 544, 550,
Siegel, Muffy E. A. 477, 478 552, 559, 560, 565, 576, 577, 593, 621,
Siewierska, Anna 543 659, 696–8, 700, 703, 725
Sigurjónsdóttir, Sigríður 345 Turpin, Myfany 352
Siloni, Tal 334
Simpson, Jane 29, 31, 34, 41, 93, 106, 107, Ulinski, Morgan 352
112, 114, 128, 134, 136, 151, 250–2, 336, Uszkoreit, Hans 189
352, 355, 356, 363, 365, 470, 544, 587,
594, 595, 600 Vallduví, Enric 369–72, 378
Sims, Andrea D. 6 van Genabith, Josef 102, 287, 293, 305,
Smolensky, Paul 365 319, 321, 517, 727
Smolka, Gert 209 Van Herk, Gerard 355
Snijders, Liselotte 40, 108, 121, 138, van Reimsdijk, Henk 80
368, 503 van Urk, Coppe 585
Spector, Ilona 102 Van Valin, Robert D. 273
Spencer, Andrew 6, 66, 102, 105, 135, Vigo, Eugenio M. 197
329, 352, 364, 436, 439, 440, 443, 451–3, Vilkuna, Maria 369
470, 601 Villavicencio, Aline 642
Srivastav, Veneeta 671 Vincent, Nigel 26, 28, 71, 73, 75, 76, 221,
Steedman, Mark J. 2, 301, 370, 634 222, 247, 354, 364, 453, 498, 636, 639,
Steele, Joshua 398 729, 730
Stewart, Thomas W. 6, 135, 442 Visser, Fredericus T. 585
Stirling, Lesley 544 Vogel, Irene 6, 398, 400
Stokhof, Martin 518, 522, 538, 722 von Stechow, Arnim 381, 382
Stowell, Timothy 103
Strahan, Tania E. 503, 544 Wada, Hajime 287
Strand, Kjetil 503, 505 Wall, Robert E. 5, 47, 199, 256, 271, 281,
Stump, Gregory T. 6, 135, 436, 437, 439, 282, 284, 302, 731
442, 443, 455, 456, 460, 464, 467 Wang, Haifeng 102
author index 815
gf
! (grammatical function) 22–4, 38 Head marking language 177
gf (metavariable over grammatical Head-Driven Phrase Structure
functions) 206 Grammar 2, 11, 41, 44n, 47n, 205,
gf-outranks, see Grammatical functions, 266, 273, 275, 352n, 378n
hierarchy human (feature) 313
ggf (metavariable over governable
grammatical functions) 177, 205n I
Glue 287–320, see also Linear logic;
Syntax-semantics interface I see Binding, semantics, index-discourse
and Categorial Grammar 301 referent relation
computational implementation 735 I-structure, see Information structure
first-order 320 ID/LP rule, see under Phrase structure rule
propositional 320 “Ignore” operator (/), see under Phrase
Grammar Writer’s Workbench 734 structure rule
Grammatical functions 3, 11–38, see also Impersonal predicate 343–5, see also
Closed grammatical functions; Subject, condition
Modification; Oblique grammatical Implication, see under Functional
functions; Open grammatical description (⇒); see Linear
functions; Restricted grammatical implication (⊸)
functions; Terms; Unrestricted Indefinite noun phrase 285–6, 558–9, 644,
grammatical functions see also def; see also under
as theoretical primitive 3, 38–42 Discourse Representation Theory
derived arguments 357–63 meaning constructor 558
encoding 176–89, see also index (f-structure feature) 69–71, 85, 523n
Configurationality; as nondistributive feature 70–1, 86,
Dependent marking language; 217–18
Head marking index (semantic structure feature) 523,
language; Lexocentric see also Binding, semantics
organization Indirect object, see under Object
governable 12–38, 50 grammatical functions
and coherence 52–3 Infinitival complement, see comp; xcomp
status of poss 36 phrase structure category 550n
gradient distinctions 357 Information structure 5, 278, 366–94,
hierarchy 10–12, 20, 22, 24, 41, 166, see also Background information;
336, 339, 347, 697, see also Mapping Completive information;
theory, markedness hierarchy; Correspondence function, ι; Focus;
see also under Binding Tail; Topic
inventory 13 features 371–3, 377–93, see also df;
tests for 13–15, 322–3 new; prom
Grammatical relations, see Grammatical Inside-out functional uncertainty, see under
functions Functional uncertainty
Ground (information structure), see Instantiated symbol, see under Functional
Background information description
Intensional logic 280, 281, 286
H Interface Harmony, see under Prosodic
structure
Head (phrase structure position) 98 Interrogative pronoun, see under Pronoun
Convention 118, 388n Intersection (∩) 214, see also under
mapping to f-structure 118, 168–70, 177 Regular language
“movement” 129–130 in feature resolution 639–40
828 subject index