Beginning Syntax
Beginning Syntax
General editors: u. a n s a l d o , p. au s t i n, b . c o m r i e , r . l a s s ,
d. l i g h t f o o t, k . r i c e , i . r o b e rt s , s . r o m a i n e , m . s h e e h a n,
i. tsimpli
Beginning Syntax
An Introduction to Syntactic Analysis
r . ca n n Formal Semantics
j. l av e r Principles of Phonetics
f. r . pa l m e r Grammatical Roles and Relations
m . a . j on e s Foundations of French Syntax
a . r a d f o r d Syntactic Theory and the Structure of English: A Minimalist
Approach
r . d. va n va l i n j r a n d r . j. l a p o l l a Syntax: Structure, Meaning and
Function
a . d u r a n t i Linguistic Anthropology
a . c ru t t e n d e n Intonation Second edition
j. k . c h a m b e r s a n d p. t ru d g i l l Dialectology Second edition
c . lyon s Definiteness
r . k ag e r Optimality Theory
j. a . h o l m An Introduction to Pidgins and Creoles
g . g . c o r b e t t Number
c . j. e w e n h . va n d e r h u l s t The Phonological Structure of Words
f. r . pa l m e r Mood and Modality Second edition
b . j. b l a k e Case Second edition
e . gu s s m a n Phonology: Analysis and Theory
m . y i p Tone
w. c r o f t Typology and Universals Second edition
f. c ou l m a s Writing Systems: An Introduction to Their Linguistic Analysis
p. j. h o p p e r a n d e . c . t r au g o t t Grammaticalization Second edition
l . w h i t e Second Language Acquisition and Universal Grammar
i . p l ag Word-Formation in English
w. c r o f t a n d a . c ru s e Cognitive Linguistics
a . s i e w i e r s k a Person
a . r a d f o r d Minimalist Syntax: Exploring the Structure of English
d. bü r i n g Binding Theory
m . bu t t Theories of Case
n. h o r n s t e i n, j. n u ñ e s a n d k . g r o h m a n n Understanding Minimalism
b . c . lu s t Child Language: Acquisition and Growth
g . g . c o r b e t t Agreement
j. c . l . i n g r a m Neurolinguistics: An Introduction to Spoken Language
Processing and Its Disorders
j. c l ac k s on Indo-European Linguistics: An Introduction
m . a r i e l Pragmatics and Grammar
r . ca n n, r . k e m p s on a n d e . g r e g o r o m i c h e l a k i Semantics: An
Introduction to Meaning in Language
y. m at r a s Language Contact
d. b i b e r a n d s . c on r a d Register, Genre, and Style
l . j e f f r i e s a n d d. m c i n t y r e Stylistics
r . h u d s on An Introduction to Word Grammar
m . l . m u r p h y Lexical Meaning
j. m . m e i s e l First and Second Language Acquisition
t. m c e n e ry a n d a . h a r d i e Corpus Linguistics: Method, Language and
Practice
j. s a k e l a n d d. l . e v e r e t t Linguistic Fieldwork: A Student Guide
a . s p e n c e r a . lu í s Clitics: An Introduction
g . c o r b e t t Features
a . m c m a h on a n d r . m c m a h on Evolutionary Linguistics
b . c l a r k Relevance Theory
b . l on g p e n g Analyzing Sound Patterns
b . da n c yg i e r a n d e . s w e e t s e r Figurative Language
j. by b e e Language Change
s . g . t h o m a s on Endangered Languages: An Introduction
a . r a d f o r d Analysing English Sentences Second edition
r . c l i f t Conversation Analysis
I A N RO BE RT S
Downing College, University of Cambridge
www.cambridge.org
Information on this title: www.cambridge.org/9781316519493
DOI: 10.1017/9781009023849
© Ian Roberts 2023
This publication is in copyright. Subject to statutory exception and to the provisions
of relevant collective licensing agreements, no reproduction of any part may take
place without the written permission of Cambridge University Press & Assessment.
First published 2023
Printed in <country> by <printer>
A catalogue record for this publication is available from the British Library.
Library of Congress Cataloging-in-Publication Data
ISBN 978-1-316-51949-3 Hardback
ISBN 978-1-009-01058-0 Paperback
Cambridge University Press & Assessment has no responsibility for the persistence
or accuracy of URLs for external or third-party internet websites referred to in this
publication and does not guarantee that any content on such websites is, or will
remain, accurate or appropriate.
Preface page xi
List of Abbreviations xiii
vii
Exercises 48
Further Reading 49
4 X′-theory 78
4.1 Introduction 78
4.2 Possible and Impossible PS-Rules 78
4.3 Introducing X′-theory 80
4.4 X′-theory and Functional Categories: The Structure of the Clause 87
4.5 Adjuncts in X′-theory 95
4.6 Conclusion 96
Exercises 97
For Further Discussion 98
Further Reading 98
5 Movement 99
5.1 Introduction and Recap 99
5.2 Types of Movement Rules 100
5.3 Wh-Movement: The Basics 104
5.4 Subject Questions, Indirect Questions and Long-Distance
Wh-Movement 112
5.5 The Nature of Movement Rules 118
5.6 Limiting Wh-Movement 122
5.7 Conclusion 126
Exercises 127
For Further Discussion 128
Further Reading 128
6 Binding 130
6.1 Introduction and Recap 130
6.2 Pronouns and Anaphors 130
6.3 Anaphors 132
6.4 Pronouns 142
6.5 R-expressions 143
6.6 The Binding Principles 144
6.7 Variables, Principle C and Movement 146
6.8 Conclusion 153
Exercises 154
For Discussion 154
Further Reading 155
Conclusion 215
Glossary 219
Index 234
The purpose of this book is to present the essential elements of the theory of
syntax, presupposing no prior knowledge of either syntax or linguistics more
generally. The first chapter introduces the thinking behind modern formal
linguistics, defining the core background concepts. The second chapter is an
extended demonstration of linguistic competence, revealing to an English
speaker their tacit, untutored knowledge of the intricacies of the syntax of the
language. The next four chapters introduce the core technical concepts of syn-
tax: phrase structure, constituency, movement rules and construal rules. Chapter
7 is a very brief introduction to comparative syntax, illustrating how the theory
developed in the earlier chapters can apply to languages other than English.
Finally, Chapter 8 discusses the overall architecture of grammar, introducing
several levels of representation, including the interface levels of Phonological
Form and Logical Form.
The book is based on a first-year lecture course I have taught at Cambridge for
most of the past ten years, called ‘Structures’. That course has successfully laid
the groundwork for more advanced study; several professional syntacticians had
their first exposure to the subject in that context.
I do not attempt to review either the history or the state of the art in contem-
porary generative syntax. Instead, the book is intended to provide a solid basis
from which the student can move on to more advanced topics. In the decades
since its inception, generative theory has yielded a range of core insights and
concepts which can be studied independently of the details of a given framework
and which form the core of any formal approach to syntax (and core background
to the formal study of other areas of linguistics, in particular semantics). These
are the focus of this introduction, where the theoretical orientation of the book is
most clearly set out. This volume is intended as the first in a series; future vol-
umes will build on the material presented here, as well as both adding an explic-
itly comparative and typological dimension and engaging more directly with the
minimalist programme for linguistic theory. The overall aim of the series is to
provide a complete course in syntax, taking the student from beginner level to
being able to engage with contemporary research literature.
This volume is intended for students starting out in syntax and assumes no
prior knowledge. As mentioned above, it corresponds to a first-year course
taught at Cambridge. It can be taught over a single term or semester with tutorial
xi
support (as I have done at Cambridge) or over an entire year as a lecture course.
The book tries to introduce the main concepts of syntactic theory in an engaging
way with a minimum of technicalities. Each chapter features exercises, some
straightforwardly testing the material covered in the chapter, others raising fur-
ther questions for reflection and discussion (e.g. in tutorials), as well as sugges-
tions for further reading. Model answers are provided separately for the first type
of exercise. The second type presupposes student participation in discussions.
Chapter 1 is somewhat different from the others in that it introduces some fairly
complex data; the point at this stage of the exposition is to demonstrate tacit
knowledge, i.e. native-speaker competence; the technical issues are, as appropri-
ate, revisited and introduced more gradually later. Instructors may wish to skip
over some of this material and/or revisit it later (for example, the Italian data can
supplement the material covered in Chapter 7).
I would like to thank the other people who have taught this course in whole or
in part over the years: Dora Alexopoulou, Theresa Biberauer, Craig Sailor and
Jenneke van der Wal, as well as the many doctoral students, postdocs and others
who have given tutorials for the course. Most of all, I’d like to thank the students
themselves: you ultimately make everything happen. Thanks also to Bob Freidin
and Dalina Kallulli for comments on early drafts. And of course, thanks to the
two Helens at Cambridge University Press: Helen Barton and Helen Shannon,
whose tolerance of missed deadlines fully matches my capacity to bring them
about.
xiii
This book is an introduction to one part of the general subject of l inguistics: syn-
tax (words which are boldfaced are either technical terms or names of languages
and language families which may not be familiar; at the end of the book you can
find a glossary where these terms are defined). Syntax is the study of the struc-
ture of sentences, how these sentences arise and how speakers of a language are
able to use and understand sentences by having a mental representation of their
structure. In this book, the language we focus on is English, not because there
is anything particularly special about English, but just because that’s the one
language I can be sure you know, since you’re reading this book.
Syntax is one of the central sub-disciplines of modern linguistics. The other
central areas of linguistics are phonology (the study of sounds and sound sys-
tems) and semantics (the study of meaning). There are many other aspects to
linguistics (e.g. morphology, the study of the form of words), but these three
form the core of language: the fundamental thing about language is that it con-
nects sound and meaning through sentences. Phonology deals with the sounds,
semantics deals with the meanings and syntax deals with the sentences. Syntax
is the bridge between the sound and the meaning of sentences. We’ll have plenty
more to say about this in the chapters to follow, but first let’s look at linguistics
more broadly.
Modern linguistics is the scientific study of language. The education systems
in many parts of the world make a major distinction between sciences (like biol-
ogy, physics and chemistry) and arts or humanities (like history, music or lit-
erature). Languages are almost always classified on the arts/humanities side,
so it might seem odd to talk about looking at language and languages scientifi-
cally. ‘Science’ conjures up visions of men (mostly) in white coats carrying out
experiments using various kinds of specialised equipment, all of which seems a
far cry from reading Shakespeare, learning French verbs or trying out Spanish
conversation.
But really, looking at something scientifically means two main things. First,
it means observing things as part of nature (rocks, stars or slugs, for exam-
ple). Language is so much a part of us that it can be difficult to think of it as a
natural object, the way rocks, stars and slugs obviously are. But imagine you
were a Martian observing planet Earth: you’d see rocks, slugs, plenty of insects
and a featherless biped building machines (including spacecraft), in control of
everything, and making noises with its breathing and eating apparatus. As an
intelligent Martian, you’d soon recognise that there is some connection between
the bipeds’ noises, their ability to build machines and their control over all the
other species. You’d also notice that only the featherless bipeds make these
noises and that the young of the species spontaneously start making the noises
when they’re very small and still dependent on their parents for survival. After
abducting some bipeds and looking inside their brains, you’d realise that there’s
just a couple of areas of the biped brain which seem to control the noise-making
ability. The noises are language (Martians would probably first notice phonol-
ogy, but being super-intelligent they’d quickly realise that there was more to
language than just that). By looking at ourselves through Martian ‘eyes’ (which
might of course be infrared heat sensors attached to their feet, but never mind),
we can appreciate language as a natural object.
The second thing about science is that it’s not just about observing things;
it’s also about putting the observations together to make a theory of things. One
of the most famous of all scientific theories is Einstein’s Theory of Relativity.
This is hardly the place to go into that, but the point is that Einstein and other
physicists wanted to construct a general understanding, based on observation, of
certain aspects of nature: the Theory of Relativity is mostly about space, time
and gravity. A good theory, like relativity, makes sense of the observations we
have already made and also predicts new ones that we should be able to make.
So saying linguistics is scientific doesn’t necessarily mean that we have to put
on white coats and start a laboratory. But it does mean that we should try to make
observations about language and put those observations together to make a theory
of language. Since we’re going to focus on syntax, we’re going to make obser-
vations about syntax and make a theory of syntax. So the goal of this book is to
present a particular theory of syntax, a scientific theory of what it is that links sound
and meaning (as you can imagine, there are separate but connected theories of pho-
nology and semantics). The theory that I will introduce here is called generative
grammar. The reasons for that name will emerge over the next couple of chapters.
Approaching language, or syntax, in this way means that there are aspects
of the more traditional, humanities-style approaches to language that we leave
behind, since they do not contribute to the goal of making observations and
building theories. First, it implies that the goal of linguistics is not to set stand-
ards of ‘good’ speaking or writing. This is not to imply that such standards
should not be set, or that linguists, or others with particular expertise in lan-
guage, should not be those responsible for setting them. Instead, it means that
setting these standards falls outside the scientific study of language, since doing
this doesn’t involve making observations about what people say, write or under-
stand but involves recommending that people should speak or write in a cer-
tain way. Zoologists don’t tell slugs how to be slimy; linguists don’t tell people
how to speak or write. Ideas of ‘good’ syntax have no place here. Prescriptive
grammatical statements of a kind once common in the teaching of English in
schools in most of the English-speaking world (see below for some examples)
are similarly irrelevant to our concerns here.
species, but a defining property of it. The official biological name of our species
(in the Linnean taxonomy) is Homo sapiens, Latin for ‘wise man’. Since a
glance at a history book or a newspaper leads one to question the wisdom of our
species, one could think that Homo loquens (‘talking man’) might have been a
better term to define us.
As just mentioned, the general concept of language, roughly defined above,
should be distinguished from individual languages. An individual language
can be taken to be a specific variant of the uniquely human capacity defined
above, usually part of a given culture: English, Navajo, Warlpiri, Basque, etc.
Concepts such as ‘English’ are thus, at least in part, cultural concepts. The dis-
tinction between language (in general) and languages (individual languages as
partly cultural entities) distinguishes two variants of a single word in English.
This is rather like the way we can distinguish the general concept of cheese
from individual cheeses such as Brie, Cheddar, Parmigiano Reggiano and so on.
Language and cheese are mass nouns; they denote general concepts. Individual
cultures in different parts of the world produce their local variants of the gen-
eral thing, associated with countable versions of the same nouns, i.e. individual
languages and cheeses. In several European languages, different words are used
to make the distinction between language and languages. French distinguishes
langage (language in general) from langue (individual languages); in Italian lin-
guaggio is distinguished from lingua in a similar way, and Spanish distinguishes
lenguaje from lengua.
writing). One of the striking features of natural languages, both spoken and
signed, is that their salient structural properties are independent of their medium
of transmission: sign-language syntax, for example, has the same structural
properties as English, French and other spoken languages.
Constructed languages, or conlangs, are also relevant here. Conlangs are of
two main kinds. First, there are languages which were deliberately constructed
as ‘ideal’ languages intended to serve for more efficient communication than
could be afforded by seemingly imperfect already existing languages. Esperanto
is probably the best known, although far from the only, language of this kind.
Esperanto was invented by L. L. Zamenhof in 1887, at a time when there was
no clear international lingua franca in the Western world (since 1945, English
has played this role; at earlier times Latin and French did). The vocabulary of
Esperanto is a mixture of Romance, Germanic and Slavic elements, and the
grammatical system is a highly simplified and artificially regularised. The sys-
tem is, however, closely based on various European, principally Romance,
languages. As such it is arguably a natural language, which happens to have a
particular origin and purpose. It is debatable, however, whether Esperanto has
any true native speakers, although it is certainly used as a second language by
up to 2 million people all around the world (compare this with the estimated 2
billion second-language speakers of English). Many other invented languages
(Interlingua, Volapük, etc.) have the same status as Esperanto, although these
days they are hardly spoken at all.
The other principal kind of constructed languages are those invented for
fictional purposes, in order to give linguistic realism to invented worlds and
their denizens. Among the best-known examples are the languages invented by
J. R. R. Tolkien in his extensive mythological writings (The Lord of the Rings,
The Hobbit, The Silmarillion, etc.). Tolkien was a professional philologist, and
so the languages he constructed have an air of authenticity. Like Esperanto,
however, they are largely based on existing, mainly European, languages. They
are thus natural languages which are unusual in their origin and purpose; they
are also barely spoken at all and certainly have no native speakers. Above all for
this last reason, we leave such languages aside here; our interest is in languages
which are acquired naturally.
Two other kinds of ‘language’ should also be mentioned. First, there is ‘body
language’: frowning, shaking or nodding one’s head, crossing one’s arms or
legs, etc. Body language can certainly communicate various attitudes or emo-
tional states (friendliness, aggression, etc.), but it is not a language in the sense
that it lacks the structural properties of natural, spoken and signed, languages;
it has no discernible syntax, for example. Second, there is music. Music is often
said to be a language, and indeed it has been shown to have structural, syntac-
tic properties which are akin to, perhaps even identical to, natural language.
However, although music may have a profound emotional impact, it lacks a clear
propositional semantics, in that it cannot communicate true or false statements.
It may be that music is a cognitive capacity which shares some, but not all, of
1
See R. Robins (1967), A Short History of Linguistics from Plato to Chomsky, London: Long-
man, and V. Law (2003), The History of Linguistics in Europe, Cambridge: Cambridge Univer-
sity Press, for discussion.
2
See R. Freidin (2020), Adventures in English Syntax, Cambridge: Cambridge University Press,
for a very illuminating discussion.
We can take this line of thought further. The notion of a standard language is
really a type of idealisation, one specially constructed by language planners and
educationalists often interested in telling people what to do. In many countries,
one of the goals of teaching the standard language is teleological, in the sense dis-
cussed above. In reality, nobody really speaks the ideal standard language; every-
body speaks some form of dialect. One very clear result of the scientific study
of the social aspects of language (the field known as sociolinguistics) is that no
language is homogenous. Everybody, including people who consider themselves
speakers of ‘Standard English’, actually speaks a slightly different variant of the
language, approximating an ideal standard in slightly different ways (which, in
practice, one may hardly ever have cause to notice). As pointed out well over a
century ago by the German historical linguist Hermann Paul, ‘we must in reality
distinguish as many languages as there are individuals’. No two speakers of a
given language actually know precisely the same things about that language or
use it in precisely the same way. Alongside dialects, then, we recognise idiolects,
the variety of a language employed by a particular person. We may also want
to recognise sociolects (varieties associated with particular social classes), eth-
nolects (varieties associated with particular ethnic groups) and so on.
All of this may leave you feeling perplexed. If all speakers have different idio-
lects, is there a notion of an individual language that can really be usefully defined
at all? Atkinson (1992:23) gave the following definition of an English speaker:
The person in question has an internal system of representation … the overt
products of which (utterance production and comprehension, grammatical-
ity judgements), in conjunction with other mental capacities, are such that
that person is judged (by those deemed capable of judging) to be a speaker
of English.3
Natural Language
Leaving aside the sociocultural dimensions, then, we concentrate on
looking at natural language from a scientific perspective. More precisely, we
3
M. Atkinson (1992), Children’s Syntax, Oxford: Blackwell.
will look at language from a cognitive perspective. This approach treats lan-
guage as a form of knowledge; the central idea is that our uniquely human lan-
guage capacity is part of our cognitive make-up. Language is really a kind of
instinct humans, and no other species, are pre-programmed with (in fact, The
Language Instinct is the title of an influential and important book by Stephen
Pinker in which this approach is explained in detail; see Further Reading at the
end of this Introduction). This is the view that forms the basis of the theory of
generative grammar and so this is the view I will adopt here. This approach, and
most of the concepts introduced in this section, are due to Noam Chomsky (see
Further Reading at the end of the Introduction for more details).
The cognitive approach treats our capacity for language as an aspect of human
psychology which is ultimately rooted in biology. In this way, it clearly treats
language, and our capacity for language, as natural objects. On this view, there
are three factors in language design. The first is contributed by genetics: an
innate aptitude for language that is unique to our species, our ‘language instinct’.
The second factor is experience, particularly in early life. The language we are
exposed to as children represents the crucial linguistic experience; when I was
growing up in England, everyone around me spoke English and so I acquired
English. My great-grandfather was surrounded by Welsh speakers when he
was growing up in nineteenth-century North Wales, so he acquired Welsh.
Whichever language, or languages, we are exposed to as small children, our
innate language-learning capacity is brought to bear on this experience in such a
way as to result in our competence as native speakers of our first language (I will
say more about language acquisition below). Third, general cognitive capac-
ities, not specific to language and perhaps not specific to humans, clearly play
a role in shaping our knowledge and use of language, although the exact role
these capacities play and how they interact with (and can be distinguished from)
the language-specific aspects of our linguistic capacities are difficult questions.
Together, these three factors constitute the human language faculty. It can be
extremely difficult to distinguish the specifically linguistic aspects that contrib-
ute to forming the language faculty from more general cognitive abilities, but
the distinction can certainly be made in principle and is of course very important
for our general cognitive theory of language.
In these terms, Universal Grammar (UG) is the theory of the first factor
which makes up the language faculty: our innate genetic endowment for lan-
guage. As already mentioned, UG is assumed to play a central role in language
acquisition. It is also vital in helping us to understand how we can make sense
of the idea that specific languages are actually variants of a single type of entity,
language. Our goal here, then, is to look at one part of UG: how words combine
to form sentences. In so doing, we will construct the theory of syntax.
We can now make an important distinction. Language can be seen from an
internal, individual perspective, I-language, or it can be seen as external to the
individual, E-language. Here we are going to focus on I-language, which arises
from the interaction of the three factors just introduced. This is a natural approach,
given our interest in how syntactic structures arise in the mind. E-language is
in fact a more complicated notion, involving society, culture, history and so on.
Concepts like ‘English’ and ‘French’ in their everyday senses are E-language
concepts; it is for this reason that they do not, strictly speaking, form part of our
object of study. I-language is a natural object; E-language is not.
Taking I-language as the central notion treats language as part of individ-
ual psychology. This is a cognitive theory of language, because it intrinsically
involves the human mind. In fact, this theory of language could be part of an
overall theory of the mind. However, we would really like our theory of language
to be part of an overall theory of the brain; being an obviously physical object,
the brain is a natural object, and it is a clearer notion than ‘mind’ (in fact, philos-
ophers have worried for centuries about whether ‘mind’ is something physical,
but we don’t need to). Unfortunately, though, at present our understanding of
how brain tissue supports cognitive processes such as language, thought and
memory is extremely limited, and so we are unable to say very much about the
relation between the physical brain and cognitive processes. That’s why I will
continue to use the older, strictly speaking vaguer, term ‘mind’.
What does it mean to say our theory is a formal theory? A formal approach to
any kind of problem or phenomenon assumes discrete, systematic ways of form-
ing complex things out of simple things. For example, the alphabet is formal: its
twenty-six letters can be combined in various different, but more or less system-
atic, ways to form a very large number of words. Arithmetic is a formal system
combining numbers of various kinds and functions such as addition, subtraction,
multiplication, etc.
In formal syntax and formal semantics, we combine simple elements (roughly
words and their meanings) to form more complex elements (sentences and their
meanings). The modes of combination must be systematic and precise; as we
will see, this is a major part of the challenge of constructing such a theory.
A further natural question to ask is why we want a formal and cognitive theory.
The answer to this lies in certain general trends of thought both in psychology
and in philosophy of mind (at least in the English-speaking world). It is widely
believed that the best way to understand the mind is to think of it as a kind of
computer. Computers manipulate symbols according to formal instructions, i.e.
algorithms and programs. We can think of I-language in these terms as a piece
of cognitive software, a program run on the hardware of the brain (this raises
the intriguing question, which I will not go into here, of who or what wrote the
program). Therefore, our theory of I-language must be a formal theory. What we
are interested in is how our knowledge of I-language is represented in our mind.
If we can get an idea of this, which I believe we can from studying syntax, then
we gain a very important insight into the human mind; what it is to be human,
what it is to be you.
To conclude this general discussion, we will concentrate here on developing
one aspect of a formal, cognitive theory of I-language: the theory of syntax.
As already mentioned, syntax is concerned with how relatively simple units,
words, are combined to form more complex units, sentences. This is taken to
be a cognitive capacity all humans have, as a reflex of their genetic endowment
which includes UG. UG interacts with linguistic experience in early life, and
with domain-general ‘third factors’ as described above, to give rise to mature
adult competence in one’s native language. This competence manifests itself
in the ability to produce and understand an unlimited number of sentences, and
to make judgements regarding both the syntax and the semantics of those sen-
tences. In the next chapter, we will look in detail at a concrete example of this
competence in action, for native speakers of English.
Just a final note: given what I’ve said, strictly speaking I shouldn’t talk about
‘English’, as it is not a scientific term since it does not designate a natural object.
Instead of talking about ‘English speakers’, I should really say something like
‘individuals who identify themselves and are identified in their cultural milieu
as possessing an I-language which corresponds to the E-concept “English”’. For
brevity I will gloss over this more accurate formulation and continue to talk about
‘English speakers’; the same goes for other E-language names (Italian, etc.).
More generally, the term ‘language’ will refer to that aggregate of I-languages
whose speakers recognise themselves and each other as belonging to the same
E-language community.
Now we can move from the rather general matters we have been considering
here to actually starting out on the study of syntax. As we delve more and more
into the detailed and intricate nuts and bolts of the theory of syntax in the chapters
to follow, the background issues we have discussed here should be kept in mind
as they form the overall conceptual underpinning to the theory we’ll develop.
But now for the nuts and bolts. Or, actually, the fish.
Exercises
1. Write a short paragraph (not more than half a page as an absolute
maximum; ten–twelve lines per concept would be ideal) to explain
in your own words what the following notions mean to linguists:
• The language faculty
• Formal approaches to the study of language
• ‘Language’ vs ‘languages’
• English
Further Reading
Adger, D. 2019. Language Unlimited. Oxford: Oxford University Press.
Pinker, S. 1994. The Language Instinct. New York: Harper Perennial Modern Classics.
Roberts, I. 2017. The Wonders of Language, or How to Make Noises and Influence
People. Cambridge: Cambridge University Press.
These three books all provide general introductions to language and linguistics assum-
ing no prior knowledge on the reader’s part. Each book has its own perspective: Adger
concentrates on how syntax is fundamental to human language and cognition, and so is
perhaps most in line with our concerns here. Pinker is a classic; it is a witty and engaging
introduction to linguistics with the emphasis on the relation between language and mind.
Roberts offers a comprehensive introduction to several different subfields of linguistics,
ranging from phonetics to historical linguistics; syntax is covered in Chapter 4.
Larson, R. 2010. Grammar as Science. Cambridge, MA: MIT Press, Unit 1. This book
is an introduction to syntax whose central idea is to present the field as an exercise in the
construction of a scientific theory. In this first unit, the central ideas regarding knowledge
of language and Universal Grammar are introduced in an attractive and original way.
Carnie, A. 2013. Syntax: A Generative Introduction. Oxford: Blackwell, Chapter 1.
Another very sound and well-written introduction to generative grammar. This first
chapter presents the central ideas behind syntactic theory, along with a section on dif-
ferent approaches to syntax (something I do not attempt here). Isac, D. & C. Reiss.
2008. I-language. Oxford: Oxford University Press, Chapters 1 and 3.As the title sug-
gests, the focus of this book is on the I-language approach to linguistics and syntax.
The first chapter starts with a presentation of linguistic data in order to illustrate the
approach. The third chapter considers different notions of language, similar to what has
been presented here but with a different overall slant; reading that chapter will comple-
ment this one nicely.
Freidin, R. 2012. Syntax: Basic Concepts and Applications. Cambridge: Cambridge
University Press, Chapter 2.This chapter contains an excellent introduction to I-language,
the nature of grammaticality, native speakers’ ability to recognise deviant sentences as
an indication of knowledge of language, language acquisition and the argument from
the poverty of the stimulus, and, in particular, language production and comprehension.
1.1 Introduction
In this chapter, we begin our study of the theory of syntax. As we will
see later, there is no end to syntax. It is also rather difficult to discern where it
begins. Here we will make a start by doing two things. First, you’ll find out some
rather surprising facts about your knowledge of English, things you didn’t know
you knew. This will give you a concrete illustration of your competence in English.
Second, by carrying out a little translation exercise, we’ll get a first glimpse of what
makes languages similar to one another and what makes them differ.
We begin with very simple sentences (see (1)) and build up to rather strange
and complex ones. It’s not necessary to follow every detail of the discussion
here; all the technical ideas are presented again in later chapters. Our goal here
is to demonstrate the nature of your tacit knowledge of English syntax. We can
do it with just one word: fish. This chapter can also be skipped and returned to
later (e.g. after Chapter 7).
Like quite a few basic words in English, fish is actually ambiguous. It can be
understood either as a verb or as a noun. As a noun, it refers to a class of aquatic
animals; as a verb, it refers to the activity of hunting those aquatic animals.
We indicate this ambiguity in the standard way, by surrounding the word with
square brackets, and writing the ‘category label’ (Noun/Verb) as a subscript to
the left bracket. So, the sentence in (1) actually has two distinct representations:
1
Thanks to my good friend and colleague Professor Robert Freidin of Princeton University for
these examples. The implications of the fish sentences are discussed and explored in more detail
in Chapter 1 of Freidin (2012), . Freidin’s discussion is more detailed than here, although it
focuses exclusively on English. The exercises given there are also well worth trying.
15
The brackets are just a way of saying ‘what is inside here is a Noun/Verb’. A
representation like that in (2) is called a labelled bracketing.
The fact that fish, like many other words including cook, book, police, report,
promise and many others, is ambiguous between a verb and a noun is a conse-
quence of the fact that English has very few inflectional endings to mark gram-
matical information of various kinds. If we compare English with a more richly
inflected language such as Italian, we see that the two versions of fish correspond
to two differently inflected words:
(3) a. [Noun fish ] = Italian pesce
b. [Verb fish ] = Italian pescare
It is quite easy to see that the Italian words share the root pesc- and distinct
inflections: -e, indicating a singular noun; and -are, indicating the infinitive of
a first-conjugation verb (the infinitive is the basic form of a verb with no tense
marking; ‘first conjugation’ refers to an arbitrary morphological class of verbs
in Italian: there are four conjugations altogether, as we will see below). To cut
a very long historical story short, English has largely lost its equivalents of -e
and -are, and so we are left with the equivalent of the ambiguous root pesc- (you
may also note a family resemblance between pesc- and fish; this is because both
are ultimately derived from an ancient root in the common ancestor language of
English and Italian, Indo European).
The ambiguous single word in (1) and (2) constitutes an entire sentence on its
own. More precisely, each interpretation of fish constitutes a sentence, so really
there are two distinct sentences here. We can represent them as follows (here,
again following standard conventions, ‘Noun’ is abbreviated as N and ‘Verb’ as
V, and S stands for ‘Sentence’):
(4) a. [S [N Fish ]]
b. [S [V Fish ]]
Interpreted as in (4a), the sentence draws attention to the presence of a single fish or
group of fish (note that fish is one of a relatively small group of English nouns that
is identical in the singular and the plural: the plural form fishes exists, but it denotes
several species of fish, not several individual fish of the same species). Interpreted
as in (4b), it is an imperative, indicating an order given to go fishing. Here there
is an implicit second-person pronoun, since an order is naturally understood as
being addressed to an interlocutor, so we could elaborate (4b) as in (5):
(5) [S [N You ] [V fish ]]
The natural interpretation of this sentence is to treat the first fish as a noun and
the second one as a verb, the combination again forming a sentence. We repre-
sent this as in (7):
(7) [S [N fish ] [V fish ]]
In this sentence, the noun fish is the subject, understood as carrying out the
action performed by the verb. The verb fish indicates the action the subject car-
ries out. So the sentence means “Fish fish stuff”. As this rough gloss indicates,
there is an implicit direct object here, indicating, somewhat vaguely, what is
being fished for, what undergoes the action of fishing.
Alternatively, we can understand the first fish as a verb, and the second fish
as a noun. This is easier to see if we add an exclamation mark to (6) (Fish fish!),
corresponding to a different intonation pattern in speech. Then the sentence has
the structure in (8):
(8) [S [V fish ] [N fish ]]
Here, as in the verbal interpretation of the single-fish example in (1), the verb
is understood as an imperative and so there is a deleted second-person pronoun
(you) as subject:
(9) [S YouN [V fish ] [N fish ]]
The second fish, the noun, is an explicit direct object, indicating that fish are
what is fished.
In Italian, where the ambiguities of the English roots are clarified by inflec-
tional endings, the two interpretations of (6) can be rendered as in (10):
(10) a. I pesci pescano. (= (7))
b. Pesca pesci! (= (8/9))
Why go to this trouble to line up English and Italian in this way? There are two
reasons. First, the English and the Italian sentences mean the same thing. Although
we tend in our everyday lives to pay more attention to cultural differences than
to cognitive similarities, it seems reasonable to assume that English speakers and
Italian speakers are cognitively alike (we could perhaps say that English speakers
and Italian speakers are I-alike but E-different, thinking of the distinction made
in the previous section between I-language and E-language). In that case, the
semantic representations of the two sentences ought to be the same. It therefore
simplifies the connection (more technically, the ’mapping’) between the syntax
and the semantics if the respective syntactic representations ‘line up’, as in (11).
Secondly, we are attempting to develop a theory of Universal Grammar (UG, see
the Introduction); we therefore want our syntactic representations of comparable
(i.e. semantically matching, or at least highly similar) sentences in different lan-
guages to be as uniform as possible. Of course, it is not always possible to achieve
a parallel as straightforward as that depicted in (11), but we should strive towards
this goal as part of the UG project. Ideally, then, we would want (a) the semantic
representations of sentences which mean the same thing to be identical across
languages, (b) the syntactic representations of such sentences to be as uniform as
possible across languages while, of course, recognising (c) that their phonologi-
cal and phonetic shapes, including the linear order of the words in a sentence, are
different. The syntax and semantics are ‘externalised’ with different phonologies
in different languages, but the syntax is (near-)uniform and the semantics entirely
uniform. We will come back to these issues in Chapters 7 and 8.
The Italian sentence in (10a) tells us two more interesting things. First, the
noun pesci is plural: it has the plural ending -i (compare the singular pesce in
(3a)). In the English version, fish is plural too; we have already noted that fish
has no overt plural ending. If we substitute a more regular noun, one which
forms its plural with -s, we can see this:
(12) Boys fish.
Second, the verb pescano consists of the root pesc- and the ending -ano (again
compare the infinitive ending -are in (3b)). This is not the place to go into the
full details of Italian verbal morphology, but this ending contains the infor-
mation that the verb is first conjugation, present tense, indicative mood (it
makes a statement of fact) and third-person plural (‘they’). That’s quite a bit
of information in just three overt phonemes. The -a- part of the ending recurs
in the infinitive ending, and this can be seen as the marker of first conjuga-
tion. The -no part of the ending shows up in many other tenses and is clearly
marking third-person plural. Present tense and indicative mood are arguably
Furthermore, we can, and for the sake of consistency must, put labelled brackets
around each part of the structure:
(14) [V [V pesc- ] [Conj -a- ] [M Indicative ] [T Present ] [Agr -no ]]
The labels are fairly straightforward: ‘V’ indicates that both the root and the
whole thing form a verb; ‘Conj’ indicates the conjugation-class marker ‘M’ indi-
cates mood (the traditional grammatical category indicating whether a sentence
describes a state of affairs believed to be true or not), ‘T’ indicates tense and
‘Agr’ indicates agreement: the verb is third-person plural because the subject, i
pesci, is third-person plural.
Now, if the Italian verb has a structure like (14), and the Italian sentence in
(10a) is an accurate translation of the English sentence in (7) and our assumption
that English and Italian verbs should be minimally different means that we want
structures to ‘line up’ across languages, our logic leads us to posit something
like (15) as the structure of the verb fish in (7):
(15) [V [V fish- ] [Conj ?? ] [M Indicative ] [T Present ] [Agr 3Pl ]]
Here, Conj, M, T and Agr are all silent. But our logic leads us to conclude that
they are all structurally present. Now we come up against the limits to this lin-
ing-up-in-the-name-of-UG approach. It seems reasonable to take the verb fish in
(7) to be indicative in mood, present in tense and third-plural in agreement, but
the conjugation-marking seems to impose an idiosyncracy of Italian morphology
onto English, moreover an idiosyncracy that appears to have no semantic corre-
late. So perhaps we should eliminate ‘Conj’ from (15). The other silent endings
are, however, justified by the semantics and our universalising methodology.
Both tense and agreement endings do audibly appear on English verbs (mood
may too, but this is a little trickier and so I’ll leave it aside). If we make the sub-
ject of our example in (12) singular, but keep the present tense, we have:
(16) The boy fishes.
Here the verb has the ending -es. If we make the verb past tense, we have:
(17) The boy fished.
Here the verb has the ending -ed. So we could give the verbs in (16) and (17) the
structures in (18a) and (18b) respectively:
Here, we have dropped the Italocentric Conj, for the reasons given above; Mood
remains silent, but T is realised as -ed in the past tense and Agr is realised as -es in
the third-person singular (agreeing with the third-person singular subject the boy)
in the present tense. So there is justification for lining up these aspects of English
verbal inflection with Italian. The fundamental difference between English and
Italian has already been mentioned: in the history of English many markers of
inflection have lost their overt phonological representation, while their Italian
counterparts have retained most of theirs. So English has more silent inflections
than Italian. Semantics and UG together lead us to postulate those inflections
nonetheless.
We can also observe another interesting parallel between English and Italian
brought out in (16). Here, unlike in (12) where boys is plural, an article must
appear with boy. Compare:
(19) *Boy fishes.
The fact that where boys is plural as in (12) no audible determiner is required,
combined with the unique, known, existing ‘natural kind’ interpretation assigned
to boys here, supports the idea that there is a silent determiner. This in turn sup-
ports the lining-up of English and Italian seen in (11).
We can now replace (11) with a fuller lined-up, (nearly) uniform English and
Italian structure, in which the determiners, silent and overt, are represented by
the category D:
(21)
[S [NP [D the ] [N [N fish-[Num Pl ]]] [V [V fish- ] [M Indicative ] [T Present ]
[Agr 3Pl ]] ]
[S [NP [D I ] [N [N pesc-[Num i ]]] [V [V pesc- ] [Conj -a- ] [M Indicative ] [T Present ]
[Agr -no ]] ]
Here I’ve added the specification of number on the noun (the ending Num);
plural number is silent in the case of fish in English as we have observed. We
could add further specifications: Person on the noun (3rd here), number marking
on the Italian article i (implying, by our now-familiar reasoning, that English
the has a silent plural ending). Furthermore, the Italian articles show gender
marking: article i is masculine. So we should add that specification to the Italian
sentence. Should we add it to the English one? The Italian gender is grammati-
cal, not semantic. Pesce is a masculine noun, but not all fish are male (or there
would be no more fish, since they reproduce sexually). In Italian, all nouns have
masculine or feminine gender, and this has no semantic basis at all in many
cases: ‘sun’ is masculine (il sole); ‘moon’ is feminine (la luna), ‘sincerity’ is
The full representations of both examples, with silent endings indicated, are
given in (22):
(22) a. [S [NP you ] [VP [V [V fish- ] [M Imperative ] [T Present ] [Agr 2 ]] [[D some ]
[N [N fish-[Num Pl ] ]
b. [S [NP tu ] [V [V pesc- ] [Conj -a- ] [M Imperative ] [T Present ] [Agr 2Sg ]] ]
[NP [D i ] [N [N pesc-[Num i ]]]
Both of these are full sentences where the first ‘fish’ is the subject, the second the
verb and the third the direct object. Clearly, what is added to the two-fish exam-
ples discussed above is an explicit direct object, compared to (7), and an explicit
subject, compared to (9). Traditional grammars tell us that a complete sentence
consists of a subject, what the sentence is about, and a predicate, which says
something about the subject. The predicate contains the main verb and the direct
object. Thus we could represent (23) as follows (leaving aside for the moment
the representation of the silent inflections):
(25) [S [N Fish ] [Predicate [V fish ] [N fish ]]]
Since the verb is the most important part of the predicate, we can give a more
strictly categorial representation of (25) as follows:
(26) [S [N Fish ] [VP [V fish ] [N fish ]]]
Here VP, the Verb Phrase, is the category which functions as the predicate.2
The ‘full’ representations for (23) and (24), complete with specification of
silent inflections and other satellite categories, are shown in (27):
(27) a. [S [NP [D the ] [N [N fish-[Num Pl ]]] [VP [V [V fish - ] [M Indicative ] [T Present ]
[Agr 3Pl ]] [NP [D some ] [N [N fish-[Num Pl ]]]]
b. [S [NP [D I ] [N [N pesc-[Num i ]]] [V [V pesc- ] [Conj -a- ] [M Indicative ] [T Present ]
[Agr -no ]] [NP [D i ] [N [N pesc-[ Num i ]]]]
These representations are almost exactly the expected combination of (21) with
an explicit direct object as in (22). Again, the subject and object have differ-
ent structures and interpretations, with the subject being generic and the object
indefinite.
Now let’s try four fish:
(28) Fish fish fish fish.
This sentence is both grammatical and interpretable, but, unlike the two-fish and
three-fish examples it requires a little thought. The intonation with which it is to
be read is important: it should be ‘ish FISH fish [short pause] FISH’, with stress
on the second fish and the strongest stress on the final fish. The structure here
is significantly more complex than our earlier examples. For this reason, from
now on I will mainly give simplified representations of these more complex
examples, leaving aside the silent inflectional material (although we should not
forget that it is there).
What does (28) mean and what is its structure? Again, the Italian translation
can give us an important clue:
(29) I pesci che i pesci pescano pescano.
The markers that and che introduce a relative clause, a clause which modifies a
noun; for our purposes, the only real difference between that and che is that che
appears obligatorily in this context, while that can be ‘dropped’ (i.e. silent). So
2
We will look in detail at the relation between grammatical categories like NP and VP and gram-
matical functions like subject and predicate in Chapter 2 of Volume II.
the second two fish, the sequence noun-verb, constitute a relative clause mod-
ifying the first noun fish (known as the head of the relative). The fourth fish is
the verb of the main clause, with an implicit direct object just as in the two-fish
example with the structure in (7). We can thus make a first pass at a structure for
(28) as follows:
(31) [S [N fish ] [Relative clause fish fish ] [VP [V fish ]]]
The relative clause is a kind of subordinate clause. In fact, we can tell that the first
fish inside the relative clause is the subject and the second fish is the verb of the
predicate. So we have an NP-VP structure here (remember that the presence of the
overt definite article in Italian shows us that the subject is an NP, not just a noun):
(32) [S [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ]]] [VP [V fish ]]]
The main clause, i.e. the independent clause containing the relative clause, indi-
cated as S here, must also have a subject NP, not just N. This NP includes the
relative clause, since the relative clause modifies the noun, so we have (33):
(33) [S [NP [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ]]]] [VP [V fish ]]]
Both the main clause and the relative have a subject and a predicate. Identifying
predicates with VP again, (33) tells us that each VP contains just the verb fish. In
the case of the main clause, this may be correct, since the object here is implicit
and has a rather vague interpretation (roughly ‘stuff’). But the object of the rel-
ative has a completely different, and very precise, interpretation: ‘fish’. Here
we see an important characteristic of relative clauses: there is always something
missing, often referred to as a ‘gap’, which corresponds semantically to the head
of the relative (which we could then call the ‘filler’). Since we know that syntac-
tic representations contain silent elements (actually, we are beginning to see that
they mostly contain silent elements) and we want to keep the syntax-semantics
mapping as straightforward as possible, the best way to account for the interpreta-
tion of the gap is that it is a silent copy of the filler. So we replace (33) with (34):
(34) [S [NP [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N fish ]]]]] [VP [V fish ]]]
We said above that English that, unlike its Italian counterpart che, can be
‘dropped’. This implies that sentences with and without that are synonymous: that
and che are just meaningless syntactic markers (saying ‘what comes next is a rel-
ative clause’ in our examples). However, (30) has another interpretation that (28)
doesn’t have. In (28) the gap inside the relative clause is the direct object, as (34)
shows; it means ‘fish that get fished fish stuff’. But (30) allows a further interpre-
tation, where the gap can be interpreted as the subject of the relative clause: ‘fish
that fish other fish fish stuff’. This example has the following structure:
(35) [S [NP [N fish ] that [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N fish ]]]]] [VP [V fish ]]]
Here it is the subject of the relative that is the silent copy of the head, and
the marker that is obligatory. The Italian translation of this English sentence is
unambiguous and should be compared to the one in (29):
As we saw, the four-fish example in (28) had an implicit direct object in the main
clause (‘Fish fish fish fish stuff’). The five-fish example in (37) has an explicit
direct object in the main clause. We can therefore give the structure as in (38),
and the Italian translation in (39):
(38) [S [NP [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N fish ]]]]] [VP [V fish ]
[NP [N fish]]]]
By now we may begin to wonder just how many repetitions of the word fish (in
the phonology, the subpart of the syntactic structure that is actually pronounced)
can give rise to a grammatical, interpretable sentence. It is clear that the step
from three fish to four fish, involving as it did the introduction of a relative
clause into the structure, places a burden on comprehension. But this burden can
be readily overcome and the sentence is fully comprehensible; in fact we can
recognise that it describes a rather odd state of affairs, the very silliness of the
sentence proves that we can understand it. We see, then, that the comprehension
burden does not increase linearly as we add extra words; the structural complex-
ity of the four-fish sentence is significantly greater than the three-fish one, and
this is reflected in our initial hesitation in understanding it.
In the next section, we’ll see some really challenging examples as we increase
the number of fish still further.
Most people are completely stumped by this sentence at first sight; it appears to
be simply a random repetition of the word fish with no meaning (and so, perhaps,
no structure) at all. But in fact this isn’t true, as we’ll see below. What is striking
though is that seven fish is easier to interpret than six:
(41) Fish fish fish fish fish fish fish.
We know from (28) that a sequence of three fish can be a relative clause with
an object gap, ‘fish that are fished by other fish’. In (28), the relative clause is in
the subject position. But relative clauses can be in object position too (in fact,
(28) can be interpreted that way too, a point that was left aside above). So (41)
has the approximate structure in (42a), given in fuller detail in (42b) (where RC
stands for relative clause, and the structure is presented as a tree diagram, as this
makes it easier to see the relations among the fish; we’ll look at tree diagrams
more systematically in the next chapter):
(42) a. [ [ Fish fish fish ] [ fish [ fish fish fish ]] ]
b.
Pronounced with the right intonation (‘Fish FISH fish [pause] fish fish FISH
fish’), this example is fairly easy to understand, especially in the light of the
four- and five-fish examples. But none of this makes (40), with just six fish, any
easier to understand.
In order to see what the structure, and hence the interpretation of (40) is, we
must go back to our earlier, simpler examples. Here, once again, is the two-fish
example with noun-verb interpretation, repeated from (7):
(7) [S [N fish ] [V fish ]]
Of course, we now know that N and V here are contained in their respective
NPs and VPs. Relative clauses are really a kind of sentence, as we pointed out
above. So let us label them as S from now on. We go from two fish to four fish
by putting the structure in (7) inside the subject NP:
(43) [S [NP [N fish ] [S [N fish ] [V fish ]] ] [VP [V fish ]]]
Actually it’s not quite accurate to say we put (7) inside the subject NP; really, we
put the three-fish sentence, with the structure seen in (26), inside the subject NP
and make the object a silent version of the head of the relative:
(44) [S [NP [N fish ] [S [NP [N fish ]] [VP [V fish ] [NP [N fish ]]]]] [VP [V fish ]]]
On the basis of (44), we can add two more pronounced fish by inserting a relative
clause inside the subject NP of the relative clause, i.e. after the second [N fish ].
This gives the string of six fish in (41), and shows us what the structure is:
(45) [S [NP [N fish ] [S [NP [N fish [S [NP [N fish ] [VP [V fish ] [NP [N fish ]]]]]]] [VP [V fish
] [NP [N fish ]]]]] [VP [V fish ]]]
By now, we know a relative clause when we see one. The essential structure of
(47) has a relative clause inside the subject containing an object gap:
(48) [S [NP the mouse [S [NP the cat [VP chased [ the mouse ]]]]] [VP died ]]
(48) is quite easy to understand. The analogue to (41), but with different words, is (49):
(49) [S [NP the mouse [S [NP the cat [S [NP the dog [VP bit [NP the cat ]]]]]]] [VP chased
[NP the mouse ]]]]] [VP died ]]
Again, the jump from (48) to (49), caused by simply adding another relative
clause modifying the cat, poses severe difficulties of comprehension. Adding
some relative markers (that again) makes (49) considerably easier to understand:
(50) The mouse that the cat that the dog bit chased died.
green ideas sleep furiously; most people find this sentence unacceptable for the
perfectly good reason that it doesn’t make any sense. It is, however, syntac-
tically well-formed, as its exact formal counterpart Revolutionary new ideas
spread quickly shows. Both sentences are grammatical in that they conform to
the rules of English syntax; the first one is semantically anomalous (in fact, it
is self-contradictory) and so unacceptable. This contrast also shows that syntax
and semantics are distinct in that a sentence can be syntactically well-formed but
semantically ill-formed; we will see more examples of this type in Section 8.3.
Of course, examples with centre-embedding like (50) become still more
difficult to understand when the same phonological word occurs six times over,
as in (40). There is, in principle, no difference in grammaticality between (40)
and (50) but a real difference in acceptability (they just go from bad to worse).
But now we can line (40) up with the variant of (50) without the relative mark-
ers, and we can see how (40) works, and see that it is in fact grammatical:
(40) Fish fish fish fish fish fish.
(51) The mouse the cat the dog bit chased died.
Finally, here are the easier versions complete with the relative pronouns which
and that inserted in the appropriate places:
(50) The mouse that the cat that the dog bit chased died.
These examples have the structures shown in (49) and (45) respectively.
Relative clauses can be put one inside another without limit. They can just go
on and on. Consider (52):
(52) This is [NP1 the guy [ S1 who loved [NP2 the girl [S2 who befriended [NP3 the boy
[S3 who lived ]]]]]].
For those familiar with the Harry Potter stories, each NP here describes one
of the central protagonists by means of modification by a relative clause, each
relative clause is inside the NP it modifies, and each NP – except NP1 – is
embedded in the next relative clause. NP3 describes Harry Potter himself, the
boy who lived. NP2 describes Hermione Granger, the girl who befriended the
boy who lived, and NP1 describes Ron Weasley, the guy who loved the girl who
befriended the boy who lived. Relative clauses give language a great deal of its
expressive power, by making it possible to modify nouns using modifiers that
can be as complicated as we want: in principle, there is no upper limit to the
number of times one relative clause can be embedded inside another. This is
reflected in nursery rhymes like ‘The House that Jack Built’: the dog who bit
the cat who ate the rat who lived in the house that Jack built. Centre-embedded
relatives can also, in principle, be constructed without limit. The problem is
that, in practice, as we have seen, they quickly become very hard to understand.
Compare the following centre-embedded sequence with the excerpt from ‘The
House that Jack Built’ just given:
(53) [ The rat [ the cat [ the dog [ bit ]] [ ate ]] [ lived in the house that Jack built ]
To answer (54a), we could keep on trying longer and longer fish sentences.
There is certainly no reason to take seven, the most we have seen here (see (41)
and the structures in (42)), to be an upper limit. If we look at the examples with
even numbers of fish, (28) and (40), we see that they have the general form in
(55):
(55) a. N1 + N2 + V2 + V1 (four fish, (28))
b. N1 + N2 + N3 + V3 + V2 + V1 (six fish, (40))
(This is not their exact structure, as we have seen, but these simplified representa-
tions make our point here.) Each pair Nn + Vn, for n = n forms a relative clause
modifying Nn-1. Hence, as Freidin (2012:13) points out ‘any sentence containing
an even number of fish from four onwards will have at least one grammatical
representation’. Furthermore, although each relative clause (i.e. each pair Nn +
Vn in (55) for n > 1) has an object gap, as we saw above (see (44) and (45)), V1
does not have an object. But of course it could have, and this object could be fish.
This is what got us from four fish to five fish and from six fish to seven fish. To
quote Freidin (2012:13) again: ‘As this procedure [adding an object to V1, IR]
can be applied to any sentence with an even number (n) of fish, there will be a
corresponding sentence with an odd number (n+1) of fish … Hence any number
of fish will correspond to a sentence of English.’ In principle, then, a sentence
containing any number of repetitions of the word fish is grammatical. Our lin-
guistic competence allows us to understand such sentences, although more than
five iterations, with the exception of seven, give rise to sentences that are difficult
to understand in practice, i.e. sentences which may be unacceptable to varying
degrees. We are able to assign syntactic representations, more technically known
as structural descriptions, to the sentences, and from there recover the meaning.
The full representations of these sentences, as we saw in our discussion of the
Italian counterparts of the two- and three-fish examples, are quite complex. The
full representation of the seven-fish example in English would be (56):
(56) [S [NP [D the ] [N [N fish-[Num Pl ]]] [S [NP [D the ] [N [N fish-[Num Pl ]]] [VP [V [V
fish- ] [M Indicative ] [T Present ] [Agr 3Pl ]] [NP [D the ] [N [N fish-[Num Pl ]]]]]]
[VP [V [V fish- ] [M Indicative ] [T Present ] [Agr 3Pl ]] [NP [D some] [N [N fish-
[Num Pl ]]] [S [NP [D the ] [N [N fish-[Num Pl ]]] [VP [V [V fish- ] [M Indicative ]
[T Present ] [Agr 3Pl ]] [NP [D the ] [N [N fish-[Num Pl ]]]]]]
You don’t need to pick your way through all the details of this representation in
order to see two things. First, most of the structure is silent. Second, sentences
of this kind are highly complex objects. But remember, and this is the really
important thing, we quite unconsciously compute such complex sentence struc-
tures all the time.
Concerning question (54b), it is highly unlikely that anyone would come
across sentences of this kind (outside of a linguistics class). But, as we have
seen, they can be fairly readily understood. It is extremely difficult to see how
any account of linguistic knowledge based on simply imitating, learning and
generalising routines (perhaps through processes of stimulus, response and rein-
forcement, as in behaviourist psychology) could account for our general capac-
ity to produce and understand, mostly instantaneously, sentences we have never
heard before and which may never have been uttered before.
Question (54c) is the question that goes to the heart of the matter. As native
speakers of English, we have very rich tacit knowledge, of even the silliest and
most awkward sentences, which we were almost certainly never ‘taught’ in any
meaningful sense of the word. Our experience in early childhood, combined with
our cognitive capacities (both language-specific and not), has made it possible
for us to perform the kinds of mental computations required to extract structure
and meaning from what appear to be merely strings of repetitions of the same
word. Of course, this ability is not restricted to native speakers of English: native
speakers of Italian can just as easily make sense of sentences like (24), (29), (36)
and (39), and the same exercise can be repeated, in principle, with any speaker
of any language.
The fish sentences clearly illustrate the distinction between competence
and performance. As we have mentioned, competence refers to a state of
knowledge, knowledge of an I-language, that a native speaker of a given lan-
guage (in roughly the ‘E’-sense as discussed in the Introduction) possesses.
Performance refers to the application of I-language in a given situation: com-
bined with other cognitive and social capacities, performance makes normal
speaking and comprehension possible. Freidin’s conclusion that ‘any number
of fish will correspond to a sentence of English’ clearly concerns competence;
the rules of English syntax are such that sentences of this kind can exist, but
of course, owing to performance limitations, no actual speaker could produce
or understand a sentence consisting of an infinite number of occurrences of
the word fish and nothing else. Since placing an upper bound on the number
of times one relative clause could be embedded in another would be arbitrary,
we regard the rules of English (and other languages) as allowing for this in
principle. Of course, the same applies to numbers: no individual can ever write
out an infinite number, although our mathematical competence tells us, and
this can of course be proved, that the series of natural numbers is infinite. In
Chapters 3 and 4, we will see the precise structure-building mechanisms that
achieve this result.
Each of these questions relates I-language to one of the three factors in language
design that underlies it. In this connection, our fish sentences are very important.
They show us that syntactic structures are potentially infinite and yet built out
of very simple elements. In particular, the structures are formed by repeating
the same operations again and again. The next few chapters, as well as much of
Volume II, are devoted to demonstrating this in full detail. The fish have given
us a first inkling of the nature of the syntactic component of I-language, but now
it is time to look more systematically into the details.
Exercises
1. Consider again the four-fish example in (28), repeated here as (i):
(i) Fish fish fish fish.
We assigned the structure in (34), here (ii), to this sentence:
(ii) [S [NP [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N fish
]]]]] [VP [V fish ]]].
We also observed that inserting the relative marker that between the
first two fish made possible a different interpretation, which gave in
(35) (= (iii)):
(iii) [S [NP [N fish ] that [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N
fish ]]]]] [VP [V fish ]]].
In all of (i–iii) there is an implicit main-clause object, meaning
something vague like ‘stuff’. It is, however, possible to interpret (i)
as having an overt object. Try to see that interpretation, indicate its
structure (in a rough way, along the lines of (ii) and (iii)), its interpre-
tation and, if you can, its approximate intonation. (HINT: look again
at the ambiguity of the two-fish example discussed in (7–9)).
2. Now look again at the five-fish example from (37) (= (i)):
(i) Fish fish fish fish fish.
We said that (i) is the four-fish example plus an overt direct object
for the main-clause verb, giving the structure in (38) (= (ii)):
(ii) [S [NP [N fish ] [Relative clause [NP [N fish ]] [VP [V fish ] [NP [N fish
]]]]] [VP [V fish ] [NP [N fish]]]]
But in fact (i) has a further interpretation, where the main-clause
object is a relative clause. Give the structure for this interpretation
of (i).
Furthermore, the relative clauses in both interpretations can be
marked with that, and in each case this gives rise to a further ambi-
guity inside the relative clause. Explain the ambiguity and give the
relevant structures.
3. Give the structure for the six-fish example in (40) (= (i)):
(i) Fish fish fish fish fish fish.
Now try to add a further level of embedding to (40), using any lexical
items you like (except perhaps fish). This will correspond to an eight-
fish sentence; give the structure for this one.
Further Reading
R. Freidin. 2012. Syntax: Basic Concepts and Applications. Cambridge: Cambridge
University Press.
There are ten! = 3,628,800 possible orders for this ten-word sentence, the vast
majority of which are ungrammatical, such as the following:
(2) a. *Hopes Goona that Vurdy will be ready for his dinner.
b. *Hopes that Goona Vurdy will be ready for his dinner.
35
Our theory must make the distinction between grammatical and ungrammat-
ical strings and structures; it must tell us how and why (1) is grammatical
while the examples in (2) are ungrammatical. Furthermore it should tell us
when two sentences consisting of different words (and therefore meaning
different things) have the same structure. This is true for (1) and the exam-
ples in (3):
(3) a. Ron said that Hermione would be happy with her homework.
b. Boris thinks that Emile must be angry with his behaviour.
c. Alex believes that Eric should be worthy of a medal.
The basic notions we use to do this are those of categories and constituents,
which we now look at in turn.
2.2 Categories
2.2.1 Introduction: Lexical and Functional Categories
Like many approaches to morphology and syntax, including tradi-
tional grammar, we make a first distinction between two main types of catego-
ries, which we will call lexical categories and functional categories.
The principal lexical categories correspond to some of the ‘parts of speech’
of traditional grammar: they are noun, verb, adjective, adverb and preposition,
abbreviated standardly as N, V, Adj, Adv and P (we saw nouns and verbs in
our fish examples in Chapter 1). These lexical categories are open: people can
and do invent nouns and verbs all the time, and even the most up-to-date online
dictionaries struggle to keep up. Words belonging to these categories have clear
non-linguistic semantic content: as we saw, fish, as a noun, denotes a certain
class of aquatic animals, while fish, as a verb, denotes the activity of hunting
these animals. New adjectives and adverbs are less readily coined, but they do
arise, e.g. downloadable, googlable, etc. Prepositions are the one lexical class it
is difficult to add to; in this and certain other respects, prepositions are more sim-
ilar to functional categories. We will return below to the possibility of defining
syntactic categories semantically.
Furthermore, lexical categories do not vary greatly across languages. The dis-
tinction between nouns and verbs, in particular, seems to recur in all known
languages (although there is some debate about this in certain cases). Adjectives
are not found in every language; some indigenous languages of North America
may lack them, for example. In other languages, e.g. the West African languages
Hausa and Igbo, they form a small, closed class. In languages where the class
of adjectives is small or non-existent, the semantic work of adjectives, roughly
describing qualities, is usually done by relative clauses (a bus which reds). It
is probable that prepositions are not universal: the World Atlas of Language
Structures (WALS) lists thirty languages lacking adpositions (a cover term
for pre- and postpositions), including several indigenous languages of North
1
WALS Map/Feature 85A, https://wals.info/feature/85A#2/16.3/153.1.
theme over the coming chapters, and return to it armed with more empirical and
theoretical knowledge in Volume III.
How can we distinguish the various categories? This is quite a tricky matter,
and there are no absolute hard-and-fast diagnostic tests. For lexical categories in
particular, we can make use of four kinds of properties, relating to the main areas
of linguistic structure. Hence we can distinguish categories on the basis of their
morphology, their syntax, their phonology and aspects of their meaning. Let us
look at each of these in turn.
All of this makes it very easy to identify verbs in terms of their morphology. What
we see in Italian is typical of the Romance languages, and, with variations, com-
mon across the Indo-European languages as a whole. Some languages show case
marking on nouns, which indicates the function of the noun (or really of the NP) in
the clause: nominative case marking typically marks subjects, accusative direct
objects, dative indirect objects, etc. So in Latin dominus (‘master’) has the -us
ending for nominative singular (in this class of nouns, known as the second declen-
sion), -um for accusative singular, -o for dative singular and in fact three more cases.
In languages of this type, nouns can be identified by case marking (adjectives agree
with the nouns they modify, so they can too).
In Modern English, common adjectives can be identified by their ability to
form the comparative with the ending -er (tall – taller) and the superlative in -est
(tallest). This applies only to adjectives of two syllables or less; hence there is no
comparative beautifuller or superlative beautifullest. This constraint applies to
the root form of adjectives, so unhappier exists, since the root is the two-syllable
happy. As usual, there is a handful of irregular adjectives, e.g. good – better –
best, bad – worse – worst.
Regular adverbs are formed from the corresponding adjective by adding -ly,
e.g. beautifully, happily, badly. There are some irregular adverbs, e.g. well, and
some which simply do not add -ly, e.g. fast. Conversely, there are some adjec-
tives which end in -ly, e.g. friendly. Prepositions are invariant: they do not inflect
at all in English.
So morphological criteria can identify lexical categories in English. In most
of the Romance and the other Germanic languages, where there is typically more
inflection than in English, these criteria are correspondingly more useful and
reliable. As we saw with our fish sentences in Chapter 1, English allows many
apparently inflectionless, and hence morphologically ambiguous, forms (we
suggested that this may be due to a large amount of silent inflection in English).
In (4a), the declarative has the order Subject – Auxiliary – Verb … . In (4b),
the interrogative shows the ‘inverted’ order of subject and auxiliary. Here the
auxiliary is will, indicating (roughly) future tense. If there is no auxiliary in the
In the declarative (5a), there is no auxiliary and past tense is marked by the -ed
ending on the verb. In (5b), the auxiliary do appears, marked for past tense as
did, and precedes the subject, while the verb has no ending. Placing the inflected
verb in front of the subject is ungrammatical (indicated by the asterisk * preced-
ing the sentence):
(6) *Flooded Cambridge after all the heavy rain last spring?
For this reason, readers familiar with Shakespeare and other writers from the
seventeenth century or earlier may find that examples like (6) have a certain
Shakespearean ring to them, but they are not really part of Modern English.
Following traditional grammar, we have said that clauses are divided into a
subject and a predicate. The predicate is usually a Verb Phrase, VP. Other cate-
gories are able to function as predicates, though, providing we include the aux-
iliary be. We can see this in examples like the following, where the predicative
category is labelled in the way we saw for the fish sentences in Chapter 1 (AP
stands for Adjective Phrase, PP for Prepositional Phrase):
(8) a. John is [AP nice ].
b. John is [AP interesting ].
c. John is [PP in a bad mood ].
d. John is [NP a nice person ].
e. John is [VP sleeping ].
In (8e), the verb sleep appears in its ‘progressive’ form marked with -ing, indi-
cating an ongoing situation.
In addition to following be, predicative categories can also follow certain
verbs, such as seem.
An important distributional property of VPs is that, unlike predicative APs,
PPs or NPs, they are unable to directly follow verbs like seem, as the ungram-
maticality of (9e) shows:
(9) a. John seems [AP nice ].
b. John seems [AP interesting ].
c. John seems [PP in a bad mood ].
d. John seems [NP a nice person ].
e. *John seems [VP sleeping ].
So this test picks out VPs, since only VPs are ungrammatical following seem.
One important distributional test which picks out NPs is that only NPs can be
subjects. Thus only an NP (which may consist of just a single noun or pronoun)
can appear in the blank in (10):
(10) ___ can be a pain in the neck.
From (10), we can form the sentences in (11a) by inserting a single noun or pro-
noun (representing a whole NP), or those in (11b) where we see more complex
NPs, but not those in (11c), where the words inserted are not nouns:
(11) a. You/kids/injections/syntax/Dave can be a pain in the neck.
b. Professors of Linguistics/other people’s kids/injections which go wrong/
fish fish fish can be a pain in the neck.
c. *Walk/tall/in can be a pain in the neck.
Similarly, only VPs (which can be just a single verb) can appear between an
auxiliary and a manner adverb, i.e. in the slot in (12):
(12) Students can ___ quickly.
In (13a), we have single-verb VPs in that position and in (13b) more com-
plex VPs. In (13c) the words inserted are not VPs and so the sentences are
ungrammatical:
(13) a. Students can talk/write/learn/understand quickly.
b. Students can dissolve in sulphuric acid/get married/conclude that you’re not
worth listening to/fish fish quickly.
c. Students can *Olly/*kids/*injections/*syntax/*tall/*in quickly.
The distributional tests we have seen so far allow us to identify NPs, VPs and
auxiliaries.
Here is a distributional test for APs. Only (gradable) APs can appear in the
blank slot in (14):
(14) The students are very ____ .
In (15a) we see simple, one-adjective APs in that slot, in (15b) we have complex
APs of various kinds, while (15c) shows that other categories cannot appear there:
(15) a. The students are very intelligent/diligent/nice/eager.
b. The students are very much more intelligent than I expected/diligent in
handing in their essays/nice to talk to/eager to please.
c. The students are very *from/*walk/*Olly.
Finally, a test for PPs is that they can be identified by their characteristic intensifiers
such as straight and right. Thus only PPs can appear in the blank slot in (16):
(16) John walked straight ____ .
In (17), we have the same pattern as in (11), (13) and (15). Example (17a) gives
simple PPs (intransitive PPs, containing just a P); (17b) illustrates PPs of various
kinds, almost always with an NP object, and (17c) shows that APs, NPs and VPs
cannot appear in this slot:
(17) a. John walked straight out/on/up/in.
b. John walked straight out of the room/on to his destiny/up the hill/into the
pub.
c. John walked straight *red/*talk/*Olly.
We see that syntactic, distributional tests can isolate the main lexical categories,
NPs, PPs, VPs and APs. In an inflectionally poor language like English, these
are probably the most reliable category diagnostics available.
In (18a), increase is a verb and has stress on the second syllable (shown by the
bold letters). In (18b), increase is a noun and is stressed on the initial syllable.
Another case where phonology indicates an aspect of syntactic structure concerns
the difference between word stress and phrasal stress. Hence blackbird, with
stress on the first syllable, is a word, in fact a compound noun. This noun denotes
a particular species of bird. On the other hand, black bird, with stress on bird, is an
NP (or part of an NP; strictly speaking there should be a determiner or plural mark-
ing). This NP denotes any bird which happens to be black, following the usual rules
for attributive modification in English (so, red bus denotes any bus which happens
to be red, and so on). Thus one can say That black bird is not a blackbird without
self-contradiction, which would be impossible if the different stress patterns did not
indicate different structures (NP vs N) and therefore different meanings.
There is something to these definitions, although they are rather vague and lim-
ited. On their own, they scarcely suffice to identify categories. Take, for example,
‘what someone does’ as part of the first definition just given, or ‘action’ in the
second and third definitions (after all, actions are what people do). The word
action clearly denotes ‘action’, but it’s a noun. Similarly for occurrence, also a
noun, or ‘state of being’ (existence is a noun). Or consider the following example:
(19) The economy worries John.
Here worries is a verb, but it’s neither an action nor a ‘state of being’ (either
of the economy, or of John). Is it an occurrence? That seems a rather difficult
question to answer, but it’s not clear that answering either way would shed much
light on things. Similarly, nothing is a noun, but it doesn’t denote a thing, and it
certainly doesn’t denote a place or a person.
Despite the vagueness of these definitions, we can discern some semantic
component in category distinctions. This can be seen relatively clearly in English
where, as we have repeatedly noted by now, words can belong to different cate-
gories without changing form. For example, consider the different meanings of
round in the following examples:
(20) a. the round church
b. Round the rugged rock the ragged rascal ran.
c. These cars round corners very nicely.
d. Time for another round.
2.2.6 Conclusion
We have now seen several different ways of identifying syntactic
categories. None of them is absolutely foolproof, but taken together, they are
fairly reliable. Arguably, the distributional syntactic criteria are the most reliable
for English. In languages with richer inflectional morphology, such as Italian,
morphology can be more informative than in English.
There is little doubt that the diagnostics for categories vary from language
to language. But do the categories themselves vary? Does every language have
the same category inventory? In discussing this question at the beginning of
this section, we pointed out that the noun–verb distinction may be universal,
while adjectives and prepositions probably are not. Moreover, there appears to
be variation in the inventory of functional categories a language can have. We
might want to include Gender as a functional category, a ‘satellite’ of N, in
Italian given the presence of grammatical gender in that language, but we would
probably not include it in English, since there is no grammatical gender. This
reasoning might lead us to exclude quite a few familiar functional elements,
determiners for example, from our grammar of Chinese. These questions remain
open and, as we said above, we will revisit them in Volume III, when we have
more facts and more theory to bring to bear on them. For now, we will assume
that English has the lexical categories N, V, Adj, Adv and P and at least the func-
tional categories Aux(iliary), D(eterminer) and C(omplementiser). Furthermore,
our conception of the nature and inventory of the functional categories will be
successively revised as we go on.
Taking our cue from how we analysed (21d) in (7) of Chapter 1, we can take
each of these sentences to consist of at the very minimum a noun and a verb, and
assign them the labelled bracketings in (22):
But we have seen that nouns belong in NPs. So we can substitute more complex
NPs for the simple nouns in (22):
(23) a. [NP The fat cat ] [Verb slept ]
b. [NP The boy who lived ] [Verb smiled ]
c. [NP The owner of the famous dog] [Verb laughed]
d. [NP Fish fish fish ] [Verb fish ]
These more complex categories are Noun Phrases, NPs. NPs consist of at least
a noun, along with other words and phrases that depend on and/or modify that
noun (adjectives/APs, PPs, relative clauses, etc.) as well as satellite functional
categories such as Num and D, as we saw in Chapter 1.
Similarly, we have seen that verbs belong in VPs, so we can substitute more
complex VPs for the simple verbs in (22) and (23):
(24) a. [NP Clover ] [VP ate a mouse ].
b. [NP Kate] [VP hopes that chocolate will not get more expensive ].
c. [NP Beth ] [VP ate her dinner with a nice Chianti ].
d. [NP Fish ] [VP fish fish fish fish ].
So here we see another kind of complex category: the Verb Phrase, VP. VPs
consist of a verb and other words and phrases that depend on/modify that verb
(objects, adverbs, adverbial phrases, subordinate clauses, PPs, etc.). We will
come back to the question of the status of satellite functional categories con-
nected to VP such as T, M and Agr in Chapter 4.
NP1 VP
N1 V NP2
Clover ate
D N2
a mouse
(Here NP1, N1, NP2 and N2 are numbered simply to keep them distinct; the num-
bers have no theoretical significance.) It is important to see that tree diagrams
and labelled bracketings present exactly the same information in typograph-
ically different ways. The choice between them is a matter of convenience.
However, most people find trees easier to work with, since the relations among
the elements of the tree are more readily visible than in labelled bracketings.
But this is just a matter of personal preference; nothing theoretical depends on
it. In what follows, I will mostly use trees to represent structural descriptions.
We can now introduce some terminology that relates to the parts of the tree
diagram. S, NP, VP, etc. are the nodes of the tree; the nodes are linked by
branches. Branches never cross and all emanate from S, which is often referred
to as the ‘root’ node (in this sense the tree is in fact upside down in relation to
botanical trees, since the root is depicted as being at the top; again this is purely a
matter of convention). The words are terminal nodes; the category symbols (S,
NP, VP, etc) are non-terminal nodes. Pursuing the tree comparison, one could
think of the terminal nodes as the leaves.
The most important relations in the tree are the vertical, hierarchical ones.
The two fundamental relations are dominance and constituency. A category A
dominates another category B just where A is both higher up in the tree than B
and connected to B. We can put this more precisely as follows:
(27) A node A dominates another node B just where there is a continuous path of
branches going down the tree from node A to node B.
Let us now apply these definitions to the tree diagram in (26), which I repeat
here for convenience:
(26) S
NP1 VP
N1 V NP2
Clover ate
D N2
a mouse
From (31a) and (31c), and the fact that immediate dominance entails dominance,
we see that S dominates all the nodes in the tree. In fact, this is a good definition of
the root node: the root node dominates all the other nodes in a tree; all the nodes in a
tree are constituents of the root. Similarly, all non-terminal nodes dominate at least
one distinct non-terminal or terminal node, and terminal nodes dominate nothing.
The dominance statements in (31) are exactly equivalent to the constituency
statements in (32):
(32) a. NP1 and VP are immediate constituents of S.
b. V and NP2 are immediate constituents of VP.
c. V, NP2, D, N2 and N1 are constituents, but not immediate constituents, of S.
d. D and N2 are immediate constituents of NP2.
e. D and N2 are constituents, but not immediate constituents, of VP.
f. N1 is an immediate constituent of NP1.
g. NP1 is not a constituent of VP or of any constituent of VP.
h. VP is not a constituent of NP1 or of any constituent of NP1.
The relations of constituency and dominance are the most fundamental syntactic
relations.
2.4 Conclusion
The central goal of the theory of syntax is to account for how words
are grouped into larger units, phrases and sentences. Categories and constituents
are the building blocks for analyses. In this chapter, we have seen how to test for
and isolate various categories of English. None of these tests is foolproof, espe-
cially when taken alone, but taken together they do allow us to identify and dis-
tinguish categories. The second thing we have seen here are the ways in which
we can present the structural description of a sentence: labelled bracketings and
tree diagrams (or phrase markers). In these terms (using a tree diagram because
it is more convenient), we defined the fundamental relations of dominance and
immediate dominance, constituency and immediate constituency.
We take the notions of category and constituent to be part of UG (they are prob-
ably not primitives, but defined in terms of more abstract aspects of UG; nonethe-
less, their existence and nature are determined, perhaps indirectly, by UG). Other
languages may have categories English lacks or may lack categories English has
(although we have mentioned that the verb–noun distinction seems to hold up pretty
well everywhere). But the notion that the words of a language fall into distinct cat-
egories, isolable by some battery of tests (which, as we saw, may relate to various
other parts of the structure of language: phonology, morphology, semantics), is a
hypothesis about the nature of UG which has certainly not been disproved to date.
Similarly, other languages have constituent structures that differ from those
of English; not all tree diagrams represent universal structures. But, again, the
fundamental ideas of (immediate) dominance and constituency are a hypothe-
sis about the nature of UG, as are the proposals that branches never cross, all
branches emanate from the root node and words are terminal nodes. As hypoth-
eses about UG, these are ultimately hypotheses about genetically given aspects
of human cognition, as we mentioned in Chapter 1.
In the next chapter, we will see ways to justify the particular constituent struc-
tures we have proposed here.
Exercises
A. Categories
1. Assign the following words to a syntactic category (noun,
verb, adjective, preposition), using the tests discussed in
this chapter along with any others you know of. In your
answer, show not only the category, but also how you
apply the tests:
cat, moon, sing, fish, to, possibly, well, dispute, round
2. Now do the same with the following words: noble, enno-
ble, near, asleep and around
B. Constituents
3. Give the labelled bracketings and tree diagrams for the
following examples:
a. Boris snorted.
b. The onx splooed the blarrg.
c. Fish fish fish fish.
d. Vurdy saw Goona.
NP VP
N V PP
Ron thought
D NP
about D N
the problem
Further Reading
All of the references below discuss the ways in which we can distinguish categories,
using tests broadly similar to those discussed here and, as here, concentrating largely (but
not exclusively) on English. All of them therefore provide useful backup, and plentiful
further examples, to the ideas introduced here.
3.1 Introduction
In the previous chapter, we saw the basics of the structural descrip-
tions of sentences, the essential features of syntactic representations. Structural
descriptions can be presented as labelled bracketings, as in (1a), or as tree dia-
grams, as in (1b):
(1) a. [S [NP [N Clover ]] [VP [V ate ] [NP [D a ] [N mouse ]]]].
b. S
NP VP
N V NP
Clover ate
D N
a mouse
We also introduced basic tree terminology: node, root node, terminal node,
non-terminal node, branch, etc., and the very important mutually definable rela-
tions of (immediate) dominance and (immediate) constituency. In addition to
constituent structures, as represented in trees and labelled bracketings (i.e. struc-
tural descriptions), categories are the other fundamental notion. We illustrated
the main categories of English, as well as the very important distinction between
lexical and functional categories, and showed how categories can be fairly relia-
bly isolated by a battery of syntactic, semantic, morphological and phonological
tests (which vary somewhat from language to language).
The goals of this chapter are to start to describe (and explain) the difference
between grammatical and ungrammatical sequences. In order to do this, we
introduce the mechanisms that generate structural descriptions, i.e. the rules that
build trees and labelled bracketings. These are the Phrase-Structure rules. The
exact Phrase-Structure rules and the structural descriptions they generate can be
justified by independent tests designed to isolate and distinguish the constituents
in a tree or labelled bracketing; these are the constituency tests. We look first
at the Phrase-Structure rules and then at some constituency tests which, for the
most part, work quite well in English.
51
The question is: how do we, as speakers of English, know that these sen-
tences are ungrammatical? Clearly we don’t merely store this information
for each individual sentence or we wouldn’t have grammaticality judgements
about novel sentences such as the fish sentences we saw in Chapter 1. What
is needed is some general schema determining which are the grammatical
sentences of English (that is, of the I-language of the typical native speaker
of English).
To put it more technically, we need a mechanism to generate the well-
formed structural descriptions of English sentences. The mechanism in ques-
tion is the set of Phrase-Structure rules, PS-rules for short. PS-rules are the
formal devices which generate constituent structure, by specifying all and
only the possible ways in which categories can combine. An example of a
basic PS-rule of English (and probably many other languages) is given in
(1), where S stands for Sentence;
(3) i. S → NP VP
The brackets indicate optional categories. Strictly speaking (4ii), for exam-
ple, abbreviates VP → V, VP → V NP, VP → V PP and VP → V NP PP, but
we collapse these rules under (4ii) using the bracket notation. These rules can
generate the structural description for the sentence Clover ate a mouse, given
as a tree diagram in (1a) and a labelled bracketing in (1b), and similar ones.
Here is another example of a tree diagram (which we saw in the Exercises
sections in the previous chapter):
(5) S
NP VP
N V PP
Ron thought
P NP
about D N
the problem
This tree can be generated by the PS-rules in (6), which combines the rules
in (3) and (4):
(6) i. S → NP VP
ii. VP → V (NP) (PP)
iii. NP → (D) N (PP)
iv. PP → P NP
Each rule generates a small piece of tree, typically (but not always) a branch-
ing node and the constituents which that branching node immediately domi-
nates. As such, each rule specifies a set of immediate dominance/constituency
relations, such that the category to the left of the arrow is the label of the node
which immediately dominates the constituent(s) to the right of the arrow, and
the category or categories to the right of the arrow label the node(s) which
is/are immediate constituent(s) of the category to the left of the arrow. So,
PS-rule (6i) generates the structure in (7):
(7) S
NP VP
PS-rule (6ii) has two categories in brackets on the right of the arrow. The
brackets indicate that the categories in question are optionally immediate con-
stituents of VP (another way to say this is to say that they are optionally part
of the ‘expansion of VP’, given that PS-rules generally expand the category
to the left of the arrow by specifying more than one category on the right).
Strictly speaking, (6ii) really collapses the four distinct PS-rules seen in (8):
(8) a. VP → V
b. VP → V NP
c. VP → V PP
d. VP → V NP PP
(9) VP
V PP
Like rule (6ii), rule (6iii) collapses four rules using the bracket notation.
These are given in (11):
(11) a. NP → N
b. NP → D N
c. NP → N PP
d. NP → D N PP
b. NP
D N
The pieces of structure in (7), (9), (10) and (12a,b) combine to form the tree
in (5). In (13), (5) is repeated with each part of the structure annotated giving
the PS-rule which generates it:
(13) S
rule (6i)
NP VP
rule rule (8c) (one subcase of (6ii))
(11a) N V PP
Ron thought rule (6iv)
P NP
rule (11b) (one subcase of (6iii))
about D N
the problem
This illustrates how the structural description of the sentence is generated by
the PS-rules. The tree diagram in (13) gives the structural description of the
grammatical English sentence in (14):
In (2a) the verb and the subject are in the wrong order for English. PS-rule (6i),
S → NP VP states “Rewrite the symbol S as the sequence NP VP in that order.’
Whatever expansion of VP we choose from the options in (6ii) (see (8)), V is
always the first immediate constituent of VP. It follows from the combination of
rules (6i) and (6ii) that the subject NP (the NP in (6i)) will precede the verb. Hence
the sentences in (2) cannot be generated by our PS-rules where the immediately
postverbal NP is the subject; if it is interpreted as the object, then the subject is
missing and again the rules in (6) cannot generate the sentence. Our PS-rules are
examples of the PS-rules of English, specifying all and only the well-formed struc-
tural descriptions of English sentences, and so the sentences in (2) are ill-formed.
It is important to see that the notion ‘ill-formed’ here means ‘not generated by
the syntactic rules of English’. Our aim is to make that notion coincide as far as
possible with native speakers’ intuitive judgements regarding the grammaticality
of English sentences, which we take to reflect their I-language competence. Here
it is important to remember what we saw in our discussion of centre-embedding in
Chapter 1: native-speaker judgements reflect performance, i.e. acceptability, rather
than competence, i.e. grammaticality. As we mentioned there, though, most of the
time grammaticality and acceptability coincide, and so most of the time matching
native-speaker judgements is a good benchmark for how well our theory is doing.
As we can see from our account of the ungrammaticality of (2), PS-rules
give information about the linear (left-to-right) order of nodes, including the
words at the terminal nodes. They also give information about hierarchical
structure, in that they specify immediate dominance and constituency rela-
tions. Finally, since the symbols they use are category symbols (S, N, V, P,
etc.), they give information about the category labels of nodes. So PS-rules
specify three kinds of information about structural descriptions simultane-
ously. We will see in Volume II that these three rather distinct kinds of
information can be teased apart.
3.3 Recursion
One of the most important things our discussion of the fish sentences
in Chapter 1 showed us was that the length of natural-language sentences is in
principle unbounded. If infinitely long sentences are grammatical (but maybe
not acceptable, as they can never be ‘performed’), they should be well-formed.
In other words, they should be generated by the PS-rules. Let us see how.
In fact, the rules we already have can generate infinite structures. Consider
again rules (6iii) and (6iv):
(6) iii. NP → (D) N (PP)
iv. PP → P NP
It is easy to see that NP appears on the left of the arrow in (6iii) and on the right
of the arrow in (6iv). Given the way PS-rules work, this means NPs can appear
inside other NPs, i.e. NPs can be constituents of other NPs, as illustrated in (15):
(15) NP1
generated by (6iii)
D N1 PP
generated by (6iv)
P NP2
(Again, the subscript numbers on the Ns and NPs serve merely to keep the two
occurrences of this category distinct; they have no theoretical significance.)
Since nothing requires us to apply rules (6iii) and (6iv) in that order, after
applying both rules so as to generate the structure in (15), we can go back and
apply rule (6iii) again so as to expand NP2 identically to NP1, thereby intro-
ducing a second PP, which we can expand by rule (6iv) to give a third NP,
which we can expand by rule (6iii) to give a fourth NP, which we expand by
rule (6iv) to give a third PP, and so on. In principle, there is no limit to how
many times we can keep applying rules (6iii) and (6iv) to their own output.
The result of iterated application of these rules is complex recursive NPs of
the following kind:
(16) [NP [D the] height [PP of [NP [D the] lettering [PP on [NP [D the ] covers [PP of [NP [D the
] manuals [PP on [NP [D the ] table [PP in [NP [D the ] corner [PP of [NP [D the ] room …
The ability of rules to apply to their own output is known as recursion, a con-
cept that originates in mathematics and logic. Rules (6iii) and (6iv) are not indi-
vidually recursive, but together they form a recursive rule system. Recursion is
an extremely important concept, as it gives rise to the possibility of sentences
of unlimited length and underlies the fact that human languages are able to
make ‘infinite use of finite means’. As we have just seen, we can construct an
infinitely long NP using just rules (6iii) and (6iv). If our minds contain a recur-
sive rule system of this kind for generating sentences in our native I-language,
then our finite brains have infinite capacity. This is clearly a very important
and interesting claim about human cognition, one which justifies looking at
language from the formal and cognitive perspective adopted here.
In fact, we already saw recursion in action in the fish sentences in Chapter 1.
Consider the structural description we gave there for the seven-fish example:
(17) [S [NP [N fish ] [RC [NP [N fish]] [VP [V fish] [NP [N fish ]]]]] [VP [V fish ] [NP [N fish ] [RC
[NP [N fish]] [VP [V fish] [NP [N fish ]]]]]]]
The relative clauses introduce NPs inside other NPs and the object relative
clause introduces a VP inside the main-clause VP. Since, as we said just after
introducing the representation in (17), relative clauses are really sentences,
i.e. of category S, we should substitute S for the RC labels in (17). Then
we see that relative clauses introduce Ss inside Ss. So the reason we can
have infinite fish sentences is that the PS-rules generating relative clauses
are recursive. They feature occurrences of the symbols S, NP and VP on both
sides of the arrow, just as rules (6iii) and (6iv) do for NP and PP.
Relative clauses are one kind of subordinate clause. I won’t give the PS-rules
that generate them here as certain details of their structure remain uncertain and
controversial. Another kind of subordinate clause whose structure appears to be
much more straightforward, however, are complement clauses. A simple com-
plement clause is illustrated in (18):
(18) Goona hopes [ that Vurdy arrived].
Rule (19i) adds to the various possible expansions of VP seen in (8). We could add
‘(S′)’ to rule (6ii); this would subsume (19i) and predict that the sequence V NP S′
is possible (it is, as in persuade Mary that John arrived), the sequence V PP S′ is
possible (it is, as in say to Mary that John arrived) and the sequence V NP PP S′
is possible (this may not be correct, but I will leave this complication aside here).
In (19ii) ‘Comp’ abbreviates ‘complementiser’. Rule (19ii) states that Comp and S
are immediate constituents of the subordinate clause S′. S′ and S are not the same
category: S is a clause and constitutes the root node of a tree, while S′ is the label
of a subordinate clause introduced by rule (19i). Thanks to rule (19ii), we have a
further case of S-recursion: S appears on the left of the arrow in rule (6i) and to
the right of the arrow in rule (19ii). These rules will together generate structural
descriptions with S inside another occurrence of S.
We can expand S introduced by rule (19ii) as NP VP (in fact, this is the only
expansion of S our PS-rules allow as we have formulated them so far). We can
then expand VP using rule (19i) and introduce S again by rule (19ii) and expand
S again as NP VP, reapply rule (6i), reapply rules (19i) and (19ii), and so on
without limit. Once again we see rules applying to their own output, the basic
property of recursion. Recursive application of these rules in this way gives rise
to unlimited sequences of subordinate clauses embedded inside one another.
This is observed in English examples like (20):
(20) Mary hopes that John expects that Pete thinks that Dave said that …
As far as competence is concerned, just as with the fish sentences and the
complex NP in (16), there is no limit to the length, or to the depth of embed-
ding, of sentences like (20). The usual performance restrictions mean that
infinite sentences cannot be uttered or written down, but three PS-rules ((6i),
(19i) and (19ii)) suffice to generate them.
The structural description of (18) in tree format is given in (21):
(21) S
NP VP
generated by rule (19i)
N V S′
Goona hopes generated by rule (19ii)
Comp S
that
NP VP
Vurdy
V
arrived
There are other complementisers in English, which introduce different kinds
of subordinate clauses. As we can see in (18) and (21), that introduces a
finite, declarative subordinate clause. In (22a), whether introduces an indi-
rect question, and in (22b) for introduces an infinitival clause:
(22) a. Goona wonders [S′ whether Vurdy will arrive on time ].
b. Goona arranged [S′ for Vurdy to arrive on time ].
As indicated here, both subordinate clauses are S′s, with whether and for in C.
What follows whether in (22a) can stand alone as a complete sentence (Vurdy will
arrive on time), and so there is no difficulty with applying rule (19ii) here. The
difference with the that-complement in (18) is that the whole subordinate S′ stands
for the direct interrogative Will Vurdy arrive on time? What we observe in indirect
interrogatives is the absence of subject-aux inversion (briefly discussed in Chapter
2, see examples (3) and (4) there): in the S′ in (22a) Vurdy precedes will just as in
a declarative. We also observe that the presence of whether contributes what we
could call ‘interrogative force’, i.e. it makes the declarative into an interrogative.
In (22b), what follows the complementiser for cannot stand alone as a com-
plete sentence: *Vurdy to arrive on time. We clearly have a subject (Vurdy) and
a predicate (arrive on time) as in (18) and (22), but what is missing is tense: there
is no finite verb or auxiliary. ‘Complete’ clauses, those able to stand alone, must
have a finite verb or auxiliary to mark tense (although the tense marker may be
silent in English, as in the following complete sentences: I/you/we/they/the boys
arrive). The particle to marks non-finiteness; it shows up in almost all infinitives
in English. It looks as though we cannot generate *Vurdy to arrive on time with
rule (6i), since there is no place for to. But it seems clear Vurdy is the subject NP
and arrive on time the VP. Assuming that to is not part of the subject NP, which
seems highly unlikely, there are two possible structural descriptions for *Vurdy
to arrive on time, which are shown in the labelled bracketings in (23):
(23) a. [S [NP Vurdy] [? to ] [VP arrive on time ]]
b. [S [NP Vurdy ] [VP to arrive on time ]]
Rule (6i) cannot generate (23a), but it could generate the NP VP structure in
(23b). But now rule (6ii) (with its various expansions) cannot generate the VP
there, which has to as its first constituent. In order to accommodate infiniti-
val subordinate clauses, one of the two basic PS-rules in (6) will have to be
modified. I will leave this question open here, but we will come back to it in
Chapter 4 (see Section 4.4).
Combining (6) and (19), we now have the following set of PS-rules:
(24) i. S′ → Comp S
ii. S → NP VP
iii. VP → V (NP) (PP) (S′)
iv. NP → (D) N (PP)
v. PP → P NP
These rules can generate a large number, in fact, given their recursive nature,
an infinite number of English sentences. However, they do not generate all
the grammatical sentences English: we have seen that there is no place for the
infinitive to here, so we cannot generate (22b).
As we said at the end of the previous section, PS-rules give us three kinds of infor-
mation: (i) information about hierarchical structure (‘vertical’ information, in terms
of tree diagrams), information about linear precedence (‘horizontal’ information)
and information about the category labels of nodes. These rules are very powerful
formal devices which are capable of generating structural descriptions in a precise
way. But how do we know that structural descriptions of the kind we have been
looking at in this chapter are the right ones for English? The descriptions make very
clear claims about constituent structure, but are they correct? We have also seen at
least one example where it is not clear where to place a constituent in a structure: the
question of where to put to in infinitival clauses like (22b). Which of the structures in
(23) is the correct one, and why is the possibility that to is a constituent of subject NP
‘highly unlikely’ as I said above? In the next section we will turn to questions of this
kind, by showing how there are various ways of testing constituent structure, just as
there are various diagnostics for categories as we saw in Section 2.2.
(6i) S → NP VP
This rule applies to give us the constituent structure (25b), rather than (25c),
for (25a):
(25) a. John ate the cake.
b. [S [NP John ] [VP ate the cake ]]
c. [S [VP John ate ] [NP the cake ]]
But why do we write the rule like this? What would be wrong with writing the
rule as (6i′), which would give the constituent structure we see in (25c) for (25a)?
(6i′) *S → VP NP
One obvious answer comes from the traditional idea that clauses consist of a
subject and a predicate. But this venerable idea does not really tell us anything
about phrase structure (despite what we said in Chapter 1): phrase structure rep-
resents hierarchy, order and categories, as we have seen. It does not represent
grammatical functions or relations such as subject and predicate. From what we
have seen up to now, these notions have no place in our theory; we may wish to
build them in somehow, and we will do this in Section 5.2 (since these notions
seem to be so useful for informal discussion we might want to have a theoretical
way to understand them, but the fact remains that we have not actually said any-
thing about this so far). We will look at the question of the theoretical status of
grammatical functions in full detail in Chapter 1 of Volume II.
So, coming back to choosing between (6i) and (6i′), what this really amounts
to is choosing between the structural description in (25b) and that in (25c) for
(25a). In (25b), which is the analysis we have been assuming up to now, the
verb and the direct object form a constituent that does not include the subject,
while in (25c) the subject and the verb form a constituent that does not include
the object. The question is: why should we prefer (25b) over (25c)? To put the
question another way, what is the evidence that ate the cake is a constituent and
John ate is not a constituent? The evidence is not directly audible on the basis
of what we hear as (25a); this is because constituent structure, like most of syn-
tax, is silent. The linear order in (25a) is clearly compatible with either (25b)
or (25c). So we must find a way to determine what the hierarchical structure is.
The same question arises in relation to the NP P sequence in (26a) or the N A
sequence in (26b):
(26) a. They gave [NP the book ] [P to ] Mary.
b. They gave [NP Mary ] [AP sweet ] cookies.
Constituency tests of various kinds can show us to a large extent what the cor-
rect constituent structures are. These tests are manipulations of sentences which
are sensitive to phrasal categories such as NP, VP, etc. There are several kinds
of constituency tests. These involve: (a) manipulating the order of elements in
such a way as to show that certain sequences of words must be manipulated
together, i.e. that they constitute a phrase (these are clefting, wh-movement
and fronting); (b) substituting a sequence of words with ‘pro-forms’ of various
kinds; (c) ellipsis, deleting a sequence of words in such a way that its interpreta-
tion is recoverable from the linguistic context; (d) coordination, conjoining two
phrases of the same category with and, and (e) fragments, whether a sequence
can stand alone and be in an intuitive sense ‘complete’, even if it is not a com-
plete sentence. We will now look at each kind of test in turn. These operations
are cases of a class of rules distinct from PS-rules, transformational rules, one
type of which (wh-movement) we will focus on in Chapter 5.
Let us begin with wh-movement. Fronting a phrase containing a wh-word (i.e.
the interrogative pronouns and determiners who, what, which, etc.; see Chapter
5) also targets constituents. This operation, known as wh-movement, is a very
important one for syntactic theory, and we will introduce it fully in Chapter 5.
Again, there is a gap in the sentence where the questioned phrase was, which we
mark with a ‘t’ (and which can also be taken as a silent copy):
(27) a. Which friends does Mary hope that John will like t ?
b. What did the Party Chairman send t to John?
c. Who did the Party Chairman send a book to t ?
d. To whom did the Party Chairman send a book t ?
These examples show us that the direct objects in (32a,b) are constituents, that
the NP following the P to in (32c) is a constituent, and that the PP to whom is a
constituent.
Non-constituents, i.e. strings of words that do not form an independent, unique
phrase, bolded in (28), cannot be fronted:
(28) a. *What to did the Party Chairman send t John?
b. *What to John did the Party Chairman send t ?
Strictly speaking the ungrammaticality of (28) does not give us information about
constituency; only the successful cases of wh-movement do this. Wh-movement
does not apply to VP: the various wh-phrases correspond to different grammatical
categories: who, what are NPs (respectively animate and inanimate), which is a
D, why an AdvP or PP (‘for what reason’), when a temporal NP or PP, where a
PP, how an AP or AdvP and how (many) a measure expression. But there is no
wh-word which questions a VP. Hence, we cannot use wh-movement as a diagnos-
tic for a VP constituent. It does not follow from this that there is no VP constituent.
A further permutation we can apply to sentences in order to isolate constitu-
ents is fronting. This operation ‘highlights’ phrasal constituents by placing them
at the beginning of the sentence. The fronted phrase often functions as a topic
which ‘is commented on’ by the rest of the sentence. So, from the neutral sen-
tence (29a) we can derive (29b), where the direct object the new car is fronted:
(29) a. Mary hopes that John will like the new car.
b. The new car, Mary hopes that John will like t.
Again, we see a trace in the position where the direct object would normally be
in (29b). Fronting can apply to CP and VP, as in (30):
(30) a. That John will like her friends, Mary hopes t.
b. (Mary hoped that John would like her friends) … and [VP like her friends ]
he did t.
For (30b) to sound natural, it helps to give some context, as shown here.
Sentence (30a) may sound a little stilted, but it certainly seems acceptable. The
examples in (30) should be contrasted with (31) (where again non-constituents
are bolded):
(31) a. *A present to, the Party Chairman sent t John.
b. *A present to John, the Party Chairman sent t.
The sentence in (32a) ‘neutral’. In (32b), the sequence to John has been clefted;
in (32c) just the NP has been clefted, ‘stranding’ the preposition to. The general
schema for clefting is given in (33):
(33) S → It was XP that S.
In (34a), the NP the Party Chairman is clefted, while in (34b) it is the NP the
present. So the clefting test has isolated three constituents for us: [PP to John ],
[NP the Party Chairman ] and [NP a book ].
If we try to cleft non-constituents, bolded in (35), the result is ungrammatical:
(35) a. *It was the Party Chairman sent that t a book to John.
b. *It was a book to that the Party Chairman sent t John.
In (35a) the sequence the Party Chairman sent is clefted, and in (35b) a book
to. In both cases the result is ungrammatical. Since there could be independent
reasons why clefting some constituents is not good, the failure of a constituency
test does not really tell us anything. For example, clefting VP yields a rather odd
result (although probably not as bad as (35)):
(36) ??It was send a book to John that the Party Chairman did.
So (37) is interpreted to mean ‘John hopes that he, John, will win.’ Since John is
an NP, we should really call pronouns NPs too, since they stand for whole NPs
rather than just nouns. We can see this from the two cases of pronoun substitu-
tion in (38b) and (38c):
(38) a. [NP The man who wears glasses ] hopes that he will win.
b. * The he who wears glasses hopes that he will win.
c. He hopes that he will win.
Other categories have pro-forms too. Most VPs can be replaced by (do) so,
as in (39):
(39) John has promised Mary a book, and Bill has done so too.
Here we see the VP promised Mary a book in the first conjunct, i.e. the verb
along with both the direct and the indirect object, are replaced in the second
conjunct by done so. As with pronoun (i.e. pro-NP) substitution in (38), do so
stands for the whole VP:
(40) a. * … , and Bill has done so Mary too.
b. *... , and Bill has done so a book too.
c. *... , and Bill has done so Mary a book too.
In (40a), just the verb and the direct object (promise a present) have been sub-
stituted; in (40b), just the verb and the indirect object (promise Mary), and (40c)
just the verb promise. Again, the fact that these examples are all ungrammatical
doesn’t necessarily imply that the substituted categories are not constituents (in
fact, the verb on its own clearly is a consitituent); what it shows is that do so
substitutes an entire VP (and not, for example, V).
In the light of the conclusion that do so substitutes for a whole VP, consider
the following examples:
(41) a. John put his car in the garage on Tuesday, and Peter did so too.
b. *John put his car in the garage on Tuesday, and Peter did so on the driveway
on Wednesday.
c. *John put his car in the garage on Tuesday, and Peter did so his bike on
Wednesday.
d. John put his car in the garage on Tuesday, and Peter did so on Wednesday.
The grammaticality of (42a) also shows us that the temporal PPs are not required
for grammaticality. Such PPs are optional modifiers, or adjuncts, giving ‘extra’
information about the time the event took place, but their presence is not required
in order for the sentence to be intuitively ‘complete’. On the other hand, (42b–d)
feel ‘incomplete’: required information about what is being put (or) where is not
given. This is because the direct object and the locative PP are arguments of the
verb put; they must appear in the VP when put is V. More technically, put cat-
egorially selects for a direct-object NP and a locative PP. But it does not select
for an adjunct temporal PP.
What the ungrammaticality of (41b,c) tells us then is that selected arguments
of V must form part of the VP with the verb, and hence must undergo do so
replacement. Then (41d) can be taken to indicate that adjuncts are outside the
VP, and so do not correspond to do so replacement. But what about (41a)? The
fact that the substituted VP is interpreted as put the car in the garage on Tuesday
indicates that the adjunct PP is part of the VP. So (41d) appears to be telling us
that the adjunct PP is outside VP, and (41a) appears to be telling us it is inside
VP. We conclude for now that adjuncts (of this kind, at least) are optional con-
stituents of VP, while selected arguments are obligatory constituents of VP. This
conclusion accounts for the pattern seen in (41) as well as the examples in (42)
and is consistent with the idea that do so corresponds to VP. We will see a more
sophisticated treatment of adjuncts in the next chapter.
The do so test and the fronting test allow us to find VPs. We can now apply
these tests to our two structures for John ate the cake in (25), repeated here as (43):
With a slight tweak to the tense of the verb to create a natural context for
VP-fronting, (44a) is perfectly grammatical. Example (44b), on the other hand,
is strongly ungrammatical (bordering on what is sometimes called ‘word salad’,
a sequence of words so unintelligible that it is almost impossible to impose any
kind of structure or meaning on it).
Do so substitution, seen in (45), gives rise to a similar contrast:
(45) a. John ate the cake, and Mary did so too.
b. *John ate the cake, and did so the cake too.
Example (45b) is not word salad, but it is not acceptable. We can also notice that
Mary has completely disappeared from this example. This is a reflection of a
very general fact about substitution and ellipsis operations: substituted and elided
material is subject to a recoverability condition, i.e. the operations of substitu-
tion and ellipsis (the latter of which we will look at in depth in Volume II) must
be such that the semantic representation can find, i.e. recover, what has been sub-
stituted or elided. In (45b), Mary is not recoverable. The contrast in (45) clearly
supports the structure in (43a) over that in (43b). This is a good result because, as
we noted above, it seems that the constituents line up with the functions of sub-
ject and predicate (although our PS-rules cannot and do not state this).
As a final point on substitution, consider the following examples:
(46) a. %The Party Chairman sent it John.
b. The Party Chairman sent it.
Example (46a) is ungrammatical for most English speakers, although there are
varieties in the north-west of England, around Liverpool and Manchester, where
it is accepted (the ‘%’ is used to indicate a form which is acceptable in one vari-
ety of a language but not another; of course, this kind of variation simply reflects
the fact that the sociocultural E-language concept ‘English’ does not correspond
exactly to aggregates of I-languages – people from the relevant area of England
have slightly different I-languages from those from elsewhere, a fact reflected
in their differing grammaticality judgements of examples like (46a). However,
in these varieties it is interpreted as the direct object (a/the book); it cannot be
interpreted as substituting for a present to. Similarly, (46b) is grammatical, but it
can only be the direct object. An indirect object (i.e. to John) cannot be recovered
here. Again, then, a book to and a book to John fail the constituency test.
Ellipsis is the next kind of constituency test. Ellipsis elides, or deletes, mate-
rial, subject to the recoverability requirement we have already seen in relation to
substitution. VP-ellipsis is quite natural in English, as in:
(47) a. John can speak Mandarin and Mary can speak Mandarin too.
b. John will leave tonight and Mary will leave tonight too.
c. John has passed the exam and Mary has passed the exam too.
d. John is writing a book and Mary is writing a book too.
(48) John ate a cake and Mary did eat a cake too.
This example illustrates an aspect of the English auxiliary system that will play
a major role in our analysis of clauses from the next chapter on. The struck-out
sequence in the second conjunct here consists of the verb and its direct object,
and so we can quite reasonably analyse this as VP. This would be consistent with
our other results for constituency tests for VP: fronting and do so substitution (see
(44a) and (45a)). Example (48) differs from (47) in that there is no auxiliary in the
first conjunct and the auxiliary do appears in the second one. There is a difference
between the two conjuncts here: in the first one, the verb bears the tense marking:
we have ate, not eat. In the second conjunct, the past-tense marking is carried by
the auxiliary do, in the form of did, and so the struck-out verb form lacks tense
marking (hence eat, not ate). We see from (47) and (48) that, generally speaking,
auxiliaries do not have to be deleted under VP-ellipsis. This suggests that finite
auxiliaries are not part of VP. VP-fronting confirms this:
(49) a. (I expected John can speak Mandarin) and speak Mandarin he can t!
b. (I expected John will leave tonight) and leave tonight he will t!
c. (I expected John to pass the exam) and pass the exam he has t!
d. (I expected John was writing a book) and writing a book he is t!
This raises the question of where auxiliaries are situated. Our basic PS-rule (6i)
has nothing to say about this and will need to be revised. We defer this question
to Chapter 4.
The exact counterpart of (47), but with did also apparently deleted is gram-
matical too:
(50) John ate the cake and Mary did eat the cake/ate the cake too.
(The struck-out forms indicate that it’s unclear where the tense marking is inside
the elided sequence: on do or on ate.) The sequence in (50) is ambiguous in
that it can also mean ‘John ate both the cake and Mary too’; this latter reading
involves coordination inside the direct object of ate and no ellipsis. The reading
that the strike-out in (50) indicates is the one where Mary is the subject, not the
object, of ate. Alongside (50), we have ellipsis patterns comparable to those in
(48), but where the auxiliary is also deleted:
(51) a. John can speak Mandarin and Mary can speak Mandarin too.
b. John will leave tonight and Mary will leave tonight too.
c. John has passed the exam and Mary has passed the exam too.
d. John is writing a book and Mary is writing a book too.
These examples, along with (50), indicate that the auxiliary can form a constit-
uent with VP. It may be then that there is a mystery constituent in the clause,
indicated in (52):
(52) [S [NP John ] [?? [Aux has ] [VP passed the exam ]]]
The VP is elided in (47) and (48), and the mystery category ?? in (52). In (23a)
we saw that infinitive to may occupy a mystery category in a similar position;
there we noted that PS-rule (6i) needs modification, a point we will return to in
the next chapter.
Consider once more our alternative structures for the simple sentence John
ate the cake in (43a,b):
(43) a. [S [NP John ] [VP ate the cake ]]
b. [S [VP John ate ] [NP the cake ]]
VP-ellipsis clearly favours (43a) here, just like do so substitution and VP-fronting,
as we have seen. Applying VP-ellipsis to (43b) yields:
(53) John ate the cake and Mary eat did the cake.
(53) is not exactly acceptable, but it is difficult to work out what the second
conjunct means. Moreover, Mary is unrecoverable in (53). Hence we must take
Aside from being rather odd and redundant, here again Mary is unrecoverable.
Once again, (43b) seems to give the wrong results, while (43a) makes correct pre-
dictions. We conclude that (43a) is the correct structure. Thus the traditional func-
tional division of the clause into subject and predicate has a structural correlate.
Coordination can also function as a constituency test. By and large, only con-
stituents of the same type can be coordinated. Thus we can observe the follow-
ing(examples from Radford 2016:134):1
(55) a. The chairman has resigned from the board and the company.
b. The chairman has resigned from the board and from the company.
c. The chairman has resigned from the board and gone abroad.
d. The chairman has resigned from the board and is living in Ruritania.
e. *The chairman has resigned from the board and company has replaced him.
f. The chairman has resigned from the board and the company has replaced him.
In (55a), we see coordination of the NPs the board and the company; in (55b)
we have coordination of the PPs from the board and from the company; in (55c),
we have VP-coordination of resigned from the board and gone abroad. Example
(55d) confirms our suggestion above of a mystery constituent combining the aux-
iliary and the VP, since two occurrences of this constituent appear to be coor-
dinated: has resigned from the board and is living in Ruritania. Example (55e)
is unacceptable, and so we do not have evidence that the sequences chairman
resigned from the board and company has replaced him are constituents; in fact
each seems to consist of the N of the subject NP (chairman, company) and our
mystery category. This confirms the suggested structure in (52) for S where the
mystery category is a constituent of S and not of the subject. The structure for
The chairman has resigned from the board, on analogy with (52), would be (56):
(56) [S [NP [D The ] [N chairman]] [?? [Aux has ] [VP resigned from the board ]]]
Here we clearly see that the sequence chairman has resigned from the board is
not a constituent. This is consistent with the fact that the coordination test yields
an ungrammatical result when we attempt to combine strings of this type as in
(55e). Finally, (55f) shows us that S is a constituent.
The final kind of constituency test relates to fragments. These are most naturally
found in answers to questions. Generally speaking, only phrasal constituents form
grammatical fragments. So, for example, the following answers are ungrammati-
cal (here again, I borrow, and slightly adapt, examples from Radford 2016:137–8):
1
A. Radford (2016), Analysing English Sentences, Cambridge: Cambridge University Press.
The plot thickens around our mystery category given the ungrammaticality of
the response in (61), however:
(61) A: What has the chairman done?
B: *Has resigned from the board.
Recall that the failure of a constituency test does not show us that a given
sequence is not a constituent; there may be other reasons certain constituents
cannot pass a given constituency test. We have seen evidence in (50), (51) and
(55e) that the mystery category, the sequence Aux + VP, is a constituent. Some
independent factor must then be responsible for the failure of the fragment test
in (61). If fragments can only be phrasal and the mystery category is not phrasal,
that could be the solution; again, we return to this in the next chapter.
The various constituency tests isolate NP, PP, S, VP and the mystery cate-
gory, as we have seen, although not all of them isolate every category. What
about APs? APs come in two different syntactic and semantic contexts, attribu-
tive (modifying a noun, as in sweet cookies) and predicative (constituting, usu-
ally with the help of be, a predicate: the cookies are sweet). Here the simple
adjective sweet forms an AP on its own (just as John and he form NPs on their
own). Let us see how the different APs fare under our constituency tests, begin-
ning with predicative APs:
(62) Permutation:
a. ??It’s sweet that the cookies are t. (clefting)
b. How sweet are the cookies t? (wh-movement)
c. ?(I expected those cookies to be sweet), and sweet, they really are!
(fronting)
(63) Pro-form:
John likes his cookies sweet → John likes his cookies (just) so.
(64) Ellipsis:
The cookies are sweet and the cake is sweet too.
(65) Coordination:
The cookies are sweet and extremely fattening.
(66) Fragment:
A: How are the cookies?
B: Sweet/very sweet/*very.
(68) Pro-form:
John likes his sweet cookies → *John likes his so cookies.
(69) Ellipsis:
*John ate the sweet cookies and the sweet cake.
(70) Coordination:
John ate the sweet and extremely fattening cookies.
(71) Fragment:
A: What kind of cookies did John eat?
B: Sweet ones/*sweet.
Here the only test that works is coordination. In (67), we can interpret the AP
as predicative, with the examples deriving from John ate the cookies sweet, but
this, as we have seen, is a different syntactic context. For this reason, I brack-
eted the NP to indicate that sweet here corresponds to the prenominal position
inside NP. Understood this way, these examples are clearly ungrammatical.
Similarly, pro-form so and ellipsis are impossible in the NP-internal position.
In (69), the sequence John ate the sweet cookies and the cake is of course gram-
matical, but sweet cannot be recovered as a modifier for cake and so the ellipsis
indicated is ungrammatical. In (71), sweet seems to need a nominal pro-form
ones for grammaticality. Nevertheless, the fact that coordination yields a gram-
matical result supports the idea that adnominal APs are constituents.
Finally, let’s compare the two possible structures for (26b), repeated here:
Here are the results of the tests applied to Mary sweet in (72a):
(73) Permutation:
a. *It’s Mary sweet that they gave t cookies. (clefting)
b. *What/how/who did they give t cookies? (wh-movement)
c. *Mary sweet, they gave t cookies. (fronting)
(74) Pro-form:
John gave Mary sweet cookies → *John gave her so cookies.
(75) Ellipsis:
*John gave Mary sweet cookies and Bill gave her sweet cake.
(76) Coordination:
?John gave Mary sweet and Bill extremely fattening cookies.
(77) Fragment:
A: Who did John give what kind of cookies?
B: *Mary sweet.
The only test which does not yield an ungrammatical result is coordination.
However, the (slightly marginal) structure is a special kind of coordinate
structure known as ‘right node raising’, involving permuting cookies to the
right and deleting a copy of this element. So the structure of (76) is actually
something like:
(78) ?John gave Mary sweet cookies and Bill extremely fattening cookies.
What this shows is that cookies is a constituent (a noun), but nothing about
the sequence Mary sweet. So no tests reveal Mary sweet to be a constituent as
in (72a). Combined with the absence of any semantic or functional motiva-
tion for grouping these words together, in that sweet does not modify Mary,
we may conclude that the structure in (72a) is mistaken.
Now let us apply our tests to sweet cookies in (72b):
(79) Permutation:
a. It’s sweet cookies that they gave Mary t. (clefting)
b. What kind of cookies did they give Mary t? (wh-movement)
c. Sweet cookies, they gave Mary t. (fronting)
(80) Pro-form:
John gave Mary sweet cookies → *John gave her them.
(81) Ellipsis:
*John gave Mary sweet cookies and Bill gave Mary sweet cookies too.
(82) Coordination:
John gave Mary sweet cookies and Bill gave Mary extremely fattening cakes.
(83) Fragment:
A: What did John give Mary?
B: Sweet cookies.
Here all the tests work straightforwardly except ellipsis. English does not
allow ellipsis of argument NPs. We can see this from the following contrast:
(84) a. John brought his friend but Bill didn’t bring his friend.
b. *John brought his friend but Bill didn’t bring his friend.
VP, but not NP, elides in English. The opposite is true in Japanese:
(85) a. *Kai-ga piza-o tabeta, Lina-mo piza-o tabe sita. [Japanese]
Kai-nom pizza-acc eat.pst, Lina-also pizza. acc eat do.pst
b. Taroo-wa zibun-no tomodati-o turetekita, demo Hanako-wa zibun-no
tomodati-o
Taroo.top self.gen friend.acc bring.pst, but Hanako.top self.gen
friend.acc
turetekona katta.
bring not.pst
‘Taroo brought his friend, but Hanako did not bring her friend.’
This sentence has two meanings, depending on who was wearing the pyja-
mas. The natural reading is that I was wearing my pyjamas (perhaps because
the elephant broke into my house one night). The other reading is that the
elephant was wearing the pyjamas (this was the one Groucho intended as he
went on ‘I’ve no idea why he was wearing them’).
The two interpretations of the sentence correspond to the following structural
descriptions:
(87) a. I shot [NP an elephant [PP in my pyjamas ]].
b. I shot [NP an elephant] [PP in my pyjamas ].
(89) Pro-form:
i. I shot it. (Groucho/vague)
ii. I shot it in my pyjamas. (normal)
(90) (VP-)ellipsis:
i. I shot an elephant in my pyjamas and John did shoot an elephant in
my pyjamas too. (Groucho)
ii. I shot an elephant in my pyjamas and John did shoot an elephant in
my pyjamas too. (John)
(91) Coordination:
i. I shot an elephant in my pyjamas and a tiger in my favourite teeshirt.
(ambiguous)
ii. I shot an elephant and a tiger in my pyjamas. (ambiguous)
(92) Fragment:
A: What did you shoot?
B: i. An elephant in my pyjamas. (Groucho)
ii. An elephant. (normal/vague)
The permutation tests, especially clefting in (88a) and fronting in (88c), clearly
distinguish the two structures in (87) and each reading is consistent: where an
elephant in my pyjamas is a constituent, we have Groucho’s interpretation, where
it is not, we have the ‘normal’ one. Example (88b,i) is slightly equivocal, in that
in my pyjamas may be completely unspecified, hence the designation ‘vague’.
However, if forced to choose between the two readings in (87) for (88b,i), the
Groucho one is clearly preferred. Similar considerations apply to (89i): in my
pyjamas may be left unspecified. But (89ii) allows only the ‘normal’ reading.
VP-ellipsis gives an interesting result in (90ii), as the interpretation is that John
was wearing my pyjamas. This is consistent with the structure in (87b), where
pyjamas modifies the subject. The fragment test in (92) gives a completely
unambiguous result in (92i), the Groucho reading, but, again, the possibility of a
vague reading in (92ii). Leaving aside the coordination result for a moment, all
of these tests, with one or two provisos, disambiguate the structural ambiguity
of (86) and reveal that the ambiguity is attributable to the different structural
descriptions, associated in turn with different semantic interpretations, in (87).
The coordination examples in (91) are both ambiguous, however. This is due
to the ambiguity of coordination more generally. The bracketings given in (93)
illustrate the ambiguities:
(93) i. I shot [ an elephant [ in my pyjamas ]] and [ a tiger [ in my favourite teesh-
irt]]. (Groucho)
I shot [ an elephant ] [ in my pyjamas ] and [ a tiger ] [ in my favourite teesh-
irt]. (normal)
ii. I shot [[ an elephant and a tiger ] [ in my pyjamas ]]. (Groucho)
I shot [ an elephant and a tiger ] [ in my pyjamas ]. (normal)
The first example in (93i) coordinates two direct objects each with the internal
structure seen in (87a), and so unambiguously yields the Groucho interpretation.
The second example coordinates a constituent consisting of the sequence NP
PP, where PP is not a constituent of NP and so does not modify NP. This constit-
uent is probably a VP, where the verb in the second conjunct has been ‘gapped’
(a form of ellipsis affecting just V, not VP, as in I spoke to Mary about John
and to Bill about Pete). In the first example of (92ii), both the elephant and the
tiger were wearing my rather capacious pyjamas; clearly a case of the Groucho
reading which is rather more absurd than even he intended. The second exam-
ple in (92ii) is a straightforward case of the normal reading. So the ambiguities
in (91) result from ambiguities regarding exactly which constituents are being
coordinated, disambiguated by the bracketings in (93). Finally, the bracketing in
(94) gives a further example of the ambiguities of coordination:
(94) I shot [ an elephant ] and [ a tiger [ in my pyjamas ]].
Example (94) is clearly interpreted to mean that the tiger, but not the ele-
phant, was wearing my pyjamas. This reading arises from the fact that in my
pyjamas modifies tiger here, since it is a constituent of the same NP. Once
again, we see that constituency tests can probe for structural ambiguity and
reveal that ambiguity to be a consequence of the different structural descrip-
tions, each unambiguously associated with different semantic interpretations.
In this section we have introduced, explained and applied the principal kinds
of constituency tests. Two important points should always be borne in mind in
connection with these tests. First, only a positive result is a real result: if a test
for constituent X fails, it does not show that X is not a constituent; in fact, it does
not show anything. Second, one aspect of syntactic variation across languages
involves the range of permutation operations, pro-forms, ellipses, coordinations
and even fragments a given language allows. So what may work as a good test
in one language (e.g. VP-ellipsis in English) may not work in another language,
as we have just seen in relation to VP-ellipsis in Japanese.
Exercises
1. Phrase-Structure rules
Take the following set of PS-rules:
i. S → NP VP
ii. VP → V (NP) (PP)
iii. NP → (D) N (PP)
iv. PP → P NP
Give five structurally distinct, grammatical English sentences that
these rules generate, showing their structure as a tree diagram.
2. Give the tree diagrams for the following sentences:
a. Goona hoped that Vurdy would eat the curry.
b. Harry thought that Ron said that Hermione knew the answer.
c. Pep wanted to know how much money Paul would want.
d. Andy wondered whether Lou would write a song.
e. Ron wondered whether to talk to Hermione.
3. Using the constituency tests discussed in this chapter (permu-
tation, pro-forms, ellipsis, coordination, fragments), justify the
constituent structure you have assigned to one of the sentences in
(1) and (2).
4. The following sentences are structurally ambiguous. Explain the
ambiguity, assign distinct structural representations (i.e. draw a tree
or give labelled brackets) corresponding to each reading of the sen-
tence, and show how constituency tests support the structures you
propose in each case.
a. Mary ate the cake on the shelf.
b. Olly dreamt that his team won last night.
c. The chickens are ready to eat.
Further Reading
All of the readings given here introduce the notions of category, constituency and con-
stituency tests, with some variations in emphasis and details. Sportiche et al. (2014) are
particularly thorough and Freidin goes into interesting detail on coordination.
Carnie, A. 2010. Constituent Structure. Oxford: Oxford University Press, Chapter 2.
Carnie, A. 2013. Syntax: A Generative Introduction. Oxford: Blackwell, Chapters 3 and 4.
Freidin, R. 2012. Syntax: Basic Concepts and Applications. Cambridge: Cambridge
University Press, Chapter 3.
Larson, R. 2010. Grammar as Science. Cambridge, MA: MIT Press, Unit 7.
Sportiche, D., H. Koopman & E. Stabler. (2014). An Introduction to Syntactic Analysis
and Theory. Oxford: Wiley, Chapter 3.
Tallerman, M. 1998. Understanding Syntax, London: Routledge, Chapter 5.
4.1 Introduction
In the previous chapter we saw how PS-rules provide structural
descriptions of sentences, in the form of labelled bracketings or tree diagrams,
in terms of the key notions of categories and constituents. PS-rules specify the
grammatical structures (the well-formed structural descriptions) of a given gram-
mar, corresponding as far as possible to the I-language grammar of competent
native speakers. PS-rules, or systems of PS-rules, can be recursive, in that they
apply to their own output; they thereby specify an infinite set of well-formed
structural descriptions. We saw several illustrations of this is in the previous
chapter and in Chapter 1 with the fish examples. Constituency tests of various
kinds show, on the basis of native-speaker well-formedness judgements, which
structural descriptions are correct. Thus these tests can support the postulation
of a given set of particular PS-rules. In this chapter we will revise and simplify
the system of PS-rules.
In (2a–d), we see the respective structural descriptions rules that (1a–d) gener-
ate, presented as labelled bracketings. Each one corresponds to the full expan-
sion of each rule, i.e. with optional material realised:
(2) a. [NP [D the ] [N picture ] [PP [P of ] [NP [N John ]]]]
b. [AP [Mod very ] [A angry ] [PP [P with ] [NP [N John ]]]]
c. [PP [Mod just ] [P beyond ] [NP [D the ] [N frontier ]]]
d. [VP [AdvP really ] [V enjoy ] [NP [D the ] [N movie ]]]
78
The examples in (2) show us that the PS-rules in (1) generate a series of well-
formed structural descriptions and as such are PS-rules in good standing. In each
case there is a lexical category (N, V, A or P) on the right of the arrow corre-
sponding to the phrasal category on the left of the arrow.
Now compare the rules in (3):
(3) i. NP → V AP
ii. PP → N V
iii. AP → V NP
None of these rules has a lexical category on the right of the arrow correspond-
ing to the phrasal category on the left. These rules are certainly not rules of
English: there is no NP constituent which consists of a verb and an AP, no PP
which contains the sequence N V and no AP consisting of V and NP. The strings
found to the right of the arrows do exist: look nice is a case of V AP; Clover slept
is the string N V and any transitive verb followed by its object is a case of V NP
order (eat a mouse, like cake, visit London, etc.). But look nice is not an NP, as
we can see if we apply the basic distributional test for NPs given in Chapter 2.
That test consisted in seeing whether a putative NP could appear in subject posi-
tion, i.e. in the dash position in (4):
(5) shows us that the result of inserting look nice in that position is
ungrammatical:
We can safely conclude that, while the PS-rules in (2) certainly do gener-
ate well-formed structural descriptions for English, those in (3) do not.
Languages differ somewhat from one another regarding their PS-rules, but
we might reasonably ask whether any language would counter structures of
the kind generated by the rules in (3). Since each PS-rule specifies a relation
of immediate dominance/constituency, the rules in (3) generate trees of the
kind shown in (8):
(8) i. NP
V AP
ii. PP
N V
iii. AP
V NP
Structures of this kind have never been postulated, and it seems highly unlikely
that they ever will be. If it is true that these structures are part of the grammar of
no natural language, then our theory of UG must rule them out. That means that
the PS-rules in (3) should be ruled out of our theory. The theory of PS-rules, as
we presented it in the previous chapter, does not do this. Hence we must develop
a way of doing this.
(3) i. NP → V AP
ii. PP → N V
iii. AP → V NP
The main difference between the rules in (1) and the rules in (3), as we observed
in the previous section, is that those in (1) have ‘different versions’ of the same
category on each side of the arrow, while that is not true of those in (3). In fact,
the rules in (1) have phrasal categories on the left of the arrow and word-level
versions of the same categories on the right of it. Not only that, but everything
else in the expansion of the respective categories is optional. So the rules in (1)
could be written in a more schematic form as:
(9) a. NP → … N …
b. AP →… A …
c. PP →… P…
d. VP →… V …
So one generalisation about the well-formed PS-rules in (1) is that they all consist
of a phrasal category to the left of the arrow and at least a single-word version of
that category to the right. Call the single-word version of the phrasal category the
head of the phrase, or just the head. So, the generalisation is that whatever else
appears to the right of the arrow, the head of the phrase that appears to the left must
appear there. To put it another way, every phrase must have a head. The technical
term for ‘having a head’ is endocentric (as opposed to exocentric, not having a
head). So every phrase generated by the well-formed PS-rules has a head. This is
the endocentricity requirement: virtually all phrasal categories are headed.
Now we can see what is wrong with the PS-rules in (3): they all generate exo-
centric categories. None of the rules in (3) meets the endocentricity requirement,
while all of those in (1) do. We can generalise the endocentricity requirement by
generalising over the partial rules in (9) as follows:
(11) XP → … X …
Here ‘X’ is a categorial variable ranging over the syntactic categories (just as in
algebra x is a variable ranging over numbers). Since all the good PS-rules in (1)
conform to (11) and none of the crazy ones in (3) do, let us make (11) a general
requirement on PS-rules. So all PS-rules must conform to the template in (11).
This significantly limits the number and the form of PS-rules. So we have now
successfully eliminated the PS-rules in (3).
Looking again at the well-formed PS-rules in (1), we can make a further gen-
eralisation. In addition to requiring a head of the same category as the category
on the left of the arrow, all the rules allow optional material both to the left and
to the right of the head. In the case of AP, PP and VP, the material to the left of
the head is always modificational and, as such, optional. In the case of NP, it is
D(eterminer), which we will take for now to also be a kind of modifier although
we know that Ds are sometimes obligatory (e.g. singular count nouns cannot
stand alone in subject or object position: recall *Boy fish from Section 1.4).
Turning to the material following heads in (1), we see that N and A are followed
by PP, while V and P are followed by NP. In (2), we exemplified these as follows:
(12) a. The picture of John
b. very angry with John
c. just beyond the frontier
d. really enjoy the movie
In (12c) and (12d) the NPs are clearly objects of P and V respectively. We can
maintain the same for N and A in (12a,b), noting that these objects are obligato-
rily prepositional, i.e. they must be introduced by a Preposition (of, with).
(Strictly speaking, verbs and other categories c-select the head of their comple-
ment, but the endocentricity requirement has the effect that c-selection introduces
phrasal categories corresponding to the selected head.) On the left in (14a–d),
we see examples of the lexical entries of verbs. The lexicon is the mental dic-
tionary (stored in the long-term memory of every competent native speaker of
an I-language) containing idiosyncratic information about all the words in the
I-language (i.e. all the words a given speaker knows). We will say more about
selection and the lexicon in Chapter 8 and in Volume II; for now, the simplified
lexical entries given in (14) suffice.
The lexical entry on the left of (14a) says that watch is a verb that c-selects an
NP; the ‘__NP′ notation means the verb must appear in the position immediately
before NP. On the right in (14a), we have the PS-rule which expands VP so as
to create the structure required by the c-selection property of watch. This rule
generates NP as the complement of V. Clearly, watch here represents the very
large class of basic transitive verbs, verbs which take just a direct object. Having
a lexical entry like this (with further information about the phonological and
semantic properties of the word) stored in long-term memory is a part of what it
means for a native speaker to ‘know’ the word.
The lexical entry in (14b) states that rely c-selects a PP, which in fact must
be headed by on: rely on/*with/*for/*up John. The PS-rule generating the corre-
sponding expansion of VP, needed to accommodate this requirement, generates a
PP in the complement of V. The lexical entry in (14c) is the one for put, which we
briefly saw in Chapter 3. Here VP needs both an NP and a PP position, since put
c-selects both; see the discussion around (42) in Chapter 3. Finally, in (14d) we
have the lexical entry for say, which specifies that this verb c-selects S′: say [S′
that the world is round ]. The corresponding expansion of VP is given on the right.
Putting all the PS-rules on the right of (14) together, we come to something fairly
close to the expansion of VP given in (24iii) of Chapter 3.
In (15), we see the first version of the full ‘Xʹ-schema’ for PS-rules (pro-
nounced ‘X-bar’):
(15) a. XP → Specifier X′
b. X′ → X complement
specifier X′
X complement
In (16) we can readily see how the X′-level creates different levels of syntactic
‘closeness’ for the complement and the Specifier in relation to the head. This
relative ‘closeness’ is structurally represented. The complement is closer to the
head, since it is obligatory and its category depends on the lexical c-selection
properties of the head, as illustrated for various verbs in (14). The Specifier is
usually (but not always, as we have seen) optional and modifies or ‘determines’
the head in various ways. The structure in (16) represents the greater relative
closeness, both syntactic and semantic, of the complement to the head.
The terms ‘Specifier’ and ‘complement’, although more abstract than ‘sub-
ject’, ‘object’, etc., are equally relational. We have repeatedly stated that
PS-rules do not (directly) give relational information. These relational terms
should be replaced with categories. With the apparent exception of S′, all
the complements in (14) are phrasal; in X′-theory terms, they are all XPs.
To distinguish them categorially from the head X that c-selects them, let us
call them YP (where Y is just another, distinct categorial variable). We will
return to the status of S′ below. The Specifier elements, with the exception
of D, are all phrasal modifiers. For uniformity, let us also then treat D as
phrasal, DP. Then Specifiers are always phrasal. To distinguish them from
the X′-category they specify and from the YP complement to X, call them
ZP (Z is yet another distinct categorial variable). So the correct version of the
X′-schema for PS-rules is (17):
(17) a. XP → ZP X′
b. X′ → X YP
The tree diagram representing the structural description (or the template for
structural descriptions, where X, Y and Z are all categorial variables) generated
by (17) is given in (18):
(18) XP
ZP X′
X YP
Y and Z are further categorial variables, substitutable with different values from
X. So YP and ZP are each expanded as in (17) and so have the internal structure
seen in (18). Clearly, then, the X′-system is fully recursive.
We have posited a new kind of constituent in (17) and (18): the interme-
diate level category X′. So far, we have only given a functional justification
for X′, saying that it allows us to distinguish the relative closeness of comple-
ments to ‘their’ heads, as opposed to Specifiers. But we must be able to prove
its presence using constituency tests of the kind introduced in Section 3.4.
The coordination test clearly isolates X′, in that X′-level constituents can be
coordinated:
(19) a. the [ pictures of John ] and [ books about Mary ]
b. just [ beyond the frontier ] and [ down the road ]
c. very [ angry with John ] and [ worried about Mary ]
d. really [ enjoy the movie ] and [ dislike the book ]
In (19a), the modifies the head of both conjuncts: books about Mary can be
interpreted as definite despite the fact that the string books about Mary can stand
alone as a plural indefinite NP (with a silent determiner). Similarly, in (19b–d)
Mod can apply to both conjuncts in each case.
The permutation tests do not give good results, however:
(20) a. *It’s pictures of John that we saw the t. (clefting)
b. *Which pictures of John did you see the t? (wh-movement)
c. *Pictures of John, we saw the t. (fronting)
The pro-form test does give a good result, using one as the pro-form, as the fol-
lowing examples illustrate:
(21) a. John read a book of poems and Mary read one too.
b. *John read a book of poems and Mary read one of stories.
c. John read a book with a red cover and Mary read one too.
d. John read a book with a red cover and Mary read one with a blue cover.
In (21a), one appears to stand for the whole NP the book of poems. The unac-
ceptability of (21b) shows that one must stand for the complement of book, the
PP of stories. In (21c), one once again appears to stand for the whole NP the
book with a red cover. But in (21d) we see that with a blue cover is not obligato-
rily part of the constituent one stands for. Here we see a pattern similar to the one
we saw with the examples in (41) in Chapter 3 the with-phrase, which is not a
complement of book but an optional modifier, does not have to be substituted by
the pro-form one. On the other hand, the of-phrases in (21a,b) are complements
of book (they are not obligatory, but few complements to nouns are fully oblig-
atory) and, as the unacceptability of (21c) shows, must be part of the constituent
one stands for. Since the occupies the Specifier of N′ (abbreviated SpecN′), and
the complement must be part of what one stands for along with the head, we
conclude that one is an N′-proform. We will come back to the question of the
with-modifiers in Section 4.5 below; clearly they cannot be in the complement
of N. So the structure of the NP one with a blue cover in (21d) is as in (22):
(22) NP
N′
N PP
one with a blue cover
Other determiners do not allow N′-ellipsis, either requiring the pro-form one, or
not allowing ellipsis at all. We do not need to go into the details of this here; it
suffices to note, once again, that the conditions under which ellipsis is possible
are hard to define.
Finally, N′-fragments are not allowed:
(24) A: What did you read?
B: *Book of poems.
We conclude that the coordination, pro-form and ellipsis tests reveal that N′is a
constituent.
Concerning the other X′ categories, neither P′ nor A′ can be permuted, like N′:
(25) a. *On the nose, she hit him right t. (Radford 2016:126)1
b. *Proud of him, she certainly seems to be very t. (Radford 2016:127)
1
See Chapter 2, Further Reading.
PPs and APs can elide in predicative position (we did not look at these exam-
ples of PP-ellipsis in the discussion of ellipsis in Chapter 3), but P′ and A′
cannot:
(28) a. John is intelligent/in London, and Bill is intelligent/in London too.
b. *John is very intelligent/right in London, and Bill is very intelligent/right in
London too.
So coordination and pro-forms provide some evidence for P′ and A′. I will leave
the question of evidence for V′ aside, since we will entertain in later chapters
a very different view of what is in SpecV′ from that suggested in (1d) and (2d)
above.
X′-theory presents a category-neutral view of hierarchical relations and linear
precedence; the central idea is that these are the same for all categories. As such,
idiosyncratic categorial information is removed from the PS-rules and supplied
independently by categorial features: thus, the way in which we give a value to
the categorial variable is by assigning a feature to it. However, the endocentric-
ity of all X′-structures has one important consequence: if we specify the head
X of a category as, say, N, then X′ must be N′ and XP must be NP. This is the
notion of projection of categorial features. This idea leads to the following ter-
minological conventions for X′-structures:
(29) XP
ZP X′
X YP
In a representation like this one, XP is called the maximal projection (of X), X′
is called the intermediate projection (of X), X is the head (of both X′ and XP)
and, as we have seen, ZP is the Specifier of X′ (abbreviated SpecX′) and YP is
the complement of the head X. The dominance and constituency relations are
clear: XP dominates everything in XP and immediately dominates ZP and X′,
while X′ immediately dominates X and YP.
As we have seen, the nature of the complement depends on the c-selection
properties of the head. We saw examples of this with various verbs in (14).
Certain heads (common nouns, intransitive verbs, simple adjectives and some
prepositions) may not c-select a complement at all:
(30) a. Clover [VP slept ].
b. [NP the cat ]
c. [AP fat ]
d. John walked [PP out ].
Here, the relevant lexical entries specify that X c-selects nothing. For example,
compare the lexical entry for sleep with those in (14):
(31) sleep(V)__
Given (31), we must allow for a unary-branching expansion of V′. The same
holds for N′, A′ and P′ in (30). Nonetheless, as a default we will assume that the
hierarchical structure is complete. A proper name like Goona, for example, has
the full non-branching structure shown in (31):
(32) NP
N
Goona
Let us begin with auxiliaries. How do they fit the X′-schema? One possibility
would be to treat the combination of the main verb and auxiliary as a complex
verb form inside VP, along the lines shown in (33):
(33) [V′ [V Aux V ] NP ]
This evidence rules out a structure like (33). We also saw in Chapter 3 that the
exact counterpart of (34c), but with has also apparently deleted, is grammatical
too:
(35) I have talked to Mary and John has talked to Mary too.
The sequence in (35) is ambiguous, in that it can also mean ‘I have talked both
to Mary and to John’; this latter reading involves coordination inside the indirect
object and no ellipsis. The reading that the strike-out in (35) indicates is the one
where John is the subject, not the indirect object, of talked. We also saw that
alongside (34d), we also have coordination patterns where the auxiliary is coor-
dinated along with the verb and its complements:
(36) The chairman has resigned from the board and is living in Ruritania.
So we see that constituency tests tell us that the sequence containing the verb
and its complements is separate from the auxiliary, and that the auxiliary can
combine with the verb and its complements to form a constituent. In Chapter 3,
we referred to the latter as the mystery constituent, treating the former as VP:
(37) [S [NP John ] [?? [Aux has ] [VP passed the exam ]]]
(37) represents one possibility. Another possibility is that Aux is in SpecV′ and
what is labelled VP in (37) is actually V′. In other words, (38) could be the cor-
rect representation:
(38) [S [NP John ] [VP [Aux has ] [V′ passed the exam ]]]
There are four reasons to favour (37) over (38). First, the verb+complement
sequence (VP in (37), V′ in (38)) acts more like an XP than an X′ in relation
to constituency tests. Second, complex sequences of auxiliaries can appear
which do not act like a single constituent, but in which each auxiliary acts
like a complement of the preceding one. Third, the auxiliaries carry sentential
information regarding negation and interrogatives; SpecV′ does not seem the
right place for carriers of such information. Fourth, we will see that when we
bring the apparently exocentric category S into line with the X′-template there
is a natural head position that we can place auxiliaries in. Let us look at these
reasons one by one.
First, as we saw above several constituency tests for X′ fail: permutation,
ellipsis (except N′-ellipsis under demonstratives and numerals) and the fragment
test. But the putative V′ in (38) passes them all (permutation is restricted to
fronting, for reasons given in Chapter 3):
(39) a. (I expected John to pass the exam), and [ passed the exam ] he has!
b. John passed the exam and Mary did pass the exam too.
c. A: What did John do?
B: Pass the exam.
The other XPs pass the permutation, ellipsis (with the exception of NP) and
fragment tests and none of the X′s do. This makes the putative V′ in (38) look
very different to the other categories. On the other hand, if we treat this sequence
as VP, as in (37), it falls into line with the other XPs. Since X′-theory treats all
lexical categories as participating in the same structural configurations, (37) thus
appears to be a more desirable option than (38).
The second argument relates to sequences of auxiliaries. English allows up to
four auxiliaries in a row, rigidly ordered as Modal > Perfect have > Progressive
be > Passive be, as in:
(40) (When you arrived), John must have been being interviewed.
If auxiliaries are Specifiers of V′, then (40) must have the structure in (41):
(41) [S [NP John ] [VP [Aux? must have been being ] [V′ interviewed ]]]
Constituency tests, particularly ellipsis, show us, however, that the sequence
‘Aux?’ in (41) comprises several separate constituents, along with V′ (some
speakers, especially Americans, find (42d) ungrammatical, but I don’t; this is
further evidence that E-English does not correspond to a homogeneous set of
I-languages and does not affect the argument being made here):
(42) John must have been being interviewed right then, and …
b. Bill must have been being interviewed too.
c. Bill must have been being interviewed too.
d. Bill must have been being interviewed too.
e. Bill must have been being interviewed too.
Any sequence of auxiliaries and the main verb can be elided, but not a passive
main verb on its own (recall that this doesn’t show that the passive main verb is
not a constituent). So we must have a structure like (43) for the auxiliary+main
verb sequence where each bracketed constituent can be elided:
(44) [VP must [V′ have [V′ been [V′ being interviewed ]]]]
But now V′ looks very different from other X′ categories, in that it can have three
Specifiers (or probably four, since being could be a fourth SpecV′ although the
ellipsis test doesn’t show this). Furthermore, X′-ellipsis is hardly found with the
other lexical categories. Instead, then, we could treat each V′ in (44) as an AuxP of
some kind. This is consistent with (37), but not with (38), so again (37) is favoured
The third argument concerns negation and question-formation (in main clauses).
The rule for negation of finite clauses in English is ‘insert not after the first auxil-
iary’. We see this rule in action in (45), where not is contracted to n’t and attached
to the auxiliary (in some cases causing a phonological change to the auxiliary):
(45) a. John can’t speak Mandarin.
b. John won’t leave tonight.
c. John hasn’t passed the exam.
d. John isn’t writing a book.
(37) [S [NP John ] [?? [Aux has ] [VP passed the exam ]]]
When we first discussed the mystery category ‘??’ here in the previous c hapter,
we pointed out that it entails a complication of the basic PS-rule for S:
(48) S → NP VP
However, rule (49) does not fit the X′-schema, and we have evidence from coor-
dination and ellipsis that Aux and VP form a constituent distinct from NP here.
At the very minimum, then, we need the two rules in (50):
(50) i. S → NP ??
ii. ?? → Aux VP
The rules in (50) will generate the structure in (37). So, the question we must
now answer is: what category is ‘??’?
X′-theory now suggests a simple answer: Aux is a head, ‘??’ is the interme-
diate projection of Aux, Aux′, and S is really the maximal projection of Aux,
AuxP. In this way, Aux and its projections conform to the general X′-theory
template; S ceases to be an exocentric exception to X′-theory; we know what
‘??’ is (Aux′); the structure is consistent with the results of the various constitu-
ency tests we have been applying and consistent with the fact that negation and
question-formation are sentence-level operations applying to Aux, since Aux is
the head of the sentence.
The tree diagram (51) gives the full structural description of a simple sentence
as AuxP, generated by the rules in (50):
(51) AuxP
NP Aux′
he
Aux VP
should
V PP
talk
P NP
to
D N
this man
Although there is no auxiliary in examples like (52), there is still a tense: (52a)
is marked for past tense, (52b) is, at least indirectly, marked for present tense
by the third-person singular agreement and, as we argued in Chapter 1, the verb
fish in (52c) has a silent present-tense marker. Auxiliaries often mark tense too,
notably English future will and shall, while have marks perfect, which is at least
partly a tense. So let us rechristen AuxP as TenseP, or TP. This is arguably a
less parochially English label than AuxP, as very many languages have tense
marking but auxiliaries are arguably somewhat rarer. Very few languages have
an auxiliary system as rich as English, allowing, as we saw in (42), up to four
auxiliaries in a row.
So we replace (51) with (53):
(53) TP
NP T′
he
T VP
should
V PP
talk
P NP
to
D N
this man
Here the subject is in SpecT′, the head of the clause is T and VP is the comple-
ment of T in T′.
We can make some sense of the English phenomenon of ‘do-support’ in terms
of the structure in (52). The basic rule is that if no auxiliary is independently
present in the main clause, do appears where VP is elided, where the sentence is
negated or where subject-auxiliary inversion takes place:
(54) a. Sarah finished her paper, and Mary did finish her paper too.
b. Sarah didn’t finish her paper.
c. Did Sarah finish her paper?
The appearance of do where the verb is elided implies that tense and agreement
must always be expressed. If there is no verb or auxiliary independently availa-
ble, do is inserted. Similarly, we can think that the presence of negation in (53b)
and the intervening subject in (53c) mean that the operation combining T with
V, which clearly takes place in examples like (51a,b) (and covertly in (51c)),
is somehow prevented from applying, and so the meaningless auxiliary do is
inserted, in order to mark tense and agreement. The auxiliary do should be dis-
tinguished from the homophonous main verb do, as in do the tango: the latter do
also requires do-support in the relevant contexts, as (54) shows:
Two of the four arguments we gave for treating VP in finite clauses as a con-
stituent separate from Aux/T apply to the infinitives with to. Constituency tests
such as fronting, ellipsis and fragments give very similar results to those seen in
(39) for finite clauses, for example (although (55a) is not quite as natural as the
other examples):
(57) a. ?(Goona arranged for Vurdy to arrive on time) and [ arrive on time ] he
managed to!
b. Goona wanted to arrive on time and Vurdy wanted to arrive on time too.
c. A: What did Goona want Vurdy to do?
B: Arrive on time.
Since modals are intrinsically finite in Modern English, they are in complemen-
tary distribution with to, which is intrinsically non-finite. Modals and to are
always followed by an uninflected verb or auxiliary; similarly, do is in examples
like (54a,b). The arguments related to negation and inversion given above do not
apply to to: the contracted negation n’t is restricted to finite clauses and inver-
sion can only apply in main clauses, while infinitives are always embedded. But
we can conclude from (57) and (58) that the mystery category in (56a) is T and
that to is the marker of non-finite T.
So we can bring S fully into line with X′-theory as TP. The next question con-
cerns S′. Can this apparently exocentric category be reduced to the X′-template?
Here we will see how this can be done, but we will leave one major question
open until the next chapter. Let us look again at the representation of a sentence
containing an embedded complement clause introduced by the complementizer
that (here S has been replaced by TP, Aux by T, and T′s have been added, along
with specification [Pres] in the main-clause T):
(59) TP
NP T′
T VP
[pres]
V S′
N
hopes
Mary
Comp TP
that NP T′
N T VP
John will
V
arrive
This raises the question of what corresponds to ZP in (62i), i.e. what can func-
tion as SpecC′. I will leave this question open until Chapter 5.
Another question is whether we need CP in main clauses. Up to now we
have treated S′/CP as the category for subordinate clauses and S/TP as that
for main clauses. This is justified to the extent that the material following that
in a finite subordinate clause corresponds to a stand-alone main clause. But
this does not, in itself, preclude the possibility of a silent C introducing main
clauses. Furthermore the rules in (62) would clearly generate this structure.
Some languages, e.g. some varieties of Welsh and Gascon, seem to have overt
Cs introducing declarative main clauses, and many languages have interroga-
tive particles similar to English if and whether introducing main-clause ques-
tions, supporting the idea that main clauses are CPs, i.e. introduced by the
rules in (62). In Chapter 7, when we look at the syntax of languages other than
English, we will see that German supports the presence of a main-clause CP.
We saw that (63d) can be taken to indicate that adjuncts are outside the VP and
so are not substituted under do so replacement, but the fact that the substituted
VP in (63a) is interpreted as put his car in the garage on Tuesday indicates that
the adjunct PP is part of the VP. We concluded that adjuncts are optional constit-
uents of VP, while selected arguments are obligatory constituents of VP.
In terms of X′-theory, selected arguments are complements of X inside X′, so
the NP his car and the PP in the garage in (63) are in V′. This, combined with
the two-layer structure of XPs given by the X′-template, allows us to make a
structural distinction between arguments and adjuncts. We can treat adjuncts as
categories adjoined to V′ in a structure like (64):
(64) VP
spec V′
V′ PP
V NP PP P NP
put the car in the on Tuesday
garage
(21) a. John read a book of poems and Mary read one too.
b. *John read a book of poems and Mary read one of stories.
c. John read a book with a red cover and Mary read one too.
d. John read a book with a red cover and Mary read the one with a blue cover.
Since selected arguments are complements of X inside X′, the PPs of poems
and of stories in (21a,b) are in N′, as complements of the noun book. Since one
substitutes for N′, it must substitute the complement to N, hence the ungram-
maticality of (21b). On the other hand, the two with-PPs in (21c,d) are optional
N′-adjuncts in a structure like (65):
(65) NP
D N′
the
N′ PP
N PP P NP
book of stories with the blue/red cover
In cases of N′-substitution the adjunct PPs are optionally substituted, while the
complement PPs are obligatorily substituted.
Putting this approach to adjuncts together with the rest of the X′-template, we
have the following set of category-neutral PS-rules:
(66) a. XP → ZP X′
b. X′ → X′ WP
c. X′ → X YP
We can generalise the notion of adjunction in (67b) to ‘sister of Xn’ and ‘daugh-
ter of Xn’, where Xn may be any bar-level. This allows adjunction to maximal
projections (XPs) and heads (Xs). We will largely leave these possibilities aside
here and return to the nature of adjunction in Volume II.
The universalist assumption is that, by default, all phrasal categories in all lan-
guages will have the internal structure generated by the rules in (64) and so can
be described in terms of (65), along, of course, with the central notions of head
(X) and maximal projection, XP. This represents a very strong claim about UG
and about the possible structures we expect to find in the languages of the world.
4.6 Conclusion
In this chapter we have seen how we can posit a single X′-template
which subsumes the PS-rules of English which we saw in Chapter 3 and explains
why only rules of that kind and not ‘crazy’ rules like those in (3) are possible.
The template, complete with the possibility of X′-adjunction, is given in (68):
(68) XP
ZP X′
X′ WP
X YP
Exercises
1. Give the tree diagrams for the following NPs, and comment on how
they instantiate, or fail to instantiate, the X′-schema:
a. him
b. the composer of Yesterday
c. two pints of beer
d. my cat
e. his professor of phonetics
f. John’s cat
2. Give the tree diagrams for the following APs and PPs, and comment
on how they instantiate, or fail to instantiate, the X′-schema:
a. in Paris
b. just over the fence
c. out from under the bed
d. just there
e. very worried about the result
f. really amused that Boris hates Michel
3. Give the tree diagrams for the following clauses, and comment on
how they instantiate, or fail to instantiate, the X′-schema:
a. Andy visited.
b. Andy has insulted Lou.
c. We had guessed that Andy would insult Lou.
d. We had thought that Andy would admit that she had
insulted Lou.
e. I have very much enjoyed this lecture.
f. Lou asked if he should ignore Andy.
g. Lou asked whether to ignore Andy.
h. Lou could have ignored Andy.
i. Lou decided on the plane.
Further Reading
Carnie, A. 2013. Syntax: A Generative Introduction. Oxford: Blackwell, Chapters 6–8.
Freidin, R. 2012. Syntax: Basic Concepts and Applications. Cambridge: Cambridge
University Press, Chapter 5.
Fromkin, V., R. Rodman & N. Hyamset. 2000. Linguistics: An Introduction to
Linguistic Theory. Oxford: Blackwell, Chapter 3.
Haegeman, L. 1995. Introduction to Government and Binding Theory. Oxford:
Blackwell, Chapter 3.
Haegeman, L. & J. Guéron. 1999. English Grammar: A Generative Perspective.
Oxford: Blackwell, Sections 1.2.4 and 1.2.5.
Larson, R. 2010. Grammar as Science. Cambridge, MA: MIT Press, Units 14–18.
Sportiche, D., H. Koopman & E. Stabler. 2014. An Introduction to Syntactic Analysis
and Theory. Oxford: Wiley, Chapters 4–6.
specifier X′
X′ adjunct
X complement
Here, as we saw, the complement is the sister of X and daughter of X′, the
adjunct is the sister of X′ and the daughter of X′ (we mentioned that XP- and
head-adjunction may also be possible) and the Specifier is sister of X′ and
daughter of XP. Specifier, adjunct and complement are all relational notions,
defined in terms of two positions in the tree. The truly primitive notions are the
structural ones X, X′ and XP. These relations are in turn defined in terms of the
categorical value of X (remember that X is a variable ranging over categories)
and the notion of projection from X to X′ to XP.
More generally, what we have seen up to now is that PS-rules generate well-
formed structural descriptions (whether formulated in terms of X′-theory or not,
although in the last chapter we saw good reasons to prefer X′-theory over the
more ‘traditional’ kind of PS-rules). The structural descriptions are represented
as tree diagrams or labelled bracketings, those generated by X′-theory being
more abstract, in particular in that they generate a category-neutral template
which is in principle the same for all lexical and functional categories. In either
format, the PS-rules can create recursive, and therefore potentially infinite,
structures.
All of this gives us a good basis for the theory of constituent structure, but
there are two questions which we need to address, one quite specific and one
much more general. The specific question is one we have held over from our
discussion in the previous chapter of how X′-theory applies to functional cate-
gories: what goes in the Specifier of CP?
99
The general question relates to the kinds of rules PS-rules are, whether
restricted to X′-theory or not. An obvious way to think about PS-rules is as
‘building rules’: as we apply them, we build up the constituent structure. As we
saw in Chapter 3, we can relate each PS-rule to a specific ‘treelet’, a little piece
of a tree diagram; look again at (13) of Chapter 3 and the preceding discussion.
The same can be said of X′-theory. We can relate the X′ PS-rules given in (63)
of Chapter 4 and repeated here as (2) to the tree in (3):
(2) a. XP → ZP X′
b. X′ → X′ WP
c. X′ → X YP
(3) XP
rule (2a)
ZP X′
rule (2b)
X′ WP
rule (2c)
X YP
The auxiliary will is in one position in the declarative (4a) and in another in the
interrogative (4b). Informally, we say that will ‘moves’ from one place in the
structure to another.
C TP
NP T′
Cambridge
T VP
will
T
flood
b. C′
C TP
T NP T′
will Cambridge
T VP
(will )
T
flood
to a new position and pasted in there. From the fact that the original copy of will
in the T position is not heard, we can add that there is a general copy-deletion
rule applying to moved elements in their original position (we will look at this
operation in more detail in Section 8.6). Since T is a head, this is a case of
head-movement. Head-movement does what it says: it takes a head, copies it to
a new position and deletes the content of the original head (but not the actual
position). The basic tree is unchanged by the movement operation, aside from
the substitution of T into a position dominated by C. So English subject-auxil-
iary inversion is a case of T-to-C movement, itself a case of the general opera-
tion of head-movement. There is much more to say about head-movement, and
we will come back to it in Chapter 7 and at greater length in Volume II. But for
now we have the basics, and an illustration of how movement operates.
The second kind of movement, NP-movement, is found in passives in English.
We saw an example of a passive in the discussion of ellipsis and English auxil-
iaries in Chapter 4:
(6) (When you arrived), John must have been being interviewed.
If we take away the complex auxiliary sequence, we can construct the following
active-passive pair:
(7) a. Oprah interviewed John. (active)
b. John was interviewed (by Oprah). (passive)
These two sentences are very close to synonymous; in particular, the who’s-
doing-what-to-whom relations (known as argument structure) are the same:
both sentences describe a situation in which Oprah is the interviewer and John
is the interviewee. Furthermore, interview is a transitive verb; in terms of the
c-selection frames given in (14) in Chapter 4 it acts like watch, so it has a lexical
entry like (8):
(8) interview (V) ___ NP
This c-selection requirement is clearly satisfied in (7a), where John is the NP.
We also saw in Chapter 4 that this c-selection frame is associated with the
PS-rule which generates NP as the complement of V. In turn, this position cor-
responds to the direct-object function for NP, so in (7a) John is the direct object
of interview. The position of complement of V and corresponding direct-object
function is typically associated, in English and in many other languages, with
the semantic, or thematic, role of Patient or Undergoer of the action described
by the verb. In other words, John is the interviewee, as we already observed.
In (7a), Oprah is the subject; NPs in this position are typically associated, in
English and in many other languages, with the thematic role of Agent: Oprah is
the interviewer.
Now, if we compare the passive sentence in (7b) with the active one in (7a),
we can make two observations. First, John has the subject function in (7b): we
can see this from the word order (it precedes both the auxiliary and the main
verb) and from subject-verb agreement, in that if we replace John with a plural
NP (the boys) then the auxiliary changes to its plural-agreement form: The boys
were interviewed (by Oprah). There are other ways to see that John is the subject
of (7b), but these will do.
We have already made the second observation: the who’s-doing-what-to-
whom/argument structure relations are the same as in (7a).
So active–passive pairs change grammatical function: the NP which is the
direct object of the active is the subject of the passive (while the NP which is
the subject of the active goes into the by-phrase of the passive and can even
disappear altogether: (7b) is grammatical if we leave out by Oprah). In our brief
discussion of subjects and predicates in Chapter 3, we observed that phrase
structure does not represent grammatical functions: it merely deals with cate-
gories, hierarchies and linear order. However, we can define the functions of
subject and direct object in X′-theoretical terms as follows:
NP T′
T VP
V NP
interviewed John
b. TP
NP T′
John
T VP
was
V NP
interviewed (John )
(The trees in (10) leave aside the active subject NP Oprah; accounting for the
relationship between the subject in the active sentence and the by-phrase in the
passive sentence is a tricky business which would take us too far afield here.)
Looking at the representation of the direct-object NP John in (10), we can see
that it is moved, by the same copy/paste + deletion operation as that applied
to T in (5), from the VP-internal direct-object position to the Spec,T′ subject
position. As with the head-movement operation in (5), a silent copy, or trace,
is left behind in the original position (indicated by strikeout). Because what is
moved here is an NP we call this NP-movement. In (10), again as in (5), the
tree is unchanged; just the position of one NP is changed. The object NP here is
substituted into the subject NP position in SpecT′.
As with head-movement, there is much more to say about NP-movement, and
we will come back to it at great length in Volume II. But, again, for now we have
the basics, along with a further illustration of how movement operates.
For the rest of this chapter, though, we will focus on the third type of move-
ment, wh-movement.
Here the wh-phrase what, an NP, moves from the direct-object position (indi-
cated here by the copy in parentheses) to the initial position of the sentence to
form a direct question. The question is interpreted as ‘asking about’ the nature of
the direct object, so a natural answer could be ‘fish’, for example. Wh-questions
contrast with yes/no-questions of the kind seen in (4b), as the latter simply ask
whether a given situation (e.g. Cambridge flooding as a consequence of global
warming) holds or not.
Wh-questions are so-called because, in English, the words and phrases that
introduce them, i.e. which stand in initial position of the interrogative clause,
nearly all begin with the letters ‘wh’: when, who, which, what, where, why, how,
etc. Wh-movement is the most important and interesting type of movement for
various reasons, some of which we will see directly.
In order to get a reasonably full picture of wh-movement, there are several
questions to be addressed. These include the following (note in passing that
these are all themselves wh-questions):
(12) a. Which XPs can move?
b. Where do they move from?
c. Where do they move to?
Looking first at question (12a), we can see from the following examples
that quite a range of categories can move (although, as (12a) implies, these
are all phrasal categories, XPs, so wh-movement is clearly distinct from
head-movement):
(13) a. Which man will David see t ?
b. Who should John talk to t ?
c. To whom can John talk t ?
d. How disappointed is Fred t ?
e. What does Don believe t ?
In all of these examples and from now on, I abbreviate the copy in the starting
position as ‘t’ for trace; this is purely for convenience; strictly speaking, there
is a whole phrasal copy in the relevant positions. In (13a), as in (11), we see
wh-movement of the object NP. The difference with (11) is that here we have the
wh-determiner which followed by the Noun man, while in (11) we have the pro-
nominal form what which substitutes the whole NP. In (13b), the indirect object
NP is moved; here we have the animate wh-pronoun who. The NP who is moved
from inside the indirect-object PP, headed by to. Since to is unaffected by the
movement, this kind of example is known as Preposition stranding. Preposition
stranding is very natural in English, as (13b) shows. It is, however, quite rare in
the languages of the world; the only other languages which clearly allow it are the
Scandinavian languages. In (13c), we see movement of the whole indirect object
PP. This is known as pied-piping (the terminology alludes to the Pied Piper of
Hamelin, who having led the town’s rats to a watery death and not been paid, led
the town’s children to a similar fate in revenge). Here the Preposition, which is
not itself questioned in (13c), is ‘pied-piped’ along with the questioned NP. In
languages which do not allow Preposition stranding (which, as just observed, is
most languages), pied-piping of the Preposition is the only way to question the
indirect object. In English, on the other hand, pied-piping of this kind, although
grammatical, sounds rather ‘stilted’, at least in the spoken language. The sense of
artificiality is added to by the use of the archaic accusative form of the wh-word:
whom. This form is all but moribund in most varieties of colloquial English and
is associated with formal, literary registers.1
1
However, it is arguably more natural in (13c) than the corresponding example with who:
(i) ??To who can John talk t?
Compare also the Preposition-stranding example (13b) with whom:
(ii) ?Whom did you talk to t?
Strictly speaking, whom should also appear in direct-object function in (13a):
(iii) ?Whom will David see t?
Here the capitals and other punctuation are intended to indicate that the
wh-phrase is heavily accented. Most important for our present purposes, though,
these examples clearly show the wh-phrases in their unmoved positions.
The second construction is multiple questions. These are questions featuring
more than one wh-phrase, and hence questioning more than one constituent.
In English, only one wh-phrase can move per clause, and so any others must
remain in situ. Some examples are:
(15) a. Which prizes did we award t to which students this year?
b. Which girls did you say t danced with which boys at the party?
In these examples one wh-phrase moves to the beginning of the sentence, while
the other one stays in situ: we can see that in both of these examples the in-situ
wh-phrase is inside a PP. If we make the moved wh-phrase into a non-wh-phrase,
then the other one must move, and we have the usual choice of Preposition
stranding or pied-piping:
(16) a. Which students did we award the Schlumpfenberger Prize to t this year?
b. To which students did we award the Schlumpfenberger Prize t this year?
c. Which boys did you say Mary danced with t at the party?
d. With which boys did you say Mary danced t at the party?
Native-speaker judgements vary on these marginal examples. My own judgement is that (i)
sounds very strange, as it involves a clash of colloquial (preposition-stranding) and formal reg-
isters, while both (ii) and (iii) sound rather archaic and/or very formal. Native-speaker read-
ers should judge for themselves (don’t take the question marks in front of these examples too
seriously). Here again, we see that the cultural E-language concept ‘English’ masks a range of
I-language diversity.
Clearly, the simplest assumption about in-situ wh-phrases in both echo ques-
tions and multiple questions is that they simply do not move; in that case, these
examples confirm our semantically based intuitions about the original position
of wh-phrases in wh-movement examples in general.
These points are further confirmed by evidence that the position and func-
tion of the wh-phrase are occupied in wh-movement examples, even if the
wh-phrase has apparently moved away from that position. We can see this in
two ways. First, if we insert a wh-phrase in the beginning of a sentence with
no corresponding functional or positional ‘gap’ in the sentence, the result is
ungrammatical:
(17) a. *Who does Grandma like Fido?
b. *Which student did we award the prize to Bloggs?
c. *How many people did you say Mary had been to Chicago?
The contrasts between (17) and (18) can be straightforwardly accounted for if
we take seriously the copy/paste + deletion idea about movement. In each of
the examples in (18), the wh-phrase is moved from the position indicated by the
trace; hence it fills the position occupied by the trace and, derivatively therefore,
has the function associated with the trace’s position: direct object in (a), indirect
object in (b) and subject in (c). The examples in (17) are ungrammatical because
either the wh-phrase or the constituent in the respective trace positions has no
syntactic home; there is no position or function available for it in the structure.
They are ungrammatical for the same reason as (19):
(19) a. *Grandma likes Fido Bill.
b. *We awarded the prize to Bloggs Smith.
c. *You said Mary more people than me had been to Chicago.
A second way to make the same point is to take, for example, an intransitive
Verb of a kind that really cannot take a direct object however hard we try, e.g.
arrive:
(20) a. The train/Mary arrived.
b. *John arrived the train/Mary.
This example cannot allow what is sometimes called the causative reading of
otherwise intransitive verbs, in that it cannot mean ‘John made the train/Mary
arrive.’ Compare break in (21):
(21) a. The vase broke.
b. John broke the vase.
Again, in (22a) the wh-phrase, or more precisely its copy/trace, has no syntactic
‘home’ in the structure, no position and therefore no function. Therefore the
result is ungrammatical (and semantically nonsensical).
So the answer to question (12b) is that wh-phrases move from independently
needed structural positions associated with particular grammatical functions:
direct object, indirect object, subject and predicate positions for example. Hence,
going back to our original example (11), we can give the pre-wh-movement
structure in (23):
(23) C′
C TP
T NP T′
did you
T VP
(did )
V NP
catch what
Here we see what in the direct-object position, which we can motivate from
the semantic interpretation, echo questions, multiple questions and examples
equivalent to (19) and (20b). Since (11) is a direct question, we also have sub-
ject-auxiliary inversion (i.e. head-movement of T to C, as we saw in the previous
section) here, this time moving T containing the ‘dummy’ auxiliary do to C.
Now we can turn to question (12c): where do wh-phrases move to? In the light
of the structure in (11) and the general X′-template introduced in Chapter 4, the
answer to this question is now quite clear. Let us look just at the top part of the
structure in (23) and compare it with the X′-template in (1):
(24) C′
C TP
(25) XP
ZP X′
X′ WP
X YP
Mapping the general template in (25) onto the structure in (24), we can see
that C′ contains the head C (the target of head-movement of T, in fact) and the
complement TP. There is no adjunct WP adjoined to T′ here; recall that adjuncts
are always optional. What is missing in (24) is ZP Specifier of C′. Given the
linear order of the wh-phrases and the inverted auxiliary in all our examples of
wh-movement so far, the obvious conclusion is that the wh-phrase is in SpecC′.
So the structure for (11) after wh-movement has applied is (26):
(26) CP
NP C′
what
C TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
ZP C′
C TP
NP T′
T VP
V NP
Here we see the three phrasal categories CP, TP and VP. Loosely speaking, each
one seems to play a characteristic syntactic-semantic role in building up a full-
fledged clause.
The three clausal domains are, first, the VP, which specifies basic argument
relations, i.e. who does what to whom. In (27), the subject NP is outside VP; this
may be because subjects have a special status in relation to the VP because of the
importance of the subject-predicate relation, or it may be that subjects originate
In Chapter 4, we considered and rejected a structure like (28) for the auxiliary
sequence in (6):
(28) [VP must [V′ have [V′ been [V′ being interviewed ]]]]
Each auxiliary occupies the head of its functional category and takes the next
one to its right as its complement, with Voice taking VP as its complement. The
subject occupies SpecT′; for now we leave open the question of what, if any-
thing, occupies the other Specifier positions (remember that Specifier positions
don’t have to be filled, as in simple NPs like John). Heads expressing Mood,
Tense, Aspect and Voice are often realised by inflections in other languages
(and indeed there are inflections in the English auxiliary sequence, e.g. the pro-
gressive -ing). So what we might loosely call the ‘extended TP′ is the domain
where notions such as Tense, Aspect and Mood are realised. Depending on the
language, these may be realised by auxiliaries, inflections, both together, or not
at all. English relies mainly, but not exclusively, on auxiliaries.
The third domain is the one relevant for wh-movement. We could take CP,
as the highest category in the clause, to be the one that relates the otherwise
complete clause (TP complete with subject and satellites, the VP saying what
the subject is doing to whom) to the ‘outside’. What’s outside the clause will
vary depending on whether the clause is a main clause or not. If the clause is a
main clause, there is no further, higher, syntactic structure. The clause is a root
clause (the CP node is the root of the tree). But root clauses have a function in a
wider discourse context; in particular, it is relevant for the discourse whether the
clause is a declarative, stating something, an interrogative, asking something,
an imperative, ordering someone to do something, etc. In other words, the root
clause must express discourse-relevant information regarding clause type. There
is much more to say about this matter, but for now it suffices to observe that
declaratives and interrogatives are two of the most basic clause types.
C TP
T NP T′
will Cambridge
T VP
(will )
T
flood
With this word order, i.e. with subject-auxiliary inversion, the sentence is inter-
preted as a yes/no question. So we could add a special clause-type feature to C,
to indicate this. Let us call the feature Q (for Question) and write CQ to designate
an interrogative C, a particular subcategory of C. So now we can say that CQ
attracts T, i.e. it has the syntactic effect of triggering subject-auxiliary inversion
and the discourse-semantic effect of saying ‘I’m asking you a yes/no question’
(however, exactly that is to be analysed by the semantic component of the gram-
mar, but we leave this aside since it is not a matter for syntax).
Wh-questions are obviously a type of question, so we can further type C with the
feature Wh; CQ, Wh is another subcategory of C. We can say this type of C has the
syntactic effect of triggering wh-movement of some XP into SpecC′. Since in main
clauses wh-questions also involve subject-auxiliary inversion, we could say that Q
is also present on C. So the full structure of a simple wh-question like (11) is (30):
(30) CQ, WhP
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
Following the X′-theoretic notion of projection, we can take it that the Q and
wh-features of this kind of C are marked on C′ and CP. Hence, in the discourse,
CP is a wh-question, with its particular discourse semantics (which, again, we
leave aside here as extra-syntactic). The syntactic property of CQ, Wh in a main
Here it looks as though the wh-phrase is in situ, since the example appears to
correspond point-for-point with the declarative in (31a). However, it is clear that
this is not an echo question; that would presumably be (32):
(32) WHO saw John?!
NP CWh′
who
CWh TP
NP T′
(who )
T VP
V NP
saw John
Here, the subject wh-phrase has moved from SpecT′ to SpecC′ in the standard
copy/paste + deletion manner, and this movement is presumably triggered by
the wh-feature on T, as in the examples we saw in the previous section. This
example of wh-movement does not change the linear order of the words in the
sentence, though; it is string-vacuous movement. That is why (31a) and (31b)
appear to line up (while, if (33) is the correct analysis of (31b), they actually
don’t). There is nothing in the notion of movement rule that states that move-
ment operations, copy/paste + deletion, have to change the linear order of ele-
ments; the fact is that they usually do, and all the examples we have seen so far
have involved such a change. But, as long as we don’t require this, then there is
no bar to positing the structure in (33) for (31b).
The problem, of course, is the absence of subject-auxiliary inversion, sig-
nalled in the absence of other auxiliaries by do-support, in (31b). In fact, if we
apply subject-auxiliary inversion to (31b), the result is strongly ungrammatical,
as long as do is left unstressed:
(34) *Who did see John?
Since, as (35b) shows, stressed do can appear in T, we can assume that the
stressed do in (35a) is also in T. In that case, what seems to give rise to the
ungrammaticality of (34) is subject-auxiliary inversion.
We can connect (34) to another strange feature of subject wh-movement.
As we saw with the multiple wh-question in (15b) in the previous section,
movement of wh-subjects from a subordinate clause to a main clause is
grammatical:
(36) a. Who did you say [ t likes John ] ?
b. Which student do you think [ t will win the prize ] ?
Here the brackets show the subordinate clause, ambiguously indicating either
CP or TP. In these examples the complementiser that is absent; this is consistent
with the general optionality of this complementiser introducing the clausal com-
plements to verbs like say and think in English. However, if we take the option
of inserting that in (36), the examples become ungrammatical:
(37) a. *Who did you say [CP that [TP t likes John ]] ?
b. *Which student do you think [CP that [TP t will win the prize ]] ?
We only see this contrast with subject wh-movement. If we wh-move the object
in examples like this, that is fully optional, just as in subordinate clauses where
no wh-movement takes place:
(38) a. Who did you say [CP (that) [TP John likes t ]] ?
b. Which prize do you think [CP (that) [TP Bloggs will win t ]] ?
Similarly with for: many speakers of English allow optional for with certain
infinitive-taking verbs, e.g. (in my English at least) prefer:
(42) I would prefer (for) John to win the prize.
We see the same disappearing optionality with subject wh-movement, but not
with object wh-movement, as we saw with that in (36) and (37):
(43) a. Who would you prefer [ t to win the prize ] ?
b. *Who would you prefer [CP for [TP t to win the prize ]] ?
c. Which prize would you prefer [CP (for) [TP John to win t ]] ?
types the clause for discourse semantics. Its syntactic effect is to trigger sub-
ject-auxiliary inversion, T-to-C movement. But subject-auxiliary inversion
is impossible in (34), where we have subject wh-movement. What does this
mean for the Q-feature? Clearly, since (31b) is interpreted as a question, we
want to say that C still has a Q-feature, but (39) prevents it from trigger-
ing subject-auxiliary inversion where there is an adjacent subject trace, i.e.
where there is subject wh-movement. We will see directly that there are fur-
ther examples where C has a Q-feature but does not trigger subject-auxiliary
inversion.
This brings us to the topic of wh-movement in indirect questions. As in direct
questions, we see a range of possibilities for wh-movement:
(44) I wonder …
a. which man John will see t ?
b. who John should talk to t ?
c. to whom John can talk t ?
d. how angry Alex is t ?
e. what John believes t ?
We mentioned in the previous section that CP types the clause and we saw there
how this works for yes/no questions and wh-questions in main (root) clauses. In
subordinate clauses, CP bears clause-typing features that are c-selected by the main
verb of the superordinate clause. So, for example, the verb wonder selects CQ, since
its complement must be an interrogative of some kind (yes/no or wh-), while think
does not allow the head of its CP complement to be CQ:
(47) a. I wonder if/*that the world is flat.
b. I think *if/that the world is flat.
These restrictions on how CQ and CQ, Wh are realised in main and embedded
clauses are quite complicated and idiosyncratic. In other languages, things can
be different. For example, certain non-standard varieties of Dutch allow the
equivalent of that in indirect wh-questions. After a verb like ‘wonder’, then, we
find clauses like the following:
Here wie (‘who’) occupies SpecC′ and dat (‘that’) is in C. So this example is
exactly equivalent to the ungrammatical English sentence in (46a). Even in ear-
lier stages of English, we can find something similar. The first line of Chaucer’s
General Prologue to the Canterbury Tales, written in the Middle English of the
late fourteenth century, begins with the wh-word whan (‘when’) and is immedi-
ately followed by that and then the subject:
(50) Whan that Aprill with his shoures soote …
When that April with its showers sweet …
of the embedded clause (see (35), for example). In fact, we can find examples of all
the other kinds of wh-movement from an embedded clause to a main clause:
(51) a. Which man did you say [ (that) John will see t ] ?
b. Who did you say [ (that) John will talk to t ] ?
c. To whom did you say [ (that) John will talk t ] ?
d. How angry did you say [ (that) Alex is t ] ?
e. What did you say [ (that) John believes t ] ?
Since wh-movement can move material from a subordinate clause to the root
SpecC′, we expect this to be possible from any subordinate clause, however
deeply embedded. At first sight, this certainly appears to be true, as the follow-
ing examples show:
(53) a. Who did you say [ (that) Mary believes [ (that) John saw t ]] ?
b. Who did you say [ (that) Mary believes [ (that) Fred knows [ (that) John saw t ]]] ?
c. Who did you say [ (that) Mary believes [ (that) Fred knows [ (that) I asserted
[ (that) John saw t ]]]] ?
There doesn’t appear to be any limit to how deeply embedded the subordinate
clause from which wh-movement takes place can be. So we see that wh-move-
ment can produce constructions with an unbounded dependency. The depend-
ency between the trace/copy of the wh-phrase in the starting position and the
SpecC′ which the wh-phrase moves to appears to be unbounded: wh-movement
can be as ‘long-distance’ as we like. The same appears to be true for wh-move-
ment in indirect questions. If we change the main-clause verb to wonder in (53),
then we have wh-movement to the first embedded SpecC′, the one c-selected by
wonder. Here, too, wh-movement appears to be unbounded:
(54) a. I wonder [ who Mary believes [ (that) John saw t ]] ?
b. I wonder [ who Mary believes [ (that) Fred knows [ (that) John saw t ]]] ?
c. I wonder [ who Mary believes [ (that) Fred knows [ (that) I asserted [ (that)
John saw t ]]] ?
generalisation of (38). In the next section, we will see that there is a whole class
of important restrictions on wh-movement.
In this section we have seen some more complex aspects of wh-movement: the
special properties of subject wh-movement, including the strange C-trace gener-
alisation in (38); the behaviour of wh-movement and wh-Cs in embedded clauses
and, most importantly, the apparently unbounded nature of wh-movement. In the
next two sections we will focus on some consequences of this last observation.
b. V′
V CP
Rule (56a) will guarantee that exactly one wh-phrase appears in the Specifier
of CWh and rule (56b) introduces TP as the complement to CWh, seemingly in
the standard way (see (56i) of Chapter 4). But now we run into a problem. The
rules in (56) tell us what the properties of a CP which attracts a wh-phrase are,
but we need to also specify that the TP must contain a trace/copy of the WhP
in (56a). If we do not do this, we cannot rule out ungrammatical examples like
(17), repeated here:
(17) a. *Who does Grandma like Fido?
b. *Which student did we award the prize to Bloggs?
c. *How many people did you say Mary had been to Chicago?
But the wh-features everywhere except on the root C and CP are not the same
as the ‘true’ clause-typing wh-features we have seen until now. We can see
this particularly clearly in the embedded CPs. These are not indirect ques-
tions, but they have to be marked ‘wh’ to show that they contain a trace/copy.
So we need to find a way to distinguish the ‘trace/copy-marking’ wh-feature
from the clause-typing one. The clause-typing feature clearly has a lexical
motivation (some verbs c-select indirect questions and some don’t) and a dis-
course-semantic motivation, as we saw above. The ‘trace/copy-marking’ fea-
ture is there simply to say that there is a trace/copy somewhere ‘lower down’
in the structure. If traces/copies are not independently existing lexical or func-
tional categories, then indicating them in a PS-rule is no more than a way to
encode the movement dependency. Moreover, this approach goes against the
general locality of PS-rules, in that it states there is a trace/copy somewhere
lower in the structure, but this trace/copy could be arbitrarily deeply embed-
ded, given the unbounded nature of wh-movement. So marking movement
dependencies in PS-rules makes the PS-rules significantly more complex by
effectively introducing whole classes of new categories to mark traces/cop-
ies, and effectively makes PS-rules non-local. Furthermore, the complexity
increases greatly if we try to introduce a similar method for marking head- or
NP-movement dependencies.
So, rather than complicating the PS-rules very significantly as sketched above,
although this can be and has been done, we will stick to the idea that we there are
two relatively simple rule types. On the one hand there are PS-rules of the kind
introduced in Chapters 3 and 4 (including the X′-template), which build struc-
ture, i.e they generate phrase markers in a strictly local, incremental fashion. On
the other hand, there are movement, or transformational rules, which map phrase
markers into other phrase markers, preserving the X′-template. Movement con-
sists of the copy/paste + deletion operations, as we have seen. The former rules
are highly local, the latter, at least wh-movement, appear to be unbounded. The
former rules do not create traces/copies; the latter do.
To summarise what we have seen concerning movement so far, we have seen
three types of movement: head-movement, NP-movement and wh-movement.
Head-movement is illustrated by subject-auxiliary inversion, as in (5b), repeated
here with CP and the Q-feature added:
(5b) CQP
CQ′
CQ TP
T NP T′
will Cambridge
T VP
(will )
V
flood
By definition, head-movement moves a head. The target is also a head. The move-
ment operation copies the head T and pastes it into the ‘target head’. The copy left
behind remains in the structure but is not phonologically realised (from a phono-
logical perspective, it is deleted; from a structural perspective it is not, however).
The second type of movement is NP-movement. The example we saw was
movement of the object NP from the complement position of V to the subject
position, SpecT′, in passives. We saw the derived structure of a passive in (10b),
repeated here:
(10b) TP
NP T′
John
T VP
was
V NP
interviewed (John )
NP-movement moves NPs, as the name implies. Passives are a typical, but not
the only, case. Here the object is copied and pasted into the subject position,
and the copy in object position is phonologically deleted. NP-movement always
targets the subject position, SpecT′.
Third, there is wh-movement, which moves a wh-phrase to the Specifier of
CWh. Wh-phrases can be NPs, APs, PPs or CPs, but not VPs or TPs (or other
clausal functional categories such PerfP, ProgP or VoiceP; see (28)). A basic
case of object wh-movement is (30), repeated here:
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
Here the wh-phrase is copied and pasted into SpecCQ, wh′, and the copy is pho-
nologically silenced. Wh-movement always targets a SpecCwh′. We have been
taking Cwh to imply CQ, since wh-questions are a kind of question; we will return
to this point below.
All three cases of movement have certain things in common: they copy/
paste a constituent into a distinct position created by the X′-rules of phrase
structure. Technically, what we have informally been calling copy/paste is
known as substitution. So all the movement rules we have seen so far are
substitution transformations. (There are also adjunction transformations,
which create adjoined positions of the kind generated by PS-rules like (2b),
but we will leave these aside for now.) All of the movement rules leave a copy
in the original position and this copy is deleted phonologically; in other words,
it is present in the syntax (we have mostly abbreviated it as ‘t’ for trace) but
lacking phonological content, i.e. silent. We can also note that each movement
rule places the new copy (i.e. the ‘moved’ category) in a position which is, in
an obvious sense, higher up in the tree than the original position: C is higher
than the original position of T in (5b), the subject position SpecT′ is higher
than the object position in (10b) and SpecC′ is higher than the object position
in (30). So movement appears to always move constituents ‘upward’ in the
tree; this is an important generalisation which we will make more precise in
later chapters.
The three operations differ both in what is moved and in the target of the
movement. Head-movement moves heads, and in fact moves the head to another,
higher head position; again, this can be easily seen in (5b). NP-movement moves
NPs into subject position; in fact it moves NPs from one grammatical-function
position (object) to another (subject). This is why operations like passive are
often described in purely grammatical-function terms (‘passives turn objects into
subjects’). As usual, we take the grammatical-function terminology as a useful
shorthand for the correct, phrase-structural description. Finally, wh-movement
moves a wh-XP into SpecCWh′.
In the above description, we can note an interesting match between the start-
ing position and the target position of movement: head-movement moves heads
from one head position to another; NP-movement moves NPs from one gram-
matical-function position to another and wh-movement moves wh-XPs from
one wh-position to another (recall that in-situ wh-phrases in echo questions and
multiple questions show us that any NP position can host a wh-phrase).
Putting the last two observations together, we can give a general characterisa-
tion of movement rules as follows:
(58) Movement substitutes an element of type X into a higher position of type X,
where X ranges over heads, wh-phrases and grammatical-function NPs.
Although somewhat rough and ready, and certainly causing us to wonder why
heads, grammatical-function NPs and wh-phrases might fall together in this
way, (58) contains the germ of a very important set of generalisations, which we
investigate in full in Volume II.
Compare the fully grammatical (60), with ‘local’ T-to-C movement in the main
clause:
(60) Can you (can) say [ that Cambridge will flood ] ?
could still ask about non-subject wh-phrases. Here we will see that the answer is
definitely negative, and that the limits to wh-movement are quite significant and
very interesting from a theoretical perspective.
Let us begin by looking at NPs which contain a possessor NP, such as my
guitar or John’s cat. The structure of these NPs is shown in (63) (here the sub-
script numerals 1 and 2 are there just to keep the two NPs distinct; they have no
theoretical significance):
(63) NP1
NP2 N′1
N′2 N1
cat
N2 guitar
John’s
my
Here we see that the possessor NP, NP2, occupies the Specifier of the possessee
NP, NP1. Everything is completely in conformity with X′-theory. Since the pos-
sessor is an NP, it can be a wh-phrase, giving whose cat, for example, with the
same structure as (63):
(64) NP1
NP2 N′1
N′2 N1
cat
N2
whose
NP2 looks like a wh-phrase in good standing. But if we try to apply wh-move-
ment to it we run into ungrammaticality:
(65) * [NP2 Whose ] did you feed [NP1 t cat ] ?
NP2 is unable to move alone. Instead, the whole of NP1 must be pied-piped:
(66) [NP1 Whose cat ] did you feed t ?
Since, in (64), NP2 whose is on the left branch (i.e. the Specifier) of NP1, the LBC
prevents it from moving alone. This accounts for the ungrammaticality of (65).
Pied-piping NP1 as in (66) is unproblematic, since the entire object is not on a
left branch; being the complement of V, it is on the right branch inside V′. So
the LBC doesn’t affect object movement. (In order to allow for movement of
wh-subjects as in (31b) we have to exempt TP from (67).)
The LBC extends beyond possessors to APs. Consider an example like (68),
which we have seen several times:
(68) [AP How angry ] is Alex t ?
Here we have the wh-AP how angry, which we can analyse as containing the
Specifier how and the A′ (and A) angry. The LBC predicts that Specifiers, being
on a left branch cannot undergo wh-movement alone. This prediction is correct
for how angry, as (69) shows:
(69) *How is Alex [AP t angry ] ?
clause; there is a gap in the subject position corresponding to story here (this
indicates that relative clauses involve movement, a matter we haven’t gone into
here, with string-vacuous movement of which from SpecTP to SpecCP here; see
the discussion of subject wh-questions in Section 5.4). Compare (72b) with the
‘reduced-relative’ NP in (73), where there is no CP inside the NP:
(73) Which hobbit did you read [NP a story about t ] ?
In addition to the LBC, the CNPC and the wh-island constraint, there are several
other island constraints. But this is not the place for a thorough discussion, still
less an analysis, of the phenomenon. Here we can simply observe that, despite
appearing to be unbounded, the operation of wh-movement is not unlimited: it is
subject to island constraints. Subject wh-movement is also subject to the C-trace
generalisation in (38).
So we see that wh-movement combines two contrasting properties: it appears
to be unbounded and yet is subject to island constraints. Here we have con-
centrated on wh-movement in interrogatives, although it is also found in other
constructions, notably relative clauses, as in (74):
(74) a. the hobbit which you read a story about t
b. the hobbit [which Mary believes [ (that) Fred knows [ (that) I asserted [(that)
John saw t at the bottom of the garden ]]]]
c. *the hobbit which you read [NP a story [CP that a wizard met t ]]
Example (74a) illustrates a basic relative clause, with the wh-relative pronoun
which and an associated trace/copy as marked. We can take it that which is in
the SpecC′ of the CP which modifies hobbit. Example (74b) shows that rela-
tive-clause formation is unbounded, and (74c) that it is subject to the CNPC.
So which-movement, forming the relative clause here, is clearly a further case
of wh-movement. Presumably this case of wh-movement is not associated with
CQ, as it does not form an interrogative; we may want to define Crel for relative
clauses, or perhaps generalize the notion of Cwh beyond interrogatives. These are
among many issues we have not explored here.
English also allows relative clauses where there is no apparent wh-relative
pronoun. These can be marked with the complementizer that, which, as usual,
is optional. Where that is optional it appears that there is no overt marker of the
relative clause:
(75) a. the hobbit (that) you read a story about t
b. the hobbit [ (that) Mary believes [ (that) Fred knows [ (that) I asserted [(that)
John saw t at the bottom of the garden ]]]]
c. *the hobbit (that) you read [NP a story [CP that a wizard met t ]]
5.7 Conclusion
In this chapter we have seen that movement rules are necessary,
alongside the PS-rules which generate the X′-template. We saw in Section 5.6
that although it is technically possible to complicate PS-rules in such a way that
they can generate the structures that result from movement, the complications
involved change the nature of PS-rules, in particular in that they lose their fun-
damentally local character.
We saw that there are three types of movement: head-movement, NP-movement
and wh-movement. These three types of movement have a number of properties
in common: they all involve substitution of a category into a position generated
by the PS-rules (hence they manipulate, rather than generate, structure), they all
copy the ‘moved’ element, leaving the copy in the starting position to be deleted
by the phonology, and they all involve a ‘like-for-like’ substitution, captured by
the rough generalisation in (58), repeated here:
(58) Movement substitutes an element of type X into a higher position of type X,
where X ranges over heads, wh-phrases and grammatical-function NPs.
CQ′
CQ TP
T NP T′
will Cambridge
T VP
(will )
V
flood
b. NP-movement:
TP
NP T′
John
T VP
was
V NP
interviewed (John )
c. Wh-movement:
CQ, WhP
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
We now have quite a good picture of the basics of the theory of syntax. However,
many questions have been left open here. One of the most important ones con-
cerns the ‘upward’ nature of movement, also captured by our generalisation in
(58). In the next chapter, we will look at a different set of rules which account for
certain kinds of meaning dependencies among elements; this will give us a way
of making the ‘upwards’ aspect of movement both more precise and integrating
it into the theory properly.
Exercises
1. Provide the trees for the following sentences. Indicate the gap (t) left
by wh-movement in each case:
a. Which car did Mary buy?
b. What did Mary buy?
c. With which knife did Mary cut the bread?
d. Which knife did Mary [V′ [V′ cut the bread ] [PP with t ]]?
e. Which books did Mary say [CP (that) she will buy t ]?
f. Who did Mary say [CP (that) [TP t sold her the books ]?
g. Which knife did Mary say that she cut the bread with?
h. How did Mary say that she will solve this problem?
i. I remembered [ who Ed said [ t had visited Wanda ]].
j. Ed asked [CP who [TP t knew [CP who [TP Greg had visited
t ]]]].
k. I will ask [CP which books [TP PRO to buy t ]].
2. Provide the trees for the following sentences, both the grammatical
and the ungrammatical ones. Indicate the gap (t) left by wh-move-
ment in each case:
a. I fed [NP [NP my [N′ neighbour’s ]] [N′ cat ]].
b. Which neighbour’s cat did you feed t ?
c. I wonder [CP [NP which neighbour’s cat ] [TP he fed t ]].
d. [CP [NP Which neighbour’s cat ] [C did ] [TP you say [CP
(that) [TP you fed t ]]]] ?
e. *[CP [NP Which neighbour’s ] did you feed [NP t cat ] ?
f. *Which did you say you fed [NP [NP t neighbour’s ] cat ] ?
What constraint do (e) and (f) violate?
3. The following examples feature another island constraint:
a. *Who did [ pictures of t ] annoy John?
b. *Who did they say that [TP [TP PRO to talk to t ] [T is ] [AP
forbidden ]]?
c. *Who was [TP [CP that John spoke to t ] twas a shock?
Can you formulate this constraint (hint: try to generalize the Left Branch
Constraint).
Further Reading
Adger, D. 2003. Core Syntax. Oxford: Oxford University Press, Chapters 9 and 10.
Carnie, A. 2011. Modern Syntax: A Coursebook. Cambridge: Cambridge University
Press, Chapter 4.
In the previous chapter we made the distinction between the rules which
generate structure, PS-rules, including those which generate the X′-tem-
plate, and movement rules, which map phrase markers into phrase markers.
In this section I will introduce a third type of rule, interpretative rules,
which specify the relationships between constituents in a hierarchical struc-
ture generated by PS-rules. These rules do not generate structure, so they are
clearly distinct from PS-rules. Furthermore, they do not appear to involve
movement. As the name suggests, though, they are relevant for the seman-
tic interpretation of relations between constituents, including, as we will
see, the relations between moved elements and their copies. The particular
kind of dependency we will be concerned with involves binding of various
types; the exact nature of binding will be elucidated as we proceed. We will
see that the account of binding relations we arrive at provides important
support for the hierarchical organisation of constituents of the kind we have
been developing up to now.
130
In that case, (1) is interpreted to mean ‘John hopes that he, John, will win.’
Alternatively, he in (1) could refer to any contextually salient male individual
distinct from John. In the latter case, we say that the pronoun has a non-linguistic
antecedent, while in the interpretation where he stands for John, the pronoun has
a linguistic antecedent.
As we will see, binding theory makes an important distinction between pro-
nouns and anaphors. This distinction is not made in traditional grammar, where
anaphors are described as a particular type of pronoun: reflexive pronouns such
as herself, himself, myself, yourself, yourselves, etc., and reciprocal pronouns
such as each other. On the other hand, binding theory groups reflexives and
reciprocals together as a class of anaphors distinct from normal, simple pro-
nouns such as I/me, he/him, she/her, we/us, you, they/them. The reasons for mak-
ing this distinction will become clear below.
Both pronouns and anaphors are distinct from non-pronominal NPs, notably
R(eferring)-expressions, NPs with the intrinsic semantics not dependent on the
interpretation of some other NP in the linguistic expression in which they occur,
e.g. John, Mary, the idiot, the pet shop owner, etc.
The goal of binding theory is to account for the distributions and interpretations
of each class of NP. A simple way to do this is to adopt the coindexing convention:
whenever two NPs have the same index, one is dependent on the other for its inter-
pretation. The indices are subscript letters of the alphabet starting from ‘i’: i, j, k …
This is linked to the idea that every distinct individual in our mental representation of
the world has an integer which identifies it: John23, the man next door, the pet shop
boy666, etc. However, this last point has little bearing on what follows. Adopting the
coindexing convention, we can repeat (1) as (2):
(2) Johni hopes that hei will win.
Unlike (1), (2) is unambiguous: the coindexing of John and he means that John
is the antecedent of he, i.e. (2) only has the first of the two interpretations of (1)
discussed above.
So binding theory should be able to explain why the following sentences, with
the interpretations indicated by the indices, are ungrammatical:
(3) a. *Herself thinks that it’s going to rain.
b. *A good friend of Maryi’s has declared herselfi bankrupt.
c. *Johni likes himi.
d. *Hei likes Johni.
Here we will see how binding theory deals with these examples in terms of the
three classes of NP. We will look first at anaphors, then at pronouns and finally
at R-expressions.
6.3 Anaphors
So let’s begin with anaphors, concentrating mainly on reflexives.
Consider first (5):
(5) Goona betrayed herself.
Comparing the grammatical (6), where Goona must be the antecedent of herself,
with the ungrammatical (7), where herself has no antecedent, we can conclude
that herself needs an antecedent.
Now consider (8):
(8) *Herselfi betrayed Goonai.
Here, given the coindexing, herself has an antecedent, namely Goona. The sen-
tence is nonetheless clearly ungrammatical. How can we account for this?
The structure of both (6) and (8) is (9):
(9) TP
NP T′
N′ T VP
N′ V′
Goonai
Herselfi V NP
betrayed herselfi
Goonai
We will now review several hypotheses concerning what might be the best anal-
ysis of the required structural relation between the antecedent of the reflexive
and the reflexive. Here is the first hypothesis:
(10) Hypothesis One: The antecedent of an anaphor must be a subject.
Hypothesis One will correctly distinguish the grammatical (6) from the ungram-
matical (8). In (6), the antecedent of the reflexive, Goona, is a subject and in (8)
it is not.
However, in English non-subjects can be the antecedents of reflexives, as in:
(11) I asked Johni about himselfi.
This example is grammatical, but John is not a subject, although it can still be
the antecedent for himself. So this disconfirms Hypothesis One. Compare also:
(12) *I asked himself about John.
Here I is the subject, but it still cannot be the antecedent for himself. This is
because the antecedence relation requires agreement in number, gender and per-
son. First-person I therefore cannot bind third-person himself (contrast Ii asked
myselfi about John where myself and I agree for first person and so I can be the
antecedent of myself). We will see further examples of the agreement condition
on antecedence below. Hence subjecthood, on its own, is neither necessary nor
sufficient for being a well-formed antecedent for a reflexive.
To a good approximation, the structure of (11) and (12) is as in (13):
(13) TP
NP T′
N′ T VP
(past)
N V NP PP
/ asked Johni
*/ asked himself P′
P NP
about himselfi
about John
Hypothesis Two accounts for the grammaticality of (6) and (11), where the
antecedent precedes the anaphor, and for the ungrammaticality of (8) and (12),
where anaphor precedes the antecedent. In fact, precedence may be a necessary
condition for a well-formed antecedent antecedent–anaphor relation, but it is
not a sufficient one, as there are examples where the antecedent precedes the
anaphor which are nonetheless ungrammatical, such as the following:
(15) a. *Johni’s friends talked about himselfi.
b. *Friends of Johni’s talked about himselfi
Here we must pay careful attention to the indices. The coindexation in (15a)
indicates that John is the intended antecedent of himself, and this is impossible.
We can assign a grammatical interpretation to the string in (15a) if we inter-
pret John’s friends as the antecedent of the plural anaphor themselves; here we
see the requirement that the antecedent and anaphor must agree for number in
action. But this would correspond to the coindexing in (16):
(16) [ John’s friends ]i talked about themselvesi.
Here the index i is associated with the whole NP John’s friends (incidentally,
this shows that indices are associated with NPs, not with Nouns). Similar rea-
soning applies to (15b): with the indexing given, John cannot be the antecedent
of himself. John’s friends can be, if himself is changed to themselves as just men-
tioned, but in that case the index is associated to that whole NP, and not to John.
Keeping these points in mind, we can get a very important clue to what is
going on if we look at the structure of (15a) (here, as in our discussion of pos-
sessive NPs in Section 5.3, the numerical subscripts on the NPs serve merely to
keep them distinct; they are not to be taken as indices):
(17) TP
NP1 T′
NP2 N1′ T VP
N2′ N1 V PP
friends talked
N2 P′
John’s
P NP3
about
N3′
N3
themselves
Here, if NP1, John’s friends, is the antecedent of themselves, the relation is struc-
turally well-formed, as we pointed out above. On the other, NP2, John’s, can-
not be the antecedent. What seems to make the difference here is the relation
between the antecedent and the anaphor in terms of hierarchical structure. We
could therefore tentatively formulate Hypothesis Three, as follows:
(18) Hypothesis Three (to be revised): An antecedent must be structurally higher
than the anaphor which it binds.
Since NP1 is evidently structurally higher (higher up in the tree) in (17) than
the anaphor NP3, it is able to be its antecedent. NP2, John’s, on the other hand,
is seemingly not in the requisite structural relation with the anaphor for this to
be possible. Clearly, we need to make this intuitive notion of ‘structural height’
more precise.
The formal indicator of relative structural height is the relation of
constituent-command, or c-command. This structural relation plays a central
role in binding theory and, as we will see in Volume II, many other dependen-
cies. The definition of c-command is given in (19):
(19) A node X c-commands another node Y if and only if:
i. X does not dominate Y;
ii. Y does not dominate X;
iii. The first branching node dominating X dominates Y.
Let us revisit some of the structures we looked at earlier and apply this definition
to the various antecedent–anaphor pairs. First, (9):
(9) TP
NP1 T′
N1′ T VP
[Past]
N1 V′
Goonai
V NP2
betrayed herselfi
Now let’s ask if the object NP c-commands the subject NP. As before the first
two clauses of the definition in (19) are met, in that neither NP dominates the
other. Turning to (19iii), though, we see that the first branching node dominating
the object NP is V′, and this node does not dominate the subject NP. Therefore
the object NP does not c-command the subject NP in (9), hence in the ungram-
matical (8), Goona cannot be the antecedent of herself.
Next, the structure in (13) (annotated with numerical subscripts):
(13) TP
NP1 T′
N1′ T VP
[Past]
N1 V NP2 PP
/ asked Johni
*/ asked himself P′
P NP3
about himselfi
about John
Here the question is whether NP2 c-commands NP3. Since it is easy to see nei-
ther NP dominates the other, the question again is whether the first branching
node dominating NP2, VP, dominates NP3. The answer is clearly yes, and so NP2
c-commands NP3 and is able to be the antecedent for NP3. Conversely, since the
first branching node dominating NP3 is P′ and P′ does not dominate NP2, NP3
cannot be the antecedent for NP2; this accounts for the ungrammaticality of (12).
Finally, let us consider (17) in the light of c-command as defined in (19):
(17) TP
NP1 T′
NP2 N1′ T VP
[Past]
N2′ N1 V PP
friends talked
N2 P′
John’s
P NP3
about
N3′
N3
themselves
Here, since the first branching node dominating NP1 is TP, which as the root node
dominates everything, NP1 c-commands NP3. This is why John’s friends can be
the antecedent of themselves in (16). On the other hand, the first branching node
dominating NP2 is NP1, which does not dominate NP3. Therefore John’s cannot
be the antecedent of themselves.
NP1 T′
N1′ T VP
[Past]
N1 V CP
*John said
C TP
that
NP2 T′
himself T VP
[Past]
V
left
Here we can easily see that NP1 c-commands NP2, since, once again, the
first branching node dominating NP1 is TP, the root node, which dominates
everything. With only Hypothesis Three, then, John should be able to be the
antecedent for himself.
We can correctly rule out (22) and retain all our earlier results if we restate
Hypothesis Three as follows:
(23) Hypothesis Four: An anaphor must be c-commanded by a clause-mate
antecedent.
By this definition, we see that John and himself are not clause-mates in (22) since
the lower CP dominates himself without dominating John. In (22), then, himself is
We can account for the ungrammaticality of (25a) exactly as for (21): John and
himself are not clause-mates by the definition in (24) and therefore John cannot
be the antecedent of himself, given Hypothesis Four. But where does this leave
(25b)? At a minimum, we have to say that, although XP looks like a clause in
that it contains a subject and a predicate (at least from a semantic perspective), it
does not qualify as a clause for the definition of clause-mates in (24). This would
be the case if XP is not a CP. So instead, we take it to be TP, where infinitival
to is in T. This is supported by the fact that verbs like believe cannot take CWh
complements, as we noted in Chapter 5. So we have contrasts of the following
kind with verbs like wonder:
(26) a. John wondered [CP whether to leave ].
b. *John believed [ whether the world to be flat ].
NP T′
N′ T VP
[Past]
N V TP
John believes
NP T′
himself
T VP
to
V NP
be
D N′
a
N
genius
Since there are either no CPs dominating John and himself here, or there is
just a root CP which is not shown here, there is no CP which dominates him-
self but does not dominate John. Therefore, John and himself are clause-mates.
Therefore John can be the antecedent of himself. The contrast in (25) is now
accounted for.
NP T′
N′ T VP
N V′
John
V TP
believes
NP T′
Mary
T VP
to
V NP
like himself
Unless we posit a CP in some strange place inside the embedded TP, John and
himself will be clause-mates by the definition in (24). It also seems clear that
John c-commands himself, since John is the main-clause subject in main-clause
SpecT′, and so the first branching node dominating John is the root TP, which
dominates everything. It seems we need a further condition in addition to clause-
mates and c-command.
Hypothesis Five can account for (28) while leaving all the earlier results
intact:
(30) Hypothesis Five: An anaphor must be c-commanded by a clause-mate
antecedent such that no other NP intervenes between the antecedent and
anaphor.
Applying Hypothesis Five to (28), we see that the subject of the infinitival TP, Mary,
c-commands the anaphor himself and is c-commanded by the antecedent John. Mary
therefore intervenes between the antecedent and the anaphor, and so the relation
cannot hold. In contrast, there is no such intervener in (25b/27) and so here the ante-
cedent–anaphor relation is well-formed, as it was under Hypothesis Four.
However, the notion of intervener in (31) is too strong, in that it rules out cer-
tain well-formed antecedent–anaphor relations, as in (32) for example:
NP T′
N′ T VP
[Past]
N V NP PP
John told Mary
P NP
about himself
The crucial cases illustrating how this definition of BD works are the biclausal
ones in (25) and (28), so let us go through those once more:
(25) a. *Johni thought [CP that himselfi was a genius ].
b. Johni believes [XP himselfi to be a genius ].
The BD for himself in (25a) is the embedded TP as this is the smallest XP con-
taining a finite T (note that this TP does not contain a subject distinct from the
anaphor). The antecedent John is outside this BD and hence Hypothesis Six is not
satisfied and the relation does not hold. The sentence is therefore ungrammatical.
The BD for himself in (25b) is the root TP, as this is the smallest category con-
taining both a subject distinct from the anaphor and a finite T. The embedded TP
is non-finite and the subject is the anaphor itself, and therefore not distinct from
the anaphor. Since that BD contains the antecedent John, the relation is well-
formed and the sentence is grammatical. This account has the advantage over the
clause-mate-based definition relying on the notion of clause-mate in (24), since
we are no longer required to consider that XP is not CP.
The BD for himself in (28) is the embedded TP. Although non-finite, this is
the smallest XP containing a subject, Mary, distinct from the anaphor. The ante-
cedent John is outside this BD, and hence Hypothesis Six is not satisfied and the
relation does not hold. The sentence is therefore ungrammatical.
Hypothesis Six therefore accounts for all the examples we have seen in this
section. We can now add a general definition of binding, which will be relevant
in the rest of this chapter, and as a consequence further simplify Hypothesis Six.
The definition of binding is (38):
(38) X binds Y if and only if X c-commands Y and X is coindexed with Y.
As long as antecedents are coindexed with their anaphors, all the cases of well-
formed antecedent–anaphor relations we have seen involve the antecedent bind-
ing the anaphor by this definition. We can now restate Hypothesis Six in simpler
form as follows:
(39) An anaphor must be bound in its binding domain.
The statement in (39) captures all the facts we have seen in this section, both the
well-formed and ill-formed antecedent–anaphor relations.
In the next sections, we will see that the definition of binding generalises to
pronouns and R-expressions. We will also see that the notion of binding domain
is relevant for capturing important aspects of the distribution and interpretation
of pronouns, which is our next topic.
6.4 Pronouns
Remember that the definition of pronouns we are working with here
excludes reflexives (himself, etc.) and reciprocals (e.g. each other); those are
anaphors and the subject of the preceding section.
The first observation we can make about pronouns is that they do not require
a linguistic antecedent (something we already noted in our discussion of (1)).
Hence both examples in (40) grammatical:
(40) a. He grew some vegetables.
b. Vurdy said he grew some vegetables.
In (41a), we see that the antecedent of the pronoun doesn’t have to c-command
the pronoun; again this contrasts with the behaviour of anaphors seen in (15). In
(41b), we see that the antecedent of the pronoun can be outside the pronoun’s
binding domain, unlike what we saw with anaphors in examples like (28). Here
we are extending the idea of binding domain to pronouns. Binding domains
were defined in (37) explicitly for anaphors; here we generalise that definition
so that it applies to pronouns:
(42) The BD for NPi is the smallest XP containing NPi and either:
a. a subject (distinct from NPi) or
b. a finite T.
This definition applies to any (indexed) NP; therefore it can still apply to anaphors
in exactly the way we saw at the end of the previous section, but now it can also
apply to pronouns (and in principle to other NPs). To repeat, then, (41b) shows
that a pronoun, unlike an anaphor, can be bound from outside its BD.
In fact, while anaphors must be in a local domain with their antecedents, as
we saw, pronouns must not be too close to their antecedents, as the following
examples show:
(43) a. Vurdyi hates him*i/j.
b. Vurdyi believes [ him*i/j to be the best ].
c. Vurdyi told Goona about him*i/j.
The definition of binding domain is the generalised one in (42). The definition of
binding is (38), repeated here:
(38) X binds Y if and only if X c-commands Y and X is co-indexed with Y.
6.5 R-expressions
In Section 6.2 we defined R-expressions as NPs with the intrinsic
semantic capacity to pick out a referent, a specific individual from the domain
of discourse (Goona, Vurdy, John, Mary, the Swedish Chef, the pet shop boy,
etc.). Exactly how NPs refer, and what the relation of reference is, are important
and difficult questions in semantic theory. For our purposes, it is sufficient to
observe that R-expressions have an intrinsic referential capacity that pronouns
and anaphors lack; at the very minimum, there is an obvious sense in which an
NP like John or that man does not require a relation with an antecedent of some
kind in order for its reference to be determined.
In terms of the notion of freedom introduced at the end of the previous sec-
tion (‘free’ simply means ‘not bound’, with binding defined as in (38)), we might
therefore expect that R-expressions are able to be free. This is, perhaps unsurpris-
ingly, true. But the condition on R-expressions is stronger: R-expressions must
always be free. What this means is that R-expressions can never have a c-com-
manding, coindexed antecedent. There is no locality condition associated with this
constraint; the notion of BD plays no role. If an R-expression is bound by a pronoun,
even at a distance, ungrammaticality results, as examples like the following show:
(47) a. *Hei thought [CP that [the idiot]i must be right ]].
b. *Hei said that John thinks that Mary believes that Nigeli is invited.
In (47a), the BD for the the idiot is the embedded TP, as this is the minimal XP
containing that NP and a finite T, but binding by he in the main clause is nonethe-
less not allowed. If we replace the idiot with a coindexed pronoun, the example
becomes grammatical; if we replace it with an anaphor it remains ungrammat-
ical. These outcomes are predicted by (46) and (39) respectively. Here we can
see the utility of the generalised definition of BD in (42). In (47b) Nigel is three
clauses away from he; nonetheless he cannot bind Nigel. This clearly illustrates
that R-expressions are subject to an absolute freedom requirement.
If R-expressions must be free in the technical sense of not bound, this means
that they can have antecedents; the requirement is that such antecedents must
not c-command the R-expression. So we find well-formed examples such as the
following:
(48) a. [ Hisi mother ] loves Johni.
b. [ The fact that hei failed the exam ] bothered Vurdyi.
c. [ The hobbit hei befriended ] trusted Gandalfi.
In all of these cases, the pronoun is embedded inside a subject constituent which
it cannot c-command out of (it is a good exercise to draw the tree diagrams for
these examples and work out the c-command relations). Hence, although there
is coindexation, and therefore coreference, the R-expressions John, Vurdy and
Gandalf are not bound by the pronouns in the sense defined in (38). Therefore
the relation is allowed and the sentences are grammatical with the interpretations
indicated by the coindexation.
The definitions of binding, free and binding domain are repeated below:
(38) X binds Y if and only if X c-commands Y and X is co-indexed with Y.
(42) The binding domain for NPi is the smallest XP containing NPi and either:
a. a subject (distinct from NPi) or
b. a finite T.
The binding principles account for a wide range of facts concerning the dis-
tribution and interpretation of NPs. We can get a glimpse of this by reconsid-
ering the grammatical and ungrammatical examples we introduced in (3) and
(4) above:
(3) a. *Herself thinks that it’s going to rain.
b. *A good friend of Maryi’s has declared herselfi bankrupt.
c. *Johni likes himi.
d. *Hei likes Johni.
(4) a. Johni said that the person who had upset himi should be fired.
b. Maryi has declared herselfi bankrupt.
c. Hisi mother thinks Johni is a genius.
there is no c-command relation between the two and so John is not bound by his,
in conformity with Principle C of (49).
So we see that the binding principles give us a compact account of a wide
range of distributional and interpretational properties of different NP-types.
The binding principles are interpretative rules regulating the dependen-
cies among NPs in structures generated by PS-rules in conformity with
X′-theory, and appear to be quite distinct from movement rules (although
they are sensitive to the operations of movement rules, as we will see in the
next section).
The reference to a ‘higher position’ here reflects the fact that we observed that
movement always appears to be upwards. This can be seen in the following
structures illustrating each kind of movement, again repeated from Chapter 5:
(51) a. Head-movement:
CQP
CQ′
CQ TP
T NP T′
will Cambridge
T VP
(will )
V
flood
b. NP-movement:
TP
NP T′
John
T VP
was
V NP
interviewed (John )
c. Wh-movement:
CQ, WhP
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
We can now observe that the correct notion of ‘upwards’ is defined by c-com-
mand: movement targets a c-commanding position. In each of (51a–c), the first
branching node dominating the target position of movement dominates the
trace/copy in the starting position. This is clear in (51b,c) where the moved
category occupies the Specifier of the root node, so that the first branching node
dominating it is the root node, which dominates everything. But it is also the
case in (51a), where the first branching node dominating CQ, the target of T-to-C
movement, is CQ′, which dominates T.
So we can replace (50) with (52):
If the moved category c-commands its trace, since c-command is part of the
definition of binding, we can ask whether a moved category binds its trace/
copy. In other words, is the moved category coindexed with its trace/copy? For
NP-movement and wh-movement of NPs the answer to this question is clearly
affirmative: if all NPs bear referential indices, then those indices will be copied
as part of the copy/paste operation of movement, so coindexation automatically
results. We may be able to extend the idea to head-movement and wh-movement
of categories other than NP (although admittedly the notion of ‘referential index’
lacks a clear semantic grounding in these cases), but let us concentrate here on
some interesting consequences of the conclusion that wh-NPs and moved NPs
in passives bind their trace/copies.
There are good semantic reasons to think that wh-NPs bind their traces/
copies quite independently of the fact that we can now see that the configura-
tion created by wh-movement creates a binding relation between them given
the definition of binding (c-command + coindexing) in (38). To see this, we
need to look in a little more detail at the semantic properties of certain kinds
of NPs.
We can start by comparing these two sentences:
(54) For every x, if x is a student, then x likes pizza (or ‘if x is in the set of stu-
dents, then x is in the set of pizza-likers’).
The crucial observation is that which is also quantificational rather than refer-
ential: the answer to the question could be ‘none of them’, showing that which
course doesn’t have to pick anything out in the domain of discourse. Like every
student, which course applies to a set of entities (courses) and asks for the iden-
tity of a member of that set such that Goona likes it. In quasi-logic:
(56) For which x, x a course, does Goona like x ?
As we have seen, pronouns don’t have to have linguistic antecedents, so she here
could refer to some contextually given female individual distinct from Goona.
Equally, she can be coindexed with and therefore bound by Goona (since Goona
c-commands she):
(58) Whoi did Goonaj say shej likes ti ?
The question in (59) cannot have the interpretation that she is the antecedent of
who or its trace/copy. So the trace/copy cannot be an anaphor subject to Principle
A, since it would be bound by shei here, but that interpretation is unavailable.
This leaves no alternative to Principle C.
Let us now look at the interpretation of (59). Since she is now coindexed with
who, it is bound by who. Pronouns bound by quantifiers are semantically inter-
preted as variables (more on this below), so (59) has the interpretation in (60)
(taking who to mean ‘for which x, x a person’):
(60) For which x, x a person, Goona said x likes x
It is clear, though, that (59) does not have that meaning. So what is the cause
of the ungrammaticality of (59)? The pronoun she does not violate Principle
B, as who is clearly outside its binding domain (the embedded TP). Therefore
it must be the trace/copy. The trace/copy’s binding domain is the embed-
ded TP (the minimal XP containing the trace/copy, a subject and a finite T),
which also contains she. So we could conclude that the trace/copy is subject
to Principle B, which would then be violated here (and not, for example, in
(58)).
However, if we swap the positions of Goona and she in (59), and maintain the
indexing, ungrammaticality results again:
(62) *Whoi did shei say Goonaj likes ti ?
Here again, the pronoun her, since it is bound by who, is interpreted as a variable
bound by it. So (64) has the interpretation:
(65) For which x, x a person, x′s mother loves x
Here the only interpretation available seems to be that where the man thinks that
the set of students as a whole are the best; the bound-variable interpretation (‘the
man who talked to every x, x a student, thinks x is the best’) is not available.
This is because every student does not c-command they here.
Second, the bound pronoun obeys Principle B, as we would expect, since it
is A-bound. Hence (69a) is ungrammatical on any interpretation, while (69b)
shows that anaphors can also have bound-variable interpretations provided
Principle A is satisfied:
(69) a. *Every boyi admires himi.
b. Every boyi admires himselfi.
of the binding principles applies to them. Here the answer seems clear from a
simple passive sentence like (51b), repeated here:
(51b) Johni was interviewed ti.
The trace/copy is clearly in the same binding domain as John, and the sentence
is grammatical. Therefore the trace/copy must be subject to Principle A. If the
trace/copy were subject to either Principle B or Principle C, the sentence would
be ungrammatical. This conclusion is supported by examples like those in (71):
(71) a. *John was believed that (John) was bankrupt.
b. John was believed (John) to be bankrupt.
c. *John was believed Mary to like (John).
6.8 Conclusion
This chapter completes our introduction to the central mechanisms
of the theory of syntax. Now we have seen how structures are generated by
PS-rules following the X′-template, how they can be modified by movement
rules and how the rules of binding theory regulate distributional and interpreta-
tive dependencies of various kinds. Obviously, very many questions have been
left open, or barely addressed at all. Nonetheless, the core ideas and mechanisms
are presented here.
One very major question needs to be addressed: how far are the mechanisms
and phenomena they are intended to account for idiosyncratic to English? We
have concentrated almost exclusively on English up to now, largely for exposi-
tory convenience. But we said at the outset that we are trying to construct a gen-
eral theory of syntactic structure, one that applies in principle to all languages.
It is therefore necessary, now that the basic notions are in place, to try to apply
the approach to some data from other languages. Unsurprisingly, it will emerge
that other languages, to varying extents, are organised differently from English.
So a very important question emerges: how can we best account for syntactic
variation across languages? This will be the topic of the next chapter.
Exercises
1. Give the trees for the following sentences:
a. Sarah likes herself.
b. Mark said that Sarah should promote herself.
c. *Sarah said that Mark should promote herself.
d. *Sarah thinks that herself/himself is the best.
e. Sarah considers Mark to like himself/*herself.
f. Sarah’s brother likes himself/*herself.
Say why the ungrammatical examples are ungrammatical. Replace
the reflexives with the corresponding non-reflexive pronouns, give
your judgements as to the grammaticality and the possible inter-
pretations of the resulting examples, and explain them as far as
you can.
2. In the following examples, take each other to be an NP with the
structure [NP each [N′ [N other ]]. There are two generalisations gov-
erning the distribution of this NP, one from binding theory and
another concerning its intrinsic nature. Try to state both (hint: for the
second one, try to make friend plural in (e) and see what happens):
a. The hippies love each other.
b. *Each other love the hippies.
c. I asked the hippies about each other.
d. *I asked each other about the hippies.
e. *The hippies’ friend talked to each other.
f. *The hippies said that each other had left.
g. The hippies believe each other to be geniuses.
h. *John believes Mary to like each other.
For Discussion
What would that structure suggest about (a) the morpheme -self and (b)
pronouns? Can you think of any support for treating self as an autonomous
morpheme?
Further Reading
Carnie, A. 2013. Syntax: A Generative introduction. Oxford: Blackwell, Chapters 4, 5, 15.
Haegeman, L. & J. Guéron. 1999. English Grammar. Oxford: Blackwell, Section 4.1.
Isac, D. & C. Reiss. 2008. I-Language. Oxford: Oxford University Press, Chapter 8.
Roberts, I. 1997. Comparative Syntax. London: Arnold, Chapter 3.
Sportiche, D., H. Koopman & E. Stabler. (2014). An Introduction to Syntactic Analysis
and Theory. Oxford: Wiley, Chapter 7.
156
The SVO order is ‘basic’ in the sense that although it can be changed by move-
ment rules (e.g. wh-movement could place the object in the first position,
SpecC′), it is the neutral order for expressing simple declarative sentences with-
out any particular emphasis on any constituent.
The structure of (1a) is given in (2):
(2) TP
NP T′
Harry et
Hermione T VP
[Pres]
V NP
follow Filch
Our comparative Occam’s Razor would lead us to assign exactly the same struc-
ture to the French sentence (1b). But consider the next pair of sentences:
(3) a. Harry and Hermione often follow Filch.
b. Harry et Hermione suivent souvent Rusard.
Here we see a small but important difference in word order between the two
languages: while the adverb often precedes the Verb follow in English, the order
of their French counterparts is the opposite: suivent (‘follow’) precedes souvent
(‘often’). Moreover, in both languages the opposite order, i.e. the one found in
the other language, is ungrammatical:
(4) a. *Harry and Hermione follow often Filch.
b. *Harry et Hermione souvent suivent Rusard.
For the English example (3a), it seems reasonable to treat often as an AdvP
adjoined to V′, since it provides extra information about the action of following
Filch. So we can give (3a) the structure in (5):
(5) TP
NP T′
Harry and
Hermione T VP
[Pres]
V′
AdvP V′
often
V NP
follow Filch
What should we say about the French sentence (3b)? We either have to say that
there’s something different about the Verb as compared to English or some-
thing different about the Adverb (we could say both are different, but that
would clearly violate Occam’s Razor). The adverb souvent means the same
thing as English often and modifies the action described by V′ in the same
way; there do not seem to be very good grounds for attributing the difference
between the two languages to this element. On the other hand, the Verb suiv-
ent does differ from English follow in that it is overtly marked for present
tense while follow does not show any overt marking for Tense (this does not
mean that there isn’t an abstract Tense affix; see Section 8.6). French verbs
are also marked to varying degrees for person and number: this Verb has the
third-person plural ending -ent. Forms of this Verb in the third-person plural
in other tenses include suivaient (imperfect, ‘used to follow’), suivirent (past,
‘followed’), suivront (future, ‘will follow’) and suivraient (conditional, ‘would
follow’). As we saw when we compared the fish sentences in English and
Italian in Chapter 2, English is rather poor in overt morphological marking
of Tense, Mood and Agreement, while Italian is quite rich. We can see now
that French is more like Italian in this respect (this isn’t surprising, as French
and Italian are both quite closely related Romance languages). Since we’ve
been assuming all along that Tense information is present on T, we could pro-
pose that the French Verb moves to T. Keeping souvent in exactly the same
V′-adjoined position as in English, this gives us the structure (6) for the French
sentence (3b):
(6) TP
NP T′
Harry et
Hermione T VP
[Pres]
V
suivent V′
AdvP V′
souvent
V NP
(suivent) Rusard
Substituting suivent into T[Pres] may force the two elements to combine as a
single head, hence the correlation between V-to-T movement and morphologi-
cal tense marking.
So we observe a difference between English and French, which we can call
the V-to-T Parameter. This is a parameter defining a specific head-movement
option, whether a given language has V-to-T movement or not. In English, this
parameter is set to negative, so there is no V-to-T head movement. In French, it
is set to positive: there is V-to-T head movement. The difference is not always
apparent, as the examples in (1) show, but it emerges where adverbs like often
are adjoined to V′. It also shows up in negative sentences and various other
environments, but our illustration here will suffice for now. The parametric dif-
ference may well be connected to the richer verbal tense inflection of French as
compared to English.
A further 189 languages are listed as ‘no dominant order’, i.e. either languages
for which a single, ‘basic’ order cannot be discerned on the basis of available
data, or for which more than one order is ‘basic’ depending on various factors.
Leaving aside these languages, we see that 564 out of the remaining 1,187
languages (just over 47.5%) are SOV and 488 (just over 41.1%) are SVO. So
together these two orders make up almost 90% (88.6%) of the world’s lan-
guages. Among the other orders, the 95 VSO languages make up 8% of the total
and just over 70% of the 135 non-SOV, non-SVO languages. The other orders
are very rare, at 2.1% (VOS), 0.92% (OVS) and 0.37% (OSV) of the total.
(9) a. French: V′
V NP
b. Japanese: V′
NP V
X YP
b. Japanese: X′
YP X
On the other hand, (10b) predicts that in Japanese the complement of P should
precede it, i.e. we expect to find Postpositions in Japanese, and we do:
(12) Nihon kara
Japan from
‘from Japan’
In Japanese C should follow TP, i.e. complementisers are expected to follow the
embedded clause:
(16) Bill-ga [CP [TP Mary-ga [VP John-ni [NP sono hon-o ] watasita ] to ] itta
Bill-nom Mary-nom John-dat that book-acc hand-pst that said
‘Bill said that Mary handed that book to John’
preponderance of SOV and SVO orders, although the disharmonic and minority patterns
must of course be accounted for. The parameter in (17) also makes predictions about
the order of Nouns and their complements and of Adjectives and their complements,
which we have not gone into at all here. The parameter-based approach to accounting for
the distributions of some of the major word-order patterns in the world’s, based on the
X′-parameter in (17), certainly shows some initial promise. We will return to this matter
in detail in Volume III, armed with a much more sophisticated approach to word-order
variation.
We have concentrated here on harmonic head-initial or head-final languages, as
defined by (17), taking French and Japanese as our respective case studies. Clearly,
though, there are quite a number of languages with no dominant order, in WALS terms,
or with disharmonic order. In the next section we will look at a well-known case of this
kind: German.
NP T′
Hermann
T VP
[Pres]
V′
NP V
den Brief schreibt
If this is the correct structure, then the question of how (18a) is related to it
immediately arises. In fact, as pointed in the commentary text to WALS Feature
81B, German is verb second rather than SVO in main clauses, while it is clearly
SOV in embedded clauses. We now look at the difference between SVO and V2.
The V2 nature of German main clauses can be seen from the following
examples:
(20) a. Am Tisch schreibt Hermann den Brief.
on table writes Hermann the letter
‘Hermann writes the letter on the table.’
Alongside the subject-verb order in (18a), what we see here is PP-verb in (20a),
wh-phrase verb in (20b), object-verb order in (20c), adverb-auxiliary order in
(20d) and subject-auxiliary order in (20e). In each case, exactly one XP precedes
the finite verb, i.e. the verb occupies the linear second position. Examples (20d)
and (20e), where we have a combination of finite auxiliary and a verbal parti-
ciple, shows that it is the finite element, not strictly speaking the lexical verb,
which occupies second position. Here, the main verb is final, following its
object, and so we have OV order.
Furthermore, the V2 constraint is very strict. Orders where two XPs precede
the finite verb in main clauses are, with certain well-defined exceptions which
we need not go into here, ungrammatical. So the word-for-word counterpart of a
simple English sentence like (21a) is ungrammatical in German, as (21b) shows.
Instead, either the subject or the adverb has to occupy a postverbal position, as
seen in (21c) and (21d):
(21) a. Yesterday John wrote a letter.
b. *Gestern Johann schrieb einen Brief.
c. Gestern schrieb Johann einen Brief.
d. Johann schrieb gestern einen Brief.
NP CQ Wh′
what
CQ Wh TP
T NP T′
did you
T VP
(did)
V NP
catch (what)
Let us assume that wh-movement behaves in the same way in German. Then
the first position in (20b) is SpecC′ and the Verb schreibt is in C. We can think
So we can say that French is just like English is having CQ cause T-to-C
movement. The difference between French and English is that main verbs
like suivent in (23) can raise to T, and from there move on to C (where C has
the right features). English, lacking V-to-T movement, has to use the dummy
auxiliary do here.
Coming back to German, we can say that German resembles French in having
both V-to-T and T-to-C, but it differs from French (and English) in that T-to-C
does not depend on CQ; it seems that any root C-type can trigger T-to-C. In
other words, this property generalizes to all (root) Cs in German, as does the
triggering of XP-movement to SpecC′, as we saw. In this way, we arrive at a full
account of V2.
Turning now to OV order, we can account for this in terms of the OV-VO
parameter in (9) in the previous section, repeated here:
(9) a. French: V′
V NP
b. Japanese: V′
NP V
If we say that German, like Japanese, chooses (9b), then V will follow the
object.
All of this will give us the structure in (24b) for the sentence in (24a):
(24) a. Hermann hat den Brief geschrieben.
Hermann has the letter written
‘Hermann has written the letter.’
b. CP
NP C′
Hermann
C TP
T NP T′
hat (Hermann)
T VP
(hat)
V′
NP V
den Brief geschrieben
To complete the picture, (25a) and (25b) represent the case of V2 where no
auxiliary is present and the main verb raises to T and C:
(25) a. Hermann schreibt den Brief.
Hermann writes the letter
b. CP
NP C′
Hermann
C TP
T NP T′
(Hermann)
V T VP
schreibt
V
(schreibt)
V′
NP V
den Brief (schreibt)
Everything is as in (24), except that the main verb schreibt moves first to T (as
in French, see Section 7.3) and then T-to-C movement raises it to C.
We see in (25b) that German has the order OV in V′. As we mentioned above,
this corresponds to the Japanese choice in (9b). However, German clearly does
not take the general category-neutral option of (10b), repeated here, unlike
Japanese:
(10b) X′
YP X
For example, we can see from our V2 examples and from the structures
in (24b) and (25b) that C precedes TP. This is confirmed by the position of
complementisers like daß (‘that’), which precedes TP, as we saw in (18b),
repeated here:
(18b) Ich weiß [CP daß Hermann den Brief schreibt ].
I know that Hermann the letter writes
In fact, it may be correct that T follows VP, contrary to what is shown in (24b)
and (25b). If auxiliaries such as hat are in T, then the fact they follow the main
verb in non-V2 clauses indicates this. This is illustrated in (26a), with the
structure in (26b):
(26) a. … daß Hermann den Brief geschrieben hat.
… that Hermann the letter written has
‘… that Hermann has written the letter’.
b. CP
C′
C TP
daß
NP T′
Hermann
VP T
hat
V′
NP V
den Brief geschrieben
If we flip the order of VP and T inside T′ in (24b) and (25b), it makes no differ-
ence to the eventual surface word order, since T is in both cases occupied by the
trace/copy of the auxiliary or the verb.
So German is head-final in VP and TP, but head-initial in CP. Furthermore,
German is head-final in A′:
(27) sich seiner Sache sicher
oneself.dat one’s subject.gen certain
‘to be sure about something’
French is like English except it has V-to-T movement; in this respect it is like
German, as we have seen. In the next section, we will look at an example of a
language lacking wh-movement.
Here we have gone into some detail on German syntax, although of course
there is much more to say and we have simplified a couple of points. The pur-
pose of the discussion was to illustrate how various parameters combine to
give us the particular properties of German, such as V2 and OV order non-
V2 clauses. We also saw that German is disharmonic in relation to the Head
Parameter, indicating that there is more to say about word-order parameters than
(17), although (17) does a lot of work in a lot of languages. In general, we see
how the parametric approach can be applied; in principle, it could be applied in
this way to any language.
b. Ni weishenme bu qu ne?
You why not go wh
‘Why aren’t you going?’
C′
TP C
ne
NP T′
ni
T VP
[Pres]
Adv VP
weishenme
Neg VP
bu
V′
qu
This analysis implies that Chinese C follows TP in C′, like its Japanese counter-
part, as we saw in Section 7.4; we mentioned at the end of the previous section
that Chinese has disharmonic word order, though, so this does not imply that
Chinese is fully head-final the way Japanese is. This particle, which is optional,
contrasts with the yes/no particle ma:
(33) a. Ni shuo Zhongwen.
You speak Chinese.
‘You speak Chinese.’
So we could say that, for root clauses, ne optionally realises CWh and ma realises
CQ. This is evidence that C is present, and that it can function, as in English, to
type a clause as an interrogative of one kind or another.
Second, there is evidence from how clauses containing an unmoved (or in-situ)
wh-phrase are interpreted which shows that CWh is present. We saw in Chapter 5
The possible positions for the wh-phrase reflect the c-selection properties of
each verb class, i.e. whether the embedded C is Cwh, Q. Furthermore, since the
presence of a wh-phrase implies an accessible SpecCwh′, where the selected C is
not CWh, as obligatorily in the complement of think and optionally in the com-
plement of remember, the root C is Cwh. Hence the scope of the interrogative
(whether the main or embedded clause counts as a question, i.e. whether the
main or the embedded clause has Cwh) is indicated by wh-movement.
The striking fact is that the corresponding sentences in Chinese have exactly
the same range of possible and impossible interpretations, despite the absence
of wh-movement:
(35) a. Zhangsan xiang-zhidao [CP Lisi mai-le shenme ].
Zhangsan wonder Lisi buy-perf what
‘Zhangsan wonders what Lisi bought.’
Clearly, the wh-phrase shenme is in the same position in all three cases (the
direct object position of the subordinate clause), but the interpretations vary as
a function of the main verb in exactly the same way as in English. This shows
us that Chinese has Cwh. The difference between Chinese and English does
not lie in the interpretation or clause-typing of wh-questions, or in the classes
of verbs that c-select Cwh optionally or obligatorily, but much more narrowly
in whether a Cwh triggers movement to its Specifier. The difference, then, is
7.7 Conclusion
In this chapter, we have taken a first look at how the theory of syntax
developed in the earlier chapters can be made into a cross-linguistically general
theory. This is of course the overall goal of our enterprise, as we want ultimately
to construct an account of the human language faculty (or at least the central
syntactic part of it). Our theory must therefore have universal scope. Although
we may still be some way off that ambitious goal, we can see the general shape
of the approach.
The central idea is that there are universal mechanisms and principles:
X′-theory, the binding principles and movement as an upward copy/paste oper-
ation. These universal features of the language faculty are subject to parametric
variation. Here we saw two types of parameters: one concerning the presence
or absence of particular movement rules (V-to-T, wh-movement) and the Head
Parameter (which may hold in maximally general form as in (17) or, for dishar-
monic languages like German, in a more attenuated form). What the full range
of possible parameters might be is a matter for ongoing research which we will
come back to in the later volumes, particularly Volume III. We can ask one sim-
ple question here though: if there are head-movement parameters (e.g. V-to-T)
and wh-movement parameters (see Section 7.6), then are there NP-movement
parameters? The only case of NP-movement we have seen here is the passive
(see Section 5.2). So even our highly limited look at parameters leads us to ask
whether there are languages without passives, and the answer is positive: WALS
(Map 107A; https://wals.info/feature/107A#2/18.0/148.9) cites 211 languages
lacking passives as opposed to 162 in which passives are found. Maybe, then,
the answer to our question is positive: there is an NP-movement parameter of
some kind.
What we have also seen here, particularly in our discussion of German in
Section 7.5, is that different languages can combine the same (or almost the
In Table 7.2, we see that each language has its own set of parameter settings; this
reinforces the idea that each language takes essentially the same set of building
blocks for a grammar but puts them together in its own distinct way. This is the
key to understanding the unity in diversity that we find when we try to reconcile
cross-linguistic diversity with the idea that there is a Universal Grammar. If we can
multiply Table 7.2 by many orders of magnitude, adding thousands of languages
and hundreds of parameters, then we may be able to get an inductive idea of the
true diversity of the world’s languages, whilst at the same time sticking to a clear
universalist perspective. That is one the goals of parametric comparative syntax.
We have been concentrating on the details of the theory of syntax in the last few
chapters. Now it is time to return to a wider perspective and try to see how everything
we’ve seen (constituent structure, X′-theory, movement, binding, parameters) fits
together and fits with the overall architecture of the language faculty, including
phonology and semantics. That is the goal of the next and final chapter.
Exercises
1. We mentioned in the text that finite verbs precede the negator pas
in French. What do we conclude from the following infinitives in
French?
(i) a. *Ne sembler pas content est une condition pour écrire
des romans.
‘To seem not happy is a condition for writing novels.’
b. *Ne posséder pas de voiture en banlieue rend la vie
difficile.
‘To possess not a car in the suburbs makes life difficult.’
c. Ne pas sembler content est une condition pour écrire des
romans.
‘To not seem happy is a condition for writing novels.’
d. Ne pas posséder de voiture en banlieue rend la vie
difficile.
‘To not possess not a car in the suburbs makes life
difficult.’
Now consider these examples:
(ii) a. N’être pas content est une condition pour écrire des
romans.
‘To not be happy is a condition for writing novels.’
1
S. Vikner (1995), Verb Movement and Expletive Subjects in the Germanic languages, Oxford: Oxford
University Press.
Further Reading
Adger, D. 2003. Core Syntax: A Minimalist Approach. Oxford: Oxford University
Press.
Haegeman, L. & J. Guéron. 1999. English Grammar: A Generative Perspective.
Oxford: Blackwell, Chapter 6.
Larson, R. 2010. Grammar as Science. Cambridge, MA: MIT Press, Unit 28.
Roberts, I. 1997. Comparative Syntax. London: Arnold, Chapter 5.
You should also take a look at the website of the World Atlas of Language Structures:
www.wals.info.
(1) LEXICON
D-STRUCTURE
S-STRUCTURE
179
All of this indicates that it is difficult to predict the precise syntactic require-
ments a verb may have (e.g. NP-complement or not) from its meaning. This
idiosyncratic information is stored in the lexicon, which specifies (among other
things) categorial selection (c-selection) and semantic selection (s-selection)
properties of lexical items. Together, these selection frames specify the argu-
ment structure for each verb, i.e. who’s doing what to whom. For the three verbs
we have been looking at here, then, the c-selection and s-selection properties are
specified as shown in (4):
(4) a. devour (V): __ NP <Agent, Theme>
e.g. John devoured the pie at 5 pm.
These entries give information as to the word’s category (V in each case here),
the c-selection frame, i.e. the syntactic material that must appear in X′ as its
sister, and the s-selection frame, i.e. the thematic roles the word’s arguments
bear. Optional material is in brackets in both the c-selection and the s-selection
specification.
The c-selection frame only specifies the arguments that appear in V′:
the internal arguments. Subjects are not c-selected, since they occupy a
Specifier position, which is always available independently of the specific
properties of any lexical item. General information of this kind, which is not
specific to lexical items, is not specified in the lexicon. Instead, the PS-rules
always generate subject positions. Verbs do differ, however, in the thematic
roles they assign to subjects. Some verbs which describe mental states or
emotions, known as psych-verbs, assign an Experiencer thematic role to
their subjects, e.g. love in (5):
(5) John loves Mary.
This shows that the subject position is not c-selected; it appears even where the
verb imposes no lexical-semantic requirement for its presence at all, i.e. it is
not s-selected either. This is a fundamental difference between subjects/external
arguments and internal arguments: internal arguments appear as a function of
c-selection and s-selection properties of lexical items (i.e. heads in X′ terms),
while subject positions always appear and they may, but do not have to, house
external arguments s-selected by a lexical head. If some verbs do not s-select an
external argument, the convention that the leftmost s-selected thematic role is
assigned to the external argument will not suffice. For most verbs, two conven-
tions will suffice:
(8) a. Agents are always lexically external arguments.
b. Themes are never lexically external arguments.
To these we can now add the s-selection frames for each verb as in (9′):
(9′) a. watch (V) ___ NP <Agent, Theme>
b. rely (V) ___ PP <Agent, Theme>
c. put (V) ___ NP PP <Agent, Theme, Location>
d. say (V) ___CP <Agent, Theme>
Examples like (10b) may be found in poetry or perhaps other registers in which
aspects of word meanings are ‘stretched’ for literary or aesthetic effect, and so it
may not be strictly correct to treat them as ungrammatical. We can in fact regard
them as syntactically well-formed in that they can be generated by X′-theory
(along with whatever movement rules might be necessary). The anomalous
nature of such examples, at least in most registers, is due to a violation of an
s-selection requirement. On the other hand, c-selection violations, as in (11),
give rise to genuine ungrammaticality, i.e. syntactic ill-formedness:
(11) a. *Sincerity delights that the world is round.
b. *Mary relies sincerity.
This is debatable however; rain can take an internal argument as in (15), although
this might be a fixed, idiomatic expression:
(15) It’s raining cats and dogs.
We will leave the very interesting question of the argument structure of meteor-
ological Verbs aside.
Some reason to think that Nouns do not have to have argument structure while
Verbs do comes from a reconsideration the fish sentences from Chapter 1:
(16) FishN fishV.
Here the Noun fish acts like a typical common Noun and appears to have no
argument structure. The Verb fish, on the other hand, presumably has the argu-
ment structure given in (17):
(17) fish (V) __ (NP) <Agent, (Theme)>
So the Verb fish has an Agent external argument and an optional Theme internal
argument, realised as an NP. The latter option is taken in the three-fish example:
(18) Fish fish fish.
The Noun fish, on the other hand, appears to have no argument structure at all.
Perhaps here we are seeing a fundamental difference between Verbs and Nouns.
As we mentioned above, functional categories are also listed in the lexicon.
We can give the lexical entries for some English auxiliaries as follows:
(19) a. can (T): __ VP
b. have (Perf): __ V-enP
b. be (Prog): __ V-ingP
These lexical entries specify the c-selection properties of each auxiliary, includ-
ing the information in (19b), for example, that perfect have requires the head
of the selected VP to be a perfect participle, here roughly indicated as V-en.
Similarly, progressive be requires the selected VP to be a progressive participle
in -ing. (Of course, there are other kinds of have, e.g. the possessive, and other
kinds of be, e.g. the copula; (19) implies that these are actually distinct lexical
items with their own lexical entries.) Modals such as can select a bare V. It is
fairly straightforward to state the c-selection requirements of these auxiliaries,
but what about s-selection? The auxiliaries appear to have semantic content, but
it is not clear that they are associated with thematic roles, at least not thematic
roles of the Agent, Experiencer, Theme (etc.) type. This is a major open question
for our understanding of auxiliaries and perhaps other functional categories.
We also stated that functional features may have their own lexical entries. For
example [Present] and [Past] may have entries like the following:
(20) a. [Present] (T): __ VP
b. [Past] (T): __ VP
These lexical entries simply specify that these features are realised on T and
c-select for VP. As with the auxiliaries in (19), they clearly have semantic con-
tent, but the question of their s-selection properties is a difficult one.
Up to now we have concentrated on the categorial, c-selection and s-selection
specifications in lexical entries. But, as with conventional dictionaries, lexical
entries must also provide phonological and semantic information. For entries like
(6), (13a) and (19a), we can take the phonological specification to simply be a
phonemic representation of the word, e.g. /lʌv/, /pɪkʧɘ(r)/ and /kæn/. But for (20)
we need to specify that these lexical items are realised as affixes. For [Past], for
example, we could perhaps try to specify something like /+d/, with the +-sign
indicating that it is obligatorily realised as an affix. This is adequate for the major-
ity of regular Verbs, it does not take into account the existence of irregular past-
tense forms like sang, went, etc. It is impossible for a lexical item such as [Past],
a member of the category T, to change its phonological form as a function of the
particular lexical item in V. The fact that sing, go and about 150 other English
Verbs have irregular past tenses can only be listed in the lexicon. Therefore, these
Verbs have to be lexically specified for their past-tense forms, roughly as in (21)
for sing:
(21) sing (V): __ (NP, PP) <Agent, Theme, Goal>. PF: /sɪŋ ~ sæŋ[Past]/.
For regular verbs, we can maintain that the PF form of [Past] is /d/. This default
(or elsewhere) realisation of [Past] is blocked by the lexically pre-specified
form for sing. The idea that a more specific realisation of a formative, such
as the particular realisation of [Past] when combined with sing, blocks the
default/elsewhere realisation, is known as disjunctive ordering, and plays a
pervasive role in morphology and phonology. All of this of course presupposes
that V and T combine when T is [Past]. This must be specified in some way
in the lexical entry, perhaps by means of a simple diacritic like ‘+’, as part
of the PF-representation. Since English lacks V-to-T movement, as we saw
in Section 7.3, it must have T-to-V movement, a rare case of a ‘downward’
movement rule (recall that we concluded in Section 6.7 that it is not clear that
head-movement must result in a binding relation, hence downward head-move-
ment, unlike wh- or NP-movement, may be possible). The downward-movement
option applies to lexical items specified as ‘+’ where upward movement is ruled
out by the negative setting of the V-to-T parameter.
Lexical entries must also give a specification of the semantics of lexical items.
It is difficult to illustrate this properly here without providing an introduction
to semantics. For many verbs, a kind of ‘action schema’ related to the thematic
roles can be given. Very approximate ‘action schemata’ for three common
English Verbs are given in (22):
(22) a. say: ‘the Agent causes a propositional Theme to exist by speaking’
b. eat: ‘the Agent consumes the Theme by oral ingestion’
c. give: ‘the Agent causes the Theme to become located at/possessed by the Goal.
(The formula in (23a) says ‘the set of individuals such that they are fish’; a
similar, much more complex formula, can be given for the LF in (23b), but this
is not necessary here. The semantics must also somehow specify the intrinsic
content of ‘fish’ as an aquatic animal or whatever, a matter at least partially
bearing on world knowledge which we can safely leave aside here.) What we see
here is that the Verb fish has a more complex lexical entry than the correspond-
ing Noun, and that the Verb’s lexical entry makes reference to the Noun’s: the
occurrence of the term FISH in the LF for the Verb fish must be connected to that
of the Noun in order to avoid unnecessary redundancy in specifying the meaning
of the concept ‘fish’ (presumably λx [fish (x)]) twice over. As we already hinted,
an intriguing question concerns the extent to which the greater complexity of the
lexical entry of the Verb is due to the fact that it is a Verb. Fully exploring this
question would take us too far afield here.
As a last point, here are the lexical entries for the Italian counterparts to the
English Noun and Verb fish, which we saw in Chapter 1:
(24) a. pesce (N; Gender: Masc): __. PF: /’pεʃe/. LF: λx
[fish (x)]
b. pescare (V; Conj: 1): __ (NP). <Agent, (Theme)>. PF: /pεs’kare/. LF:
HUNT(Agent) [FISH (Theme)]
8.4 D-structure
As we said in Section 8.2, D-structure is a syntactic representation
built by PS-rules in conformity with X′-theory. No movement rules apply in the
derivation of D-structure, however. This is the level where the c- and s-selection
requirements of lexical items are directly represented structurally. But how do
the Lexicon and Syntax ‘communicate’? How is the information regarding c-
and s-selection of the kind we saw in the previous section transmitted from the
lexical entries to the syntactic structures?
This is achieved by the Projection Principle, which can be stated as follows:
(25) All and only featural information that is stored in the Lexicon must be
reflected in the Syntax, at all levels of representation.
Looking at (23a), for example, we can take the effect of the Projection
Principle to mean that the Noun fish can only undergo lexical insertion into
an N-node. Since its c-selection and s-selection frames are null, this Noun
On the other hand, Nouns like student are specified for an optional PP com-
plement headed by of, with an associated Theme thematic role, and Nouns like
book are specified for an optional PP complement headed by about or on:
(27) a. a student (of physics)
b. every book (on/about Ancient Greece)
Adjuncts, such as with large fins in (26a), and Specifiers, e.g. the determiners
a and every here, are not c-selected and hence the Projection Principle neither
requires nor forbids their appearance in syntax. This is not to imply that there
are no constraints on adjuncts or Specifiers: a student of physics with large fins
conjures up a possibly anomalous image, but this is a matter of world knowledge
rather than grammar or syntax (and this, as we saw with (10b), may depend on
register in that such an NP would not necessarily be anomalous in a science-fic-
tion context, for example). One important lexical property of Nouns does relate
to what kind of Specifier they are compatible with: whether they are count or
mass. We introduced this distinction in Section 2.2, pointing out that mass nouns
do not have plurals. The Noun fish can function as both a mass and a count
noun, but this is hard to see with plural marking since, as we saw in Chapter 1,
the usual plural of fish (except in the sense of ‘species of fish’) has a null ending
and so looks the same as the singular; see (21) of Chapter 1. In English, mass
nouns can appear in the singular with no determiner, but count nouns cannot,
while count nouns can appear with the indefinite article but mass nouns cannot.
The Noun fish can appear in either context; compare the obligatorily mass Noun
water and the obligatory count Noun ant in (28):
(28) a. Fish/water is good for you.
b. A fish/an ant is on your plate.
So we see that mass or count nature of a Noun, which should also be stated in
its lexical entry (although we have not seen exactly how), interacts with what
can appear in its Specifier. This, too, is arguably a matter of semantics, since the
distinction has to do with countable instances of something as opposed to undif-
ferentiated quantities of something. As we mentioned in Section 2.2, many mass
nouns can be ‘coerced’ into count readings, including water if one is concerned
with distinguishing different types of water (cf. a fine water from the foothills of
the Alps), but fish cannot really be coerced into giving coherent readings where
its c-selection property is violated in examples like (26).
As we said in the previous section, many common Nouns like cat, dog and
fish lack external arguments. This is implicit in the lexical entry in (23a), since
no Agent or other potential external argument is specified.
This suggests that we may need to modify (23b) as regards the thematical roles
s-selected as external argument. Of course, there are various further possible
additions to be made to (23b), but what we have seen here suffices to illustrate
the role of the Projection Principle.
D-structure, then, is the syntactic level at which lexical items are inserted
into phrase markers generated by X′-theory in conformity with the Projection
Principle. This means that (29b) has the D-structure in (33), where the c- and
s-selection properties of all lexical items (including the functional item [Present],
see (20a)) are satisfied:
(33) TP
NP T′
D N′ T VP
Ø Present
N V′
fish
V NP
fish
D N′
Ø
N
fish
The Projection Principle also guarantees that the external argument [np [d ø ]
[n′ [n fish ]] is interpreted as the Agent and the structurally identical internal
argument is interpreted as the Theme in relation to the action schema in (23b) at
LF. It also ensures that the correct PF representations are associated with each
category, including null representations for D and T. We see that the two oper-
ative constraints at D-structure are X′-theory and the Projection Principle, with
the latter guaranteeing the connection between the lexicon and independently
generated phrase structure.
8.5 S-structure
In terms of the architecture of grammar shown in (1), the next stage
of the derivation is S-structure. As we said above S-structure is derived from
D-structure by movement rules. We introduced movement rules in Chapter 5.
There we saw three types of movement rules: wh-movement, head-movement
and NP-movement. Let us briefly look again at each of these, in the light of
what we have seen in this chapter regarding the lexicon, D-structure and the
Projection Principle.
Let us look first at wh-movement. The structure in (34), corresponding closely
to (30) of Chapter 5, is an example of an S-structure derived by wh-movement:
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
have you
T VP
(have )
V NP
catch (what )
Here we see that the wh-NP what moves from the object position, complement
of V, to SpecCQ,Wh′. The D-structure from which (34) is derived is (35) (here
head-movement of T to C has not taken place either, a point I return to directly):
(35)
CQ, WhP
CQ, Wh′
CQ, Wh TP
NP T′
you
T VP
have
V NP
catch what
As a transitive verb, catch has a lexical entry similar to that of fish in (23b) but
without the specification that the complement NP is optional:
(36) catch(V): __ NP. <Agent, Theme>. PF: /kæʧ/. LF: HUNT(Agent) [
CONCRETE-OBJECT (Theme)]
In (35), what satisfies the c- and s-selection properties that catch imposes on its
internal argument, and the external argument you is the s-selected Agent.
CQ, Wh c-selects finite TP, like any root C. As we saw in Chapter 5, it also has
the property of triggering wh-movement. So, in the derivation from D-structure
to S-structure, the wh-NP what moves to SpecCWh, Q′, as seen in (34). Since the
Projection Principle requires lexical properties to hold at all levels of the gram-
mar (as given in (1)), the c-selection property of catch that it has a complement
NP must be satisfied at S-structure. This is why there must be a trace or copy
of what in that position. We did not take a position on which of these it is in
Chapter 5, but we can now observe that positing a copy seems a simpler option
than positing a trace, at least at S-structure, since the latter involves deleting
the copy of what left by movement and inserting a trace in its place, while the
former may not affect the direct-object position at all. X′-theory generates the
landing site of wh-movement, SpecCQ, Wh′. The S-structure in (34) forms the
input to PF and LF, whose properties in relation to wh-movement we will look
at in the next two sections.
Comparing the S-structure in (34) with the D-structure in (35), we can also
see that head-movement from T to C takes place in the derivation of (34) from
(35). We know from yes/no questions, as already discussed several times in
earlier chapters, that CQ triggers T-to-C movement in main clauses whether the
wh-feature is present on C or not:
(37) a. Will Cambridge flood as a consequence of global warming?
b. Did Cambridge flood after the heavy rain last year?
We can also see that, where there is no other auxiliary present, do appears in T
in main-clause interrogatives and various other environments (e.g. negation and
VP-ellipsis). We can clearly assign will a lexical entry like the one given for can
in (19a), as a T-element c-selecting VP (leaving aside the difficult question of
the s-selection properties of modals, as we mentioned there). Then CQ will trig-
ger T-to-C movement. But in (34) and (35), T is occupied by the feature [Past].
In our discussion of the PF properties of [Past] in Section 8.3, we said that it
must combine with V by the ‘downward’ movement rule T-to-V. The operation
combining T with V is somehow prevented from applying in main-clause inter-
rogatives, and so the meaningless auxiliary did is inserted as the realisation of
[Past]. If T-to-C movement places T in a position too distant from V to allow
the PF T-to-V rule to take place, then this may be why do, in the form of did,
is inserted. This implies that at D-structure and S-structure, T is simply [Past],
with the default PF form /d/ as discussed in Section 8.3. (We will come back to
do-support in more detail in the next section and slightly revise what we have
said here.)
The Projection Principle, X′-theory, wh-movement, head-movement and the
sketch of do-support just given, along with the lexical entry for catch in (36),
together account for the paradigm in (38):
(38) a. What have you caught (what)?
b. You caught a fish.
c. *What have you caught a fish?
d. You caught WHAT?
These selection requirements are clearly satisfied in (39a), where Oprah is the
Agent external argument and John the c-selected NP complement bearing the
Theme thematic role. In the passive (39b) we see the same NP John, with the
same thematic role of Theme, in SpecT′. We account for this by saying that this
NP moves from one position to the other. The relationship can be captured by
the trees in (41) (again repeated from Chapter 5):
(41) a. TP
NP T′
T VP
V NP
interviewed John
b. TP
NP T′
John
T VP
was
V NP
interviewed (John )
for the external argument to be satisfied. In fact, there is evidence that this argu-
ment is present even in ‘short’ passives where the by-phrase is absent. A full
discussion of this evidence would take us too far afield here; perhaps the most
important observation is that in (39b) where the by-phrase is absent, the interpre-
tation of the sentence is ‘someone interviewed John’, where ‘someone’ clearly
functions as the s-selected external argument of interview. So in short passives
the external argument is interpreted as an arbitrary implicit argument, a silent
(or implicit) version of ‘someone’, referring to an arbitrary individual or indi-
viduals. The presence of this argument is presumably required by the Projection
Principle. The precise structural realisation of the arbitrary implicit external
argument is much debated; arguably the simplest approach would be to treat
it as a silent version of the by-phrase, as a silent adjunct which, unlike typical
adjuncts, is not optional because of the Projection Principle.
Here we have illustrated how S-structure is derived from D-structure by three
different kinds of movement rules. The Projection Principle requires lexically
specified c- and s-selection requirements to be satisfied at this level, as at all
others. The principal consequence of this is that copies must be present in the
positions vacated by movement; we also saw that the implicit external argument
of passives must be in some sense present.
Let us now look again at the architecture of the grammar in (1):
(1) LEXICON
D-STRUCTURE
S-STRUCTURE
Here we see that the derivation splits into two paths after S-structure, one path
leading to Phonological Form (PF) and the other to Logical Form (LF). As we
will see in more detail in the next section, PF ultimately converts the representa-
tion of the sentence into a phonological and a phonetic form; this is what we
hear. LF, on the other hand, converts the syntactic representation, which is a
purely formal object, into a representation which can be the input to semantic
interpretation (in standard semantic theory, this means a representation which
can be read as a proposition bearing truth conditions, capable of being eval-
uated as true or false); this is what we understand. It is very important to see
that the level of representation we hear (PF) and the level of representation we
understand (LF) are distinct and that they do not directly interact in that neither
is directly derived from the other. The ‘bridge’ between the two is constituted
by the earlier levels of representation D-structure and S-structure, which we can
collectively designate as the narrow syntax, and the Lexicon. The Projection
Principle plays a vital role too, in that, as we saw, it prevents PF /kæt/ from
corresponding to LF λx [ dog(x) ] (as long as PF and LF specifications in lexical
entries are taken to be featural).
The split in the derivation seen in (1) allows us to account for the ubiquitous
fact that we understand more than we hear; as we have repeatedly emphasised, a
great deal of syntax is covert, hence silent. In terms of the architecture of gram-
mar in (1), aspects of narrow syntax feed into LF representations but are inert
at PF. The opposite is also true: there are syntactic elements which we hear but
which play no role at LF, e.g. the dummy English auxiliary do. Additionally,
phonological and semantic aspects of lexical entries are inert in narrow syntax
but active at their respective interfaces, as we mentioned earlier. We will see
examples of PF/LF mismatches in the next two sections. These examples moti-
vate the split derivation of sentences entailed by the architecture in (1).
NP T′
D N′ T VP
Ø Present
N V′
fish
V NP
fish
D N′
Ø
N
fish
The PF in (42b) is seriously incomplete, in that only the segmental aspects of
the phonological representation are specified; at the very minimum the third
/fɪʃ/ receives greater accentual prominence than the preceding two (this is the
English Nuclear Stress Rule).
Clearly, large parts of the S-structure representation in (42a) have no coun-
terpart in (42b). None of the phrase structure determined by X′-theory appears
to have a PF counterpart, although this may be a little too hasty a conclusion,
as phrase structure information is required in order to determine the Nuclear
Stress, i.e. the greater accentual prominence of the third /fɪʃ/ just noted. Of the
lexical items in (42a), only the occurrences of the noun fish and the verb fish
have any PF realisation, the former twice over. The plural indefinite D and
Present Tense are null, as a matter of lexical specification. Different values of
NP T′
D N′ T VP
a +d
N V′
fish
V NP
fish
D N′
a
N
fish
If T is not raised to C in the narrow syntax, then Present/Past satisfy their affixa-
tion requirement at PF by lowering to V. This gives the following derived struc-
ture for (43a):
(46) TP
NP T′
D N′ T VP
a +d
N V′
fish
V NP
V T D N′
fish +d a
N
fish
Here we see [t +d ], i.e. Past Tense, lowered to V (and in fact adjoined to it).
Various PF rules now apply: the copy of [t +d ] in T is deleted by regular copy
deletion, on which see more below; and the derived [v [v fish ][t +d ]] is realised
(‘spelled out’) as /fɪʃ+d/ and input to the assimilation rule we described above,
converting /d/ to /t/. Phonological realisation of [t +d ], i.e. ‘spell out’ as /d/, is
blocked by disjunctive ordering where the lexical entry of the verb is specified
as having a special past-tense form, as we saw in the case of sing in (21): the
availability of sang (/sæŋ/) as sing+Past overrides singed (/sɪŋd/).
Where CQ attracts T in the narrow syntax, T-to-V lowering cannot take place.
This is presumably because T in C is too distant from V. In that case, the affixal
property of T is satisfied by insertion of do at PF, which combines with +d to
give /dɪd/, again overriding the default form, which would presumably be /
du:d/.
‘Do-support’ is also required in negative contexts. Here it is presumably the
presence of the negation, not/n’t, which prevents T-to-V lowering. It is tempt-
ing to attribute this to linear adjacency, but other adverbs, including negative
adverbs such as never, can intervene linearly between T and V without causing
do-insertion:
(47) a. John did not go on holiday in Bangor.
b. John often went on holiday in Bangor.
c. John never went on holiday in Bangor.
As we saw in our comparison of English and French verb positions in Section 7.3,
often can be analysed as left-adjoined to V′; never is semantically similar to
often in that it modifies the action described by V′, adding intrinsically negative
content. Nonetheless, it does not block T-to-V lowering, as (47c) shows. What
is the generalisation over T-to-C raising and not-insertion which blocks T-to-V
lowering in both cases?
It is clear that, after T-to-C raising, a head, namely the copy of the raised T,
sits in between T in C and V. Similarly, we can treat the clausal negator not as
a head; one piece of evidence for this is that in its contracted form it raises to T,
giving forms such as won’t, can’t, etc., which can then raise further to C giving
sentences like (48):
(48) a. Won’t Vurdy talk to Goona?
b. Can’t you understand this?
NP T′
John
T NegP
+d
Neg VP
not
V NP
interview Oprah
CQ, Wh TP
T NP T′
+d you
T VP
+d
V NP
catch a fish
In both cases, we see that there is a head in between T and V: Neg in (49) and the
copy of T in (50). So we can conclude that T-to-V lowering is blocked if a head
intervenes between T (containing +d) and V; this prevents T lowering from C to
V here. In Volume II, we will see that this is a subcase of a much more general
constraint on movement in general.
Where T is Present, everything is the same as with Past. But there is one
important difference, which shows us that PF is quite an abstract level of rep-
resentation. The PF specification of Present is as a null affix: /+ø/. We see this
clearly in (42), where the present-tense form of the verb fish is /fɪʃ/. The reason
we posit T-to-V lowering here, although it does not affect the surface phonolog-
ical form since the affix is null, is that do-insertion is obligatory in main-clause
interrogatives and in negative clauses:
(51) a. Do fish fish fish?
b. Fish do not fish fish.
Again, disjunctive ordering will force /+z/ in the more specified context where
the features [Person: 3] and [Number: Sg] are present, while /+ø/ is the else-
where form. For English, this choice between a separate Agreement (or Agr)
head or (52) may be somewhat moot. In languages with richer tense and agree-
ment inflection (e.g. Italian, as we saw in Chapter 1), distinguishing Tense and
Agr as separate heads may be advantageous. As we also mentioned in Chapter 1,
it may be desirable to ‘line up’ English and Italian verb forms as much as possi-
ble, which would entail postulating a separate Agr head in English.
The +z realisation of Present behaves exactly like Past +d, both in relation to
do-insertion (applying where T-to-V lowering is blocked by an intervening head
in the contexts in (49) and (50)) and in relation to the phonological assimilation
rule, which changes /z/ to /s/ following a voiced obstruent (cf. knots /nɒts/ vs
nods /nɒdz/) and inserts an epenthetic vowel when the root ends in a coronal
fricative or affricate as in fishes (/fɪʃɪz/). Furthermore, a few verbs show stem
allomorphy with +z: do becomes /dʌz/, not /du:z/, and says is pronounced /sεz/
by many speakers (see Section 2.2).
What we see in the contrasting PF realisations of T (and of D in (42) and (43))
is simply differences in the lexical specifications of functional heads: Present
(aside from third singular) happens to be null, just as fish happens to start with
/f/; these are simply matters of lexical arbitrariness (i.e. why we need the lexi-
con in the first place). Copies resulting from movement are different: here we
have a productive rule of copy deletion, as we have seen in our examples of all
three movement types. Since copy deletion removes phonological content, we
take it to be a PF process. We now have three PF processes: T-to-V lowering,
do-support and copy deletion. In addition we have the morphological processes
of lexically conditioned stem allomorphy forming the irregular past tenses of
verbs like sing and catch, and the regular phonological assimilation rules apply-
ing to the affixes +d and +z. It is important to see that these rules must apply in
a particular order, or ungrammaticality will result.
We can easily see that there is an intrinsic ordering of T-to-V lowering and
copy deletion in that the former creates the context for the latter to apply; this
is known as feeding order. Furthermore, lowering clearly creates the context
for the irregular past-tense forms and for the assimilation rules applying to +d
and +z; it therefore feeds and precedes those rules. Finally, if do-insertion is a
‘repair’ rule which applies when lowering fails in order to satisfy the affixation
requirements of +d, +z and +ø, then it must also follow T-to-V lowering. Do-
support seemingly applies where lowering does not apply; these rules are in
bleeding order as lowering removes the context for do-insertion. Do-support
also prevents the formation of irregular past tenses, and as we saw, do itself has
an irregular past form and so must feed, and therefore precede, that rule. It also
clearly precedes the assimilation rules. So it seems clear that we have the follow-
ing partial ordering of the PF operations we have been considering:
(53) T-to-V lowering > do-insertion > irregular past-tense formation > assimilation
NP T′
D N′ T VP
a +d
N V′
fish
V NP
V T D N′
fish +d a
N
fish
Let us now look again at the result of narrow-syntactic T-to-C raising of +d,
repeated from (50):
(50′) CQP
CQ, Wh TP
T NP T′
+d you
T VP
+d
V NP
catch a fish
(50′) differs from (50) in that the copy of +d in T has not been deleted. If T-to-V
lowering precedes copy-deletion, and T-to-C movement, as a syntactic rule,
precedes all the rules in (53), then (50’) is the representation which is the input
to PF. Following (53), we can in fact apply T-to-V lowering here, giving a form
such as [v [v catch ] +d ] in the V position (see again (46)). If copy-deletion
precedes do-support, the copy of +d in the original T position will disappear and
there will be no head in between the moved T in C and V. However, this is not
a problem since T-to-V lowering has already applied; if T-to-V lowering were
allowed to apply iteratively it could then lower T from C to V, bleeding do-in-
sertion, so we see that this is not a possibility. Copy deletion must in any case
precede the other rules in (53), irregular past-tense formation and assimilation,
as it will then delete the occurrence of +d attached to V and bleed these rules,
guaranteeing that where do is inserted in C the verb always appears uninflected.
So we arrive at the following ordering of the PF-rules:
(54) T-to-V lowering > copy deletion > do-insertion > irregular past-tense for-
mation > assimilation
Evidence from child language is relevant here. It has been observed that children
sometimes ‘repeat’ the auxiliary in inversion contexts at early stages of language
acquisition, producing examples like (55):
(55) a. Can its wheels can spin?
b. Is the steam is hot?
c. Was that was Anna?
We could account for this by saying that, at the relevant stage of first-language
acquisition, children have not fully acquired copy deletion, at least in the T-to-C
context. Very strikingly, the same applies with do-support, examples like (56)
being attested:
(56) Did the kitchen light did flash?
We predict that (56) would arise from failure of copy-deletion of the occurrence
of do+d in the original T position in (50’). This form is then input to past-tense
formation, and did results. Failure of deletion of the lowest copy of T, the one
attached to V, would give rise to examples like (57):
(57) Did the light (did) flashed?
These are also attested in the spontaneous speech of children acquiring English.
A further environment for do-support is VP-ellipsis, as in (58), repeated from
Section 3.4:
(58) John ate the cake and Mary did eat the cake too.
Here, as we argued in Section 3.4, the VP eat the cake is elided. One approach
to English VP-ellipsis is to treat it as PF deletion. We interpret the elided VP
as eat the cake because it is present at LF but absent at PF (a good example of
understanding more than we hear). If VP-ellipsis is a PF operation, we can ask
how it is ordered in relation to the operations we have been looking at here. The
answer is clear from the presence of do-support in (58): ellipsis bleeds T-to-V
lowering by deleting V and thereby feeds do-insertion by creating the context
where do is required to meet the affixal requirement of +d in T. The order of
operations is shown in (59):
(59) a. … and [TP Mary [T′ [T +d ] [VP eat the cake ]]] (S-structure)
b. … and [TP Mary [T′ [T +d ] [VP eat the cake ]]] (VP-ellipsis)
c. … and [TP Mary [T′ [T do+d ] [VP eat the cake ]]] (do-insertion)
d. … and [TP Mary [T′ [T did ] [VP eat the cake ]]] (past-tense formation)
Because T-to-V lowering doesn’t apply in (59), copy deletion doesn’t apply.
If VP-ellipsis followed T-to-V lowering, the result would be (60):
(60) John ate the cake and Mary too.
(Here the interpretation where Mary eats the cake is the relevant one, not the one
where John eats Mary; see Section 3.4.) The sentence in (60) could be derived
in the following way:
(61) a. … and [TP Mary [T′ [T +d ] [VP eat the cake ]]] (S-structure)
b. … and [TP Mary [T′ [T +d ] [VP [ eat ] + d] the cake ]]] (by T-to-V lowering)
c. … and [TP Mary [T′ [T +d ] [VP [ eat ] + d] the cake ]]] (by copy deletion)
d. … and [TP Mary [T′ [T +d ] [VP [ eat ] + d] the cake ]]] (by VP-ellipsis)
(63) VP-ellipsis > T-to-V lowering > copy deletion > do-insertion > irregular
past-tense formation > assimilation
The crucial observation is that which is also quantificational rather than refer-
ential: the answer to the question could be ‘none of them’, showing that which
course doesn’t have to pick anything out in the domain of discourse. Like every
student, which course applies to a set of entities (courses) and asks for the iden-
tity of a member of that set such that Goona likes it. In quasi-logic:
(67) For which x, x a course, does Goona like x ?
the same value. The variables that have this property are said to be in the scope
of the quantifier. So all the x′s in both (65) and (67) are in the scope of every or
the wh-operator. In predicate logic, (65) could be written as follows:
(68) "x [ student(x) → likes-pizza(x) ]
Here the symbol " is the universal quantifier, meaning roughly ‘every’ or ‘all’,
and the arrow indicates the ‘if–then’ implication relation, so (68) restates (65).
The important point is that, by convention, the scope of the quantifier is indi-
cated by the brackets following it; all and only occurrences of x inside that
bracket are in the scope of the quantifier, i.e. are interpreted as bound by it. This
is a convention developed by logicians.
If we now look at (66) and (67) in the light of these logical conventions, we can
say that the interrogative C (CQ, Wh) which triggers wh-movement to its Specifier
thereby does two things. First, as we saw in our discussion of wh-movement in
Chapter 5, it determines the position of the wh-operator by triggering wh-move-
ment, thereby creating a copy in the original position. Second, it determines
the scope of the wh-phrase. We can directly connect these two points by the
following postulate:
(69) A quantified phrase has scope over the portion of phrase structure it
c-commands at LF which contains a variable that it binds.
NP CQ, Wh′
what
CQ, Wh TP
T NP T′
did you
T VP
(did )
V NP
catch (what )
Here CQ, Wh triggers movement of what from the object position to SpecCQ, WhP.
The Projection Principle requires the presence of the copy in object position.
This copy is deleted at PF by the general copy-deletion rule we discussed in
the previous section, but given that PF and LF do not interact, it remains at LF.
In fact, it is interpreted as the variable bound by the wh-operator. The variable
is in the scope of the operator since it is c-commanded by the operator. This is
guaranteed by the general fact that narrow-syntactic movement always targets a
c-commanding position (we saw in the previous section that PF-movement need
not have this property); furthermore, the fact that the variable is bound by the
wh-operator results from the definition of binding as c-command combined with
coindexing (see Chapter 6, (38)). Hence, (70) can be read at LF as ‘for which
thing x, you caught x′ (glossing what as ‘which thing x′), and this can be read by
interpretative rules for quantifier-variables along the lines sketched above and in
a little more detail in Section 6.7.
Since it is CQ, Wh which triggers wh-movement to its Specifier, this element
effectively determines the scope of wh-operators. We saw this in Section 7.6 in
relation to the different matrix verbs in the following examples:
(71) a. We wonder what you bought.
?*What did you wonder (that) I bought?
We also saw evidence, from sentential particles, that C can bear the Q and
wh-features. So the question arises of how the wh-operator is interpreted in
Mandarin, a question we left open in Section 7.4.
The covert derivation to LF allows us to give a very simple answer to this ques-
tion: the Mandarin sentences in (72) have different structures at LF, where covert
wh-movement operations parallel to those that take place overtly in languages like
English create the required operator-variable structures. So, a simple wh-question
in Chinese such as (73a) has the S-structure in (73b) and the LF in (73c):
(73) a. Lisi mai-le shenme?
Lisi buy- perf what
‘What did Lisi buy?’
b. CQ, WhP
CQ, Wh TP
NP T′
Lisi
T VP
V NP
mai-le shenme
c. CQ, WhP
NP CQ, Wh′
shenme
CQ, Wh TP
NP T′
Lisi
T VP
V NP
mai-le (shenme)
the operation takes place. This difference is regulated by a parameter of the kind
described in Section 7.6. A consequence of this is that if Mandarin has covert
wh-movement, then we might expect it to have covert island phenomena com-
parable to the overt ones of English seen in Section 5.6. I will not go into this
question here, but return to it in Volume II.
Are there any cases of covert movement in English? If we take seriously the
idea that LF represents quantifier-variable relations in terms of binding (coin-
dexation and c-command) of a variable created by movement, then we should
extend our account of wh-questions to sentences containing quantifiers like
every, as in (64) and (65), repeated here:
(64) Every student likes pizza.
(65) For every x, if x is a student, then x likes pizza (or ‘if x is in the set of stu-
dents, then x is in the set of pizza-likers’).
If (65) approximates the LF of (64), then every student raises and its copy is
interpreted as a variable bound by it. So (64) has the LF in (74):
(74) TP
NPi TP
every student T VP
−z
V NP
like pizza
This sentence has two interpretations. In one, every student has scope over some
lecture, so the interpretation is ‘for every student x, there is some lecture y such
that x dreads y′. The dreaded lecture could be a different one in each case (e.g.
Goona dreads Syntax, Vurdy dreads Phonetics, Pete dreads Semantics, etc.). In
the second reading, some lecture has scope over every student, so the meaning
is ‘there is some lecture x such that for every student y, y dreads x′, i.e. one
lecture so dreadful that all students dread it (clearly this cannot be a Syntax lec-
ture …). The first reading is the wide-scope reading for every student (it’s also
the ‘surface scope’ reading, in that the linearly first quantifier has wider scope).
The second reading is the wide-scope reading for some lecture (it is also known
as the ‘inverse scope’ reading as the scope relations don’t follow surface linear
precedence).
If scope is determined by c-command, then we can disambiguate (75) by
applying QR in different orders where there is more than one quantified NP. Thus
the first reading of (75) (wide scope for every student) is represented in (76a),
while the second reading (wide scope for some lecture) is represented in (76b):
(76) a. TP
NPi TP
every student T VP
−z
V NPj
dread
some lecture
b. TP
NPj TP
every student T VP
−z
V NPj
dread
some lecture
The postulate in (69) that scope reduces to c-command determines that every
student has scope over some lecture in (76a) and some lecture has scope over
every student in (76b). These representations can then be input to logical rep-
resentations of the kind seen in (68) as necessary, in order to derive the semantic
interpretations.
We also see covert wh-movement in certain contexts in English. This occurs
in multiple questions (which we saw briefly in Section 5.3), illustrated in (77):
(77) Who wrote which songs?
Here we assume the subject wh-phrase who has moved from SpecT′ to SpecC′
(recall the discussion of this in Section 5.4). The object wh-phrase which songs
has clearly not moved. In English, the rule for multiple questions is that where
there is more than one wh-phrase in the clause only one can move (this isn’t
true in all languages, e.g. Russian and other Slavic languages have multiple
wh-movement). Nonetheless, which songs must bind a variable in (77) in order
to be properly interpreted as a quantifier. Moreover, it interacts with who: the
natural answer pairs people and songs (e.g. Paul wrote Hey Jude, John wrote I
am the Walrus, George wrote Here Comes the Sun, etc.). This is known as the
pair-list interpretation of multiple questions.
We can account for this by proposing that in multiple questions the wh-phrases
combine (by a rule known as Absorption) to form a complex operator, giving an
LF representation of the kind in (78):
(78) [CP [whoi, which songsj] [TP xi wrote xj ]
The interpretation of (78) is ‘for which x, x a person, and for which y, y a song,
x wrote y′, which naturally elicits a pair-list answer of the kind seen in the pre-
vious paragraph. As with covert wh-movement in Mandarin, the postulation of
covert movement in English leads to the prediction of covert island effects; we
will come back to this in Volume II.
The final LF phenomenon we will look at here is reconstruction. This is
illustrated in (79):
(79) [NP Which joke about himselfi ] does Donaldi dislike --- most ?
The lower copy is deleted at PF, giving (79) (without the trace). But, again
thanks to the independence of PF and LF, we can retain the copy at LF. In that
context, the anaphor himself satisfies Principle A: it is c-commanded by its coin-
dexed antecedent Donald within its binding domain (which is TP, the smallest
XP containing the anaphor and a subject as well as a finite T; see (42) of Chapter
6). So LF looks at the copy for satisfaction of Principle A. In fact, since the
occurrence of the anaphor inside the moved wh-phrase in (80) would otherwise
violate Principle A, it may be that copy deletion deletes this copy at LF (but not
the whole wh-phrase, as we need an operator here).
This concludes our very brief introduction to LF. We have seen three examples
of covert movement, i.e. movement that takes place after the derivation splits
into its PF and LF branches at S-structure and whose effects therefore cannot be
directly ‘heard’: wh-movement in Mandarin, QR in English and wh-movement
in multiple questions in English. Finally, we looked at reconstruction in relation
to copies and copy deletion. All of these operations are motivated by the need to
create LF representations which can be directly fed into semantic interpretation.
8.8 Conclusion
To end this chapter, let us look once more at the overall architecture
of the grammar given in (1):
(1) LEXICON
D-STRUCTURE
S-STRUCTURE
Exercises
1. There are at least four kinds of have in English, shown in (a–d):
a. John has written a book. [perfect have]
b. John has to write a book. [modal have (to)]
c. John has many books. [possessive have]
d. John has his students proofread his books. [causative have]
Which of these requires or allows (i) subject-auxiliary inversion, (ii)
negation with n’t, (iii) contraction to ‘ve? What can we conclude
about the position(s) the different kinds of have can occupy? There is
some dialectal variation concerning the possessive and modal forms.
You may also want to try replacing have with (have) got; here again
there is some dialectal variation. Try to formulate and analyse the
data based on your own judgements.
2. Modals have complex and intriguing semantic properties and have
been the object of a great deal of study. One central idea in many
analyses of modal semantics is that they involve quantification over
possible worlds. So, (ia) can be approximately glossed as in (ib), and
(iia) as (iib):
(i) a. I may have lost my keys.
b. In some possible world (consistent with what I know), I
have lost my keys.
(ii) a. I must have lost my keys.
b. In all possible worlds (consistent with what I know), I
have lost my keys.
‘Possibility’ modals like may involve existential quantification over
possible worlds and ‘necessity’ modals like must involve universal
quantification. If modals have quantificational properties, the ques-
tion of their scope, and possible scope ambiguities, arises. What are
the scopal relations between the modal and negation in the following
examples?
(iii) a. John cannot come along today.
b. Confronting the enemy cannot be avoided.
c. You mustn’t drink and drive.
d. You needn’t drink and drive.
(Can is a possibility modal, need is a necessity modal). If c-com-
mand relations at LF determine scope, what LF representations do
these examples have?
3. There is a scope ambiguity in example (i):
(i) What did everyone buy for Bill?
Explain the ambiguity and give the relevant LF representations.
What does this tell us about the nature of QR?
4. Describe the derivation of the following sentence:
(i) How many students were arrested?
Give the D-structure, S-structure, PF and LF representations and
describe how the later levels are derived from the earlier ones. At
PF, describe how the forms were and arrested are derived. At LF,
bear in mind the implicit agent (‘by someone or other’).
Further Reading
Most of the textbooks cited as Further Reading in earlier chapters discuss the organisa-
tion of the grammar (from slightly varying perspectives) and various more advanced top-
ics not covered here. On some of the specific topics covered here (the lexicon, passives,
do-support and affix lowering), see the following:
Adger, D. 2003. Core Syntax. Oxford: Oxford University Press, Chapter 3.5 (on
thematic roles) and 5.5 (on do-support).
Carnie, A. 2002. Syntax: A Generative Introduction. Oxford: Blackwell, Chapters 8 (on
the lexicon), 9 (on head-movement and do-support) and 10 (on passives
and related matters).
Larson, R. 2010. Grammar as Science. Cambridge, MA: MIT Press, Unit 13 (on the
lexicon).
Freidin, R. 2012. Syntax: Basic Concepts and Applications. Cambridge: Cambridge
University Press, Chapter 7.2 (on do).
Radford, A. 2016. Analysing English Sentences. Cambridge: Cambridge University
Press, Chapter 5.10 (on do-support).
Sportiche, D., H. Koopman & E. Stabler. 2014. An Introduction to Syntactic Analysis
and Theory. Oxford: Wiley, Chapter 6.8 and 6.9 (on the lexicon).
Tallerman, M. 1998. Understanding Syntax. London: Routledge, Chapter 7.1 (on
passives).
In the foregoing chapters, I have tried to present the core elements of the the-
ory of syntax. We began with the basic notions of categories and constituents,
and how to find them (Chapter 2). Then we saw how PS-rules can generate
the well-formed structures, creating phrase markers which we can represent
either as tree diagrams or as labelled bracketings. We saw how to justify the
proposed rules and the structures they generate using constituency tests (Chapter
3). Next, we introduced X′-theory, as a simpler, more general and more abstract
format for PS-rules (Chapter 4). In Chapter 5, we introduced movement rules,
concentrating on wh-movement (how it works and where it is unable to apply,
i.e. islands). Chapter 6 introduced the interpretative rules of binding theory,
along with the very important structural relation of c-command and the related
notion of variable binding; here we also made the distinction between A- and
A′-positions. In Chapter 7, we looked a little beyond English and introduced the
important notion of parameter, giving several examples from a small range of
languages along with some data from the World Atlas of Language Structures.
Finally, Chapter 8 introduced the architecture of grammar we adopt, namely the
‘Y-model’, so-called because the derivation looks like an inverted Y, splitting at
S-structure into PF and LF. This led to a discussion of certain PF rules and the
important idea of covert movement in the derivation to LF.
There are many questions that we can raise concerning the model of syn-
tax described here, even while remaining within a broadly Chomskyan concep-
tion of grammar and the language faculty (and of course, if we move beyond
that conception, still more questions arise). One very important question is
whether our assumption that constituency relations (dominance, constituency
and c-command) are primary and dependency relations (subject, object, predi-
cate) secondarily defined in terms of constituency is justified. It has often been
claimed that syntactic descriptions based on constituency work well in languages
like English which have fairly fixed word order and rather little inflectional mor-
phology, especially case morphology, while many languages have much freer
word order and appear to mark relations like subject, object, etc. morphologi-
cally (often with case); such languages appear to lend themselves less well to
a constituency-based approach to syntactic analysis and more naturally to a
dependency-based one. Clearly, this question is partly a cross-linguistic one,
and something we were unable to properly investigate in our brief comparative
discussion in Chapter 7. It is of central importance for syntactic theory though.
215
ZP C′
C TP
NP T′
T VP
V NP
The sentence in (2a) bears a resemblance, both syntactic and semantic, to the
passive The vase was broken. It seems natural to propose that the Theme argu-
ment the vase moves from the object position to the subject position. Finally,
there are raising verbs like seem, which show the alternation in (3):
(3) a. It seems that the world is round.
b. The world seems to be round.
In (3b), the world arguably raises from the lower subject position to the higher one.
A very general question, which we only addressed rather sketchily, is
what causes movement to happen. We said that a wh-feature on C triggers
wh-movement, although we did not go into detail as to how or why this happens,
and we didn’t mention any kind of trigger for head-movement or A-movement.
There is much more to say on these points, most of it of a rather technical nature.
Related to this is the question of what underlies the fact that wh-movement is
(mainly) overt in English but covert in Mandarin, that V raises to T in French
and German but not in English, and T raises to all main-clause Cs in German but
only in certain clause types (mainly interrogatives) in English and French. These
last observations suggest that the triggers for movement should be parametrised
in some way.
The Head Parameter clearly has great potential for accounting for typolog-
ical generalisations about cross-linguistic word-order variation, as we saw in
Section 7.4. However, the nature of disharmonic languages like German and
Mandarin needs clarification. Further, while the other parameters we have looked
at here mainly involve the presence or absence of movement (head-movement
and wh-movement, and we mentioned that there are languages which lack pas-
sives and therefore possibly NP-movement), the Head Parameter seems to con-
cern X′-theory rather than movement. Again, the question arises as to the range
of possible parametric variation: which rule systems can be parametrised?
Finally, there are two very large theoretical questions. First, we have seen that
there are two main rule systems: PS-rules (X′-theory) and movement. We have
treated these as quite distinct and we have treated movement as substitution into
positions created by X′-theory. But we also saw two examples where movement
created new adjunction structures: T-to-V lowering in Section 8.6 and Quantifier
Raising in Section 8.7. Can movement also create structure? If so, how differ-
ent are our two rule systems really? As a related question, we saw that lexical
insertion involves substituting a lexical item for a categorially matching head
generated by PS-rules (e.g. substituting fishN for N). But the noun fish has a
feature N, so if we simply place it in a structure it creates its own N-node. If we
could do this, would we need to say that lexical insertion takes place all at once
at D-structure? Nothing in the lexical entries themselves or in the formulation
of the Projection Principle requires this. Since D-sructure is an internal level,
as we observed at the end of Chapter 8, perhaps it would then not be strictly
necessary. This, in turn, would raise the question of whether we need the other
internal level, namely S-structure: perhaps it is enough to simply say that the
derivation splits into its PF and LF paths at some point, nothing more. All of
these questions are interrelated and form the central object of research in post-
1990 generative grammar; collectively they form a large part of the minimalist
programme for linguistic theory. They are addressed in depth in Volume II, but
the groundwork laid here is essential for understanding their scope and import.
In our conclusion to Chapter 1, after the discussion of the fish sentences in
both English and Italian, which was intended as a demonstration simultaneously
219
which is not I-language. The everyday concepts of languages such as English, French,
etc. are E-language concepts, since they are largely social, cultural and historical.
ellipsis elided categories are ‘understood’, i.e. they are present in the semantic
interpretation (in some way) but absent in the phonological representation (i.e. silent).
Can be used as a constituency test (q.v.) for XPs. A prominent type of ellipsis in English
is VP-ellipsis (q.v.).
elsewhere the default, non-specified or minimally specified context for the application of
a rule; for example +d for English Past Tense.
endocentric a phrasal category which has a head (q.v.). According to the X′-schema for
PS-rules (q.v.) all categories are endocentric.
ethnolect a language variety associated with a specific ethnic group. It may be a
distinguishing mark of social identity, both within the group and for outsiders. Probably
the best-studied ethnolect is Afro-American Vernacular English (AAVE).
evidential markers clausal markers which indicate the speaker’s commitment to the
evidential basis for what is asserted by the clause: typical evidentials mark direct vs
indirect knowledge, witnessed or not witnessed by the speaker, etc.
exocentric a phrasal category which has no head (q.v.). According to the Xʹ-schema for
PS-rules (q.v.) no categories are exocentric.
Experiencer a thematic role designating the sentient being affected by an emotion or
psychological state, one of the arguments of a psych-verb (q.v.).
external argument the s-selected argument of a head which appears outside the maximal
projection of that head. External arguments are never c-selected. Agents are always
external arguments, but external arguments are not always Agents.
feeding order the order of two rules, A and B, such that A applies before B and creates
the structural environment for the application of B, making it possible for B to apply. For
example, movement rules feed copy deletion by creating the environment in which copy
deletion applies.
finite finite clauses contain a tense-marked verb or auxiliary. Main clauses are typically
finite; declarative (q.v.) main clauses always are.
first person the type of personal pronoun or person inflection on a verb used to refer to
the speaker or (in the plural) a group including the speaker, I/me/we/us in English.
fragment a piece of a sentence, often occurring the answer to a question, possibly
derived from one or more applications of ellipsis (q.v.). Can be used as a constituency
test (q.v.) for XPs.
fronting a syntactic operation which places a constituent XP at the beginning of the
sentence; the remainder of the sentence contains a gap corresponding to the neutral
position of XP. A standard constituency test (q.v.) in English.
functional categories the subclass of categories (q.v.) that represent grammatical
rather than lexical information. Functional categories tend to have rather few members,
which are frequently phonologically reduced, affixal, or silent, and to be closed classes.
Some seem to lack a semantic value, others have complex or rather abstract semantics.
Examples include auxiliaries, complementisers and determiners.
natural languages, rather than being centrally planned. Interlingua literature maintains
that (written) Interlingua is comprehensible to the hundreds of millions of people who
speak Romance languages [footnote omitted], though it is actively spoken by only a few
hundred’ (https://en.wikipedia.org/wiki/Interlingua, accessed 14/9/21).
internal argument the c-selected (q.v.) argument(s) of a head X which form its
complement(s) (q.v.) inside X′. Agents are never internal arguments.
interpretative rule rules which specify relations between syntactic constituents which
affect semantic interpretation, distinct from both Phrase-Structure (q.v.) and movement
rules (q.v.). Binding of various kinds is a prominent example.
intervention the case where a syntactic relation between A and B may be blocked if there
is a third element ‘in between’ A and B. Very frequently defined in terms of c-command
(q.v.), i.e. C intervenes between A and B where A c-commands C and B, C c-commands
B and C does not c-command A; see (31) of this Chapter.
intrinsic ordering the order of two rules, A and B, such that one can only precede the
other. For example, the feeding order (q.v.) of movement rules and copy deletion is a
case of intrinsic ordering, as there are no copies to delete before movement applies.
island constraints a class of restrictions on wh-movement (q.v.) including the Complex
NP Constraint (q.v.), the Left Branch Constraint (q.v.) and wh-islands (q.v.). Island
constraints do not permit wh-movement to move a wh-phrase (q.v.) from a position
inside the island to a position outside it.
labelled bracketing in syntax, one of the standard ways of displaying constituent
structure, along with tree diagrams (which are equivalent).
language acquisition the process by which people acquire the ability to speak and
understand a language. First-language acquisition takes place in early childhood. From
the perspective of generative grammar (q.v.), first-language acquisition involves the
development of the language faculty (q.v.) on the basis of the three factors of language
design (q.v.).
language faculty whatever cognitive structure underlies our ability to acquire our native
language, to store the knowledge so acquired in the mind/brain and to put it to use in
production and comprehension. In generative grammar (q.v.), Universal Grammar
(q.v.) is the theory of the genetic endowment underlying the human language faculty.
Left Branch Constraint (LBC) an island constraint (q.v.) which prevents a wh-phrase
(q.v.) moving off a left branch, e.g. SpecN′ in English possessives.
lexical category a category (q.v.) whose head is a substantive lexical item, e.g. a noun,
verb or adjective. Lexical categories are open classes, i.e. it is possible to invent new
members and seemingly always have a phonological form, sometimes quite a complex
multisyllable one. Opposed to functional categories (q.v.).
lexical entry the specification in the lexicon of all and only the idiosyncratic properties
of a lexical item. Lexical entries are discussed in more detail in Section 8.3.
lexical insertion the operation which places lexical items in D-structure (q.v.). As
presented in Section 8.4, this is an operation which substitutes a lexical item into a head
position at that level.
Lexicon the mental dictionary, in which, in principle, all and only idiosyncratic
information about individual lexical items is listed. This information includes at least a
specification of s-selection, c-selection, phonological and semantic properties.
Linnean taxonomy a system of biological classification based on the system established
by the eighteenth-century Swiss biologist Carl Linnaeus. Linnaeus introduced the system
of two-part Latin names for species such as Homo sapiens for humans and felix catus for
domestic cats.
Logic the study of the laws of valid inference, relevant for semantics (q.v.), since sound
inferences are truth-preserving, and truth provides a way to elucidate meaning in truth-
conditional semantics.
Logical Form (LF) the level of representation which converts the syntactic representation
into a representation which can be the input to the semantic representation.
marked the marked value of an asymmetric opposition is the more complex term of
that opposition. Parameters of variation (q.v.) may have marked and unmarked values,
or perhaps a combination of parameter settings may give rise to a marked system, as
suggested in Section 7.4 for the rarer sentential word orders.
maximal projection the phrasal projection of a head (q.v.) X and the intermediate
projection X′. Can be defined as the projection of X immediately dominated by Y X.
morphology the study of the internal structure of complex words.
movement rule one kind of transformational rule (q.v.) which copies categories into a
c-commanding position.
multiple questions questions featuring more than one wh-phrase (q.v.). In English,
since only one wh-phrase can move per clause, the other(s) must remain in situ.
narrow syntax the collective term for the ‘internal’ syntactic levels of representation,
D-structure (q.v.) and S-structure (q.v.), which do not interface with non-syntactic
parts of the grammar.
natural kind ‘To say that a kind is natural is to say that it corresponds to a grouping that
reflects the structure of the natural world rather than the interests and actions of human
beings. We tend to assume that science is often successful in revealing these kinds; it is a
corollary of scientific realism that when all goes well the classifications and taxonomies
employed by science correspond to the real kinds in nature’, Stanford Encyclopedia of
Philosophy (https://plato.stanford.edu/entries/natural-kinds/#SemNatKinTer, accessed
14/9/21).
Navajo a Southern Athabaskan language spoken by 170,000 people primarily in the
Southwestern United States.
negative concord the phenomenon, found in French and many other languages, whereby
the co-occurrence of two negative morphemes in a sentence does not give rise to a double
negation but rather to single negation, also found in many varieties of non-standard
English.
node a position in a tree diagram (q.v.) from which either at least one branch (q.v.)
emanates ‘downwards’ (a non-terminal node (q.v.)) or a site of lexical insertion (q.v.).
nominative one of the case forms found in Latin, Sanskrit and many other languages. It
typically marks the subject (q.v.) of a sentence.
non-terminal node a node (q.v.) in a tree diagram (q.v.) from which at least one branch
emanates ‘downwards’ and hence which dominates (q.v.) at least one other node (q.v.).
Noun Phrase (NP) a phrase which contains a noun or pronoun and occupies the same
syntactic positions as a noun or pronoun (subject, direct object, object of preposition,
etc.).
NP-movement the operation which moves an object NP into the subject position in
passives.
number the inflectional category which distinguishes singular from plural, marked on
nouns in many languages, and on verbs in quite a few (in English, only in the present);
some languages, e.g. Mandarin, have no obligatory number marking (either on nouns or
verbs).
operator roughly equivalent to quantifier (q.v.), a category which binds a variable from
an Aʹ-position (q.v.), giving rise to quantificational interpretation. Wh-phrases are a
typical example.
operator-variable structures a syntactic configuration in which an operator (q.v.)
or quantifier (q.v.) binds a variable from an Aʹ-position (q.v.), giving rise to a
quantificational interpretation.
pair-list interpretation the usual interpretation of multiple wh-questions, in which the
answers associated with each wh-phrase are paired, e.g. John drank beer and Paul drank
whisky in answer to Who drank what?
parameters of variation a way to define and analyse syntactic variation among languages.
A given grammatical system (I-language) may be seen as the set of parameter values
characteristic of that system, combined with the invariant aspects of Universal Grammar.
Patient the thematic role (q.v.) of the argument that is affected by the action described
the verb, e.g. her dinner in Mary ate her dinner.
performance putting competence (q.v.) into practice in production and comprehension;
performance involves competence, combined with short-term memory, attention and
other non-linguistic cognitive capacities.
person the inflectional category which disambiguates between first, second and third,
marked on pronouns in very many languages, and on verbs in many languages (in English
only in the third-person singular of the present).
Phonological Form (PF) the level of representation which converts a syntactic
representation (S-structure (q.v.), given the standard architecture seen in (1) of this
chapter) into a representation which can be input to phonological and phonetic rules and
representations.
phonology the branch of linguistics dealing with how speech sounds are organized in
linguistic systems.
phrasal category in syntax (q.v.?), a unit of organization of words, intermediate
between the word level and the sentence level; phrasal categories are non-root, non-
terminal nodes (see Chapter 3).
phrasal stress the accent (marker of phonological prominence through pitch, loudness,
length or some combination of these) which falls on the designated most prominent
syllable in a phonological phrase (often close to a syntactic XP), as in Oxford Róad.
phrase marker synonymous with tree diagram (q.v.) or labelled bracketing, the means
of displaying constituent structure (q.v.).
Phrase-Structure (PS) rules rules which generate (q.v.) constituent structure (q.v.).
The recursive nature of these rule systems accounts for the fact that natural languages
make infinite use of finite means.
pied-piping a case of wh-movement (q.v.) where the moving wh-phrase (q.v.) takes
along other parts of a constituent to which it belongs in its pre-movement position.
postposition the category P is postpositional if it has the head-final (q.v.) value of the
Head Parameter (q.v.); if head-initial (q.v.) it is prepositional. Strictly speaking, P
should be taken as the abbreviation for the order-neutral term ‘adposition’.
predicate logic the branch of logic which breaks propositions (roughly, logical-semantic
representations of sentences) down into predicates and arguments; a key aspect of
predicate logic is its ability to represent quantification.
predicate traditionally seen as the part of a sentence which describes a property or action of
the subject, typically corresponding to VPs, PPs or APs, although NPs can be predicates too.
Preposition a lexical category (q.v.) typically denoting a location or path and taking
an NP complement. Unlike other lexical categories, Prepositions are closed-class items.
Preposition stranding (q.v.) of the complement of a preposition without associated
pied-piping (q.v.) of the preposition. Found in English (e.g. in Who did you talk to?) but
cross-linguistically rare.
prescriptive grammar ‘the establishment of rules defining preferred or correct usage
of language … such normative practices often suggest that some usages are incorrect,
inconsistent, illogical, lack communicative effect, or are of low aesthetic value, even in
cases where such usage is more common than the prescribed usage’ (https://en.wikipedia
.org/wiki/Linguistic_prescription, accessed 14/9/21).
present tense a form of the verb which, in English, is just the root in all persons except
the third-person singular, where it is marked by -s. The present tense situates the event or
situation described by the sentence in the present time.
pro-form a reduced form of an XP that can substitute for that XP in the appropriate
context. Pronouns are pro-NPs, do so can function as a pro-VP and so as a pro-AP.
progressive a verbal aspect indicating, roughly, ongoing action. Formed by adding -ing
to the verb and marking tense and agreement inflection on the auxiliary be, as in John
was singing.
projection the occurrences of the category of a head (q.v.) which contain further material in
addition to the head: the X′ projection contains the head and its lexically specified complement,
and the maximal projection (q.v.) XP contains the X′ projection and the Specifier (q.v.).
Projection Principle the principle which conditions lexical insertion (q.v.) and holds
lexical properties constant through the syntactic derivation by requiring these properties
to hold at all levels. A particular formulation is given in (25) of Chapter 8.
second-person pronoun the type of personal pronoun used to refer to an addressee, you
in English. English does not distinguish the singular and plural of this person, neither
does it make politeness distinctions indicating the status of the addressee, as many other
languages do.
semantics the study of meaning, particularly sentence meaning.
sign language language(s) transmitted through the visual-gestural rather than the oral-
aural medium, mostly used by deaf communities; sign languages are now known to be
languages in every sense of the word, having all the salient structural properties of oral-
aural languages.
sociolect a variety of language used by a particular socioeconomic class.
Specifier the category (itself a maximal projection (q.v.)) which combines with
the intermediate X′ projection to form the maximal projection (q.v.). Specifiers are
typically, but not only, modifiers in a rather broad sense.
string-vacuous movement movement of a category in such a way that the constituent
structure is changed, but the linear order of terminals is not. The best-known example is
wh-movement (q.v.) of subjects in English, as in Who left?
strong crossover the configuration where a wh-phrase moves ‘over’ a pronoun, which is
unable to take it as its antecedent. Falls under Binding Principle C (q.v.); see Section 6.7.
structural description the precise description of the structure of a syntactic object,
usually a sentence. The standard form of such a description is in terms of constituent
structure (q.v.) generated by Phrase-Structure rules (q.v.) and transformational rules
(q.v.).
Subject an NP which typically marks the Agent (q.v.) of an action described a VP.
substitution (transformation) a movement rule copying a category into a position
which has been created by the Phrase-Structure rules (q.v.). Both NP-movement
(q.v.) and wh-movement (q.v.) substitute into Specifier positions (SpecT′ and SpecC′
respectively).
syntax the study of the structure of sentences.
terminal node a node (q.v.) in a tree diagram (q.v.) which dominates (q.v.) no other
node; the site of lexical insertion.
thematic role lexico-semantic properties of predicates which designate the way in which
their arguments participate in the eventuality they describe. An important component of
argument structure (q.v.).
third-person plural the third-person plural pronoun in English is they/them/their.
Generally used to refer to a group of known individuals who are neither the speaker nor
the addressee(s), although it may also be used as an arbitrary pronoun (they eat pasta in
Italy) or as a singular pronoun to avoid reference to gender. In many languages, verbs
have special forms for the third-person plural distinct from other verb forms; English
does not, however.
three factors in language design the three elements that together constitute the adult
language faculty: (i) the genetic endowment (Universal Grammar), (ii) the primary
linguistic data which acts as the input to language acquisition, (iii) third factors of a
general cognitive nature.
Topic typically an XP that has undergone fronting (q.v.). Topics generally stand for
‘old’ or known information, something already talked about or given in the context.
trace a cover term for a position from which movement takes place. More precisely,
copies (q.v.) occupy these positions.
transformational rule syntactic rules which map phrase markers (q.v.) into other
phrase markers. Deletion and movement rules (q.v.) are types of transformations.
tree diagram one of the standard ways of displaying constituent structure (q.v.), along
with labelled brackets. Synonymous with phrase marker (q.v.).
truth conditions the standard way of understanding sentence meaning; the central idea is
that to know the meaning of a declarative sentence is to understand what the world would
have to be like for the sentence to be true.
unbounded dependency the property of wh-movement (q.v.) which gives the
appearance of movement across arbitrarily long stretches of material.
Undergoer a further thematic role, almost synonymous with Patient (q.v.).
Universal Grammar (UG) the initial state of the language faculty, assumed to be
genetically endowed. The first fator in language design.
universal quantifier a quantifier, written " in predicate logic (q.v.), corresponding
roughly to every, each or all in English. Its semantics involves set-inclusion, i.e. Every A
is B means that all members of A are included in B but not necessarily vice versa.
verb second the syntactic constraint operative in German (see Section 7.5) and other
Germanic languages, but not English, which requires that the finite verb or auxiliary be
preceded by exactly one XP in declarative main clauses.
Volapük a conlang created by the German Catholic priest Johann Martin Schleyer in
1879–80. It was later displaced by Esperanto.
VP-ellipsis a kind of ellipsis (q.v.) which affects VPs. Quite common in English, and a
good constituency test (q.v.) for VPs.
V-to-T Parameter the parameter which determines whether the lexical verb V raises to
T to combine with tense-agreement marking. Positive in French and German; negative
in English.
Warlpiri an Australian language spoken by about 3,000 people Australia’s Northern
Territory. It is a member of the Pama–Nyungan family.
weak crossover a further configuration where a wh-phrase moves ‘over’ a pronoun,
like strong crossover (q.v.). In this case, though, the crossed-over pronoun does not
c-command (q.v.) the variable bound by the wh-phrase. The effect of unacceptability is
weaker than in the case of strong crossover, hence the name.
wh-islands an island constraint (q.v.) which prevents wh-movement (q.v.) out of a CP
whose Specifier is already filled by a wh-phrase (q.v.).
234
S-structure (surface structure), 180, 190–5, fish and some relative clauses, 22–5
211–12 linguistic theory, 31–2
Logical Form (LF), 203–11 six- and seven-fish sentences, 25–30
Phonological Form (PF), 195–203 teleological perspective, 8, 9
satellite words, 21–2, 37 tense
scientific approach to linguistics, 1–3, 9–12 functional categories, 37–8
second-person pronouns, 16, 17 wh-movement, 109–11
semantics, 1, 42–3 TenseP (TP), 92–5, 110, 122–5
sign language, 5–6 terminal nodes (tree diagrams), 46
sociopolitical perspective, 8 thematic roles, 102, 180
sociolects, 9 third-person plural, 18–19
Specifiers, Phrase-Structure rules, 82–7 to, 92–3
Standard English, 8, 9 Tolkien, J. R. R., 6
stress (phonology), 42 topic, fronting, 61–2
string-vacuous movement, 113 trace, head-movement, 101–2
strong crossover, 151–2 transformations, 100
structural descriptions see also movement rules
constituency tests, 59–75 tree diagrams, 45–7
grammaticality, 51 local PS-rules, 118–22
Phrase-Structure rules, 52–5, 75–6, 99 Phrase-Structure rules, 52–5
possible and impossible PS-rules, 78–80 structural descriptions, 51
presentation of, 51 truth conditions, 148
recursion, 55–9
syntactic representations, 29 unbounded dependency, 117–18
subject-auxiliary inversion, 39–40, 90, 101–2, Undergoer of the action (argument structure), 102
114–15 ungrammatical language see grammaticality
subject-object-verb (SOV) word order Universal Grammar (UG)
French vs English, 157–60 competence, 12
Japanese, 161–3 language acquisition, 10
X′-theory, 160–4 syntax beyond English, 156–7
subject questions, 112–18 universal quantifiers, 186
subject-verb-object (SVO) word order
French vs English, 157–60 V-to-T movement, 171, 175–6
German, 165–71 V-to-T parameter, 159–60, 167
substitution, 121 Verb Phrase (VP)
substitution transformations, 121 fish, 22–3, 26–7
syntactic diagnostics, 39–42 phrasal constituents, 45
syntactic representations, 29 Phrase-Structure rules, 55–9
syntactic theory, goal of, 32 possible and impossible PS-rules, 79–80
syntax syntactic diagnostics, 40
as constituent structure, 35–6 VP-ellipsis, 66–8
I-language, 35 wh-movement, 61, 109–11, 122–5
as sub-discipline, 1 X′-theory, 82–4
syntax beyond English, 156 verb positions, French and English, 157–60
approaching Universal Grammar, 156–7 verb second, 165
German, 165–71 see also subject-verb-object (SVO) word order
verb positions in French and English, 157–60 verbs
wh-movement, 171–4 English vs Italian, 18–19
word order and X′-theory, 160–4 external arguments, 184–5
fish, 15–22
T-to-C movement lexical categories, 36
S-structure, 192 lexical entries, 186–7
subject-auxiliary inversion, 114–15 morphological diagnostics, 38–9
target of, 122 semantic criteria, 42–3
word order, 168–9
tacit knowledge, 15 weak crossover, 151–2
fish, 15–22 Weinreich, Max, 8