Word Grammar (Richard Hudson)
Word Grammar (Richard Hudson)
Word Grammar (Richard Hudson)
continuum
Continuum
The Tower Building 15 East 26th Street
11 York Road New York
London SE1 7NX NY 10010
Part I
Word Grammar Approaches to Linguistic Analysis:
Its explanatory power and applications 33
2. Case Agreement in Ancient Greek: Implications for a
theory of covert elements 35
Chet Creider and Richard Hudson
1. Introduction 35
2. The Data 35
3. The Analysis of Case Agreement 41
4. Non-Existent Entities in Cognition and in Language 42
5. Extensions to Other Parts of Grammar 46
6. Comparison with PRO and pro 49
7. Comparison with Other PRO-free Analyses 50
8. Conclusions 52
3. Understood Objects in English and Japanese with
Reference to Eat and Taberui A Word Grammar
account 54
Kensei Sugayama
1. Introduction 54
2. Word Grammar 56
viii WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
3. Eat in English 58
4. Taberu in Japanese 60
5. Conclusion 63
4. The Grammar of Be To: From a Word Grammar
point of view 67
Kensei Sugayama
1. Introduction and the Problem 67
2. Category of Be 68
3. Modal Be in Word Grammar 69
4. Morphological Aspects 70
5. Syntactic Aspects 71
6. Semantics of the Be To Construction 72
7. Should To be Counted as Part of the Lexical Item? 75
8. A Word Grammar Analysis of the Be To Construction 77
9. Conclusion 81
5. Linking in Word Grammar 83
Jasper Holmes
1. Linking in Word Grammar: The syntax semantics principle 83
2. The Event Type Hierarchy: The framework; event types;
roles and relations 103
3. Conclusion 114
6. Word Grammar and Syntactic Code-Mixing Research 117
Eva Eppler
1. Introduction 117
2. Constituent Structure Grammar Approaches to Intra-Sentential
Code-Mixing 118
3. A Word Grammar Approach to Code-Mixing 121
4. Word Order in Mixed and Monolingual 'Subordinate' Clauses 128
5. Summary and Conclusion 139
7. Word Grammar Surface Structures and HPSG Order
Domains 145
Takafumi Maekawa
1. Introduction 145
2. A Word Grammar Approach 146
3. An Approach in Constructional HPSG: Ginzburg and Sag 2000 154
4. A Linearization HPSG Approach 160
5. Concluding Remarks 165
Part II
Towards a Better Word Grammar 169
8. Structural and Distributional Heads 171
Andrew Rosta
1. Introduction 171
2. Structural Heads 172
WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE ix
JASPERW. HOLMES is a self-employed linguist who has worked with many large
organizations on projects in lexicography, education and IT. Teaching and
research interests include syntax and semantics, lexical structure, corpuses and
other IT applications (linguistics in computing, computing in linguistics),
language in education and in society, the history of English and English as a
world language. His publications include 'Synonyms and syntax' (1996, with
Richard Hudson, And Rosta, Nik Gisborne). Journal of Linguistics 32; 'The
syntax and semantics of causative verbs' (1999). UCL Working Papers in
Linguistics 11; 'Re-cycling in the encyclopedia' (2000, with Richard Hudson), in
B. Peeters (ed. ) The Lexicon-Encyclopedia Interface (Amsterdam: Elsevier);
'Constructions in Word Grammar' (2005, with Richard Hudson) in Jan-Ola
Ostman and Mirjam Fried (eds) Construction Grammars: Cognitive Grounding and
Theoretical Extensions (Amsterdam: Benjamins).
Email: jasper. holmes@gmail. com
Introduction to Systemic Grammar, North Holland, 1971); and his second book
was about Daughter-Dependency Grammar (Arguments for a Non-transforma-
tional Grammar, University of Chicago Press, 1976). As the latter tide indicates,
Chomsky's transformational grammar was very much 'in the air', and both
books accepted his goal of generative grammar but offered other ideas about
sentence structure as alternatives to his mixture of function-free phrase structure
plus transformations. In the late 1970s when Transformational Grammar was
immensely influential, Richard Hudson abandoned Daughter-Dependency
Grammar (in spite of its drawing a rave review by Paul Schachter in Language
54, 348-76). His exploration of various general ideas that hadn't come together
became an alternative coherent theory called Word Grammar, first described
in the 1984 book Word Grammar and subsequently improved and revised in
the 1990 book English Word Grammar. Since then the details have been
worked out much better and there is now a workable notation and an
encyclopedia available on the internet (cf. Hudson 2004). The newest version
of Word Grammar is now on its way (Hudson in preparation).
The time span between the publication of Richard Hudson's Word Grammar
(1984) and this volume is more than two decades (21 years to be precise). The
intervening years have seen impressive developments in this theory by the WG
grammarians as well as those in other competitive linguistic theories such as
Minimalist Programme, Head-driven Phrase Structure Grammar (HPSG),
Generalized Phrase Structure Grammar (GPSG), Lexical Functional Grammar
(LFG), Construction Grammar and Cognitive Grammar.
Here are the main ideas, most of which come from the latest version of the
WG homepage (Hudson 2004), together with an indication of where they came
from:
• It is monostratal - only one structure per sentence, no transformations.
(From Systemic Grammar).
• It uses word-word dependencies - e. g. a noun is the subject of a verb.
(From John Anderson and other users of Dependency Grammar, via
Daughter-Dependency Grammar; a reaction against Systemic Grammar
where word-word dependencies are mediated by the features of the mother
phrase. )
• It does not use phrase structure - e. g. it does not recognize a noun phrase
as the subject of a clause, though these phrases are implicit in the
dependency structure. (This is the main difference between Daughter-
Dependency Grammar and Word Grammar. )
• It shows grammatical relations/functions by explicit labels - e. g. 'subject'
and 'object'. (From Systemic Grammar).
• It uses features only for inflectional contrasts - e. g. tense, number but not
transitivity. (A reaction against excessive use of features in both Systemic
Grammar and current Transformational Grammar. )
• It uses default inheritance, as a very general way of capturing the contrast
between 'basic' or 'underlying' patterns and 'exceptions' or 'transforma-
tions' - e. g. by default, English words follow the word they depend on, but
exceptionally subjects precede it; particular cases 'inherit' the default pattern
WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE xv
Figure 1
xvi WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
analysis in order to address the relation between a word and the construction.
In Chapter 4, he attempts to characterize the be to construction within the WG
framework. He has shown that a morphological, syntactic and semantic analysis
of be in the be to construction provides evidence for the category of be in this
construction. Namely, be is an instance of modal verb in terms of morphology
and syntax, while the sense of the whole construction is determined by the
sense of 'to'.
In Chapter 5, Holmes in a very original approach develops an account for
the linking of syntactic and semantic arguments in the WG approach. Under
the WG account, both thematic and linking properties are determined at both
the specific and the general level. This is obviously an advantage.
In Chapter 6, Eppler draws on experimental studies concerning the code-
mixing and successfully extends WG to an original and interesting area of
research. Constituent-based models have difficulties accounting for mixing
between SVO and SOV languages like English and German. A dependency
(WG) approach is imperative here. A word's requirements do not project to
larger units like phrasal constituents. The Null-Hypothesis, then, formulated in
WG terms, assumes that each word in a switched dependency satisfies the
constraints imposed on it by its own language. The material is taken from
English/German conversations of Jewish refugees in London.
Maekawa continues the sequence in this collection towards more purely
theoretical studies. In Chapter 7, he looks at three different approaches to the
asymmetries between main and embedded clauses with respect to the elements
in the left periphery of a clause: the dependency-based approach within WG,
the Constructional HPSG approach, and the Linearization HPSG analysis.
Maekawa, a HPSG linguist, argues that the approaches within WG and the
Constructional HPSG have some problems in dealing with the relevant facts,
but that Linearization HPSG provides a straightforward account of them.
Maekawa's analysis suggests that linear order should be independent to a
considerable extent from combinatorial structure, such as dependency or
phrase structure.
Following these chapters are more theoretical chapters which help to
improve the theory and clarify what research questions must be undertaken
next.
Part II contains two chapters that examine two theoretical key concepts in
WG, head and dependency. They are intended to help us progress a few steps
forward in revising and improving the current WG, together with Hudson (in
preparation).
The notion of head is a central one in most grammars, so it is normal that it
is discussed and challenged by WG and other theorists. In Chapter 8, Rosta
distinguishes between two kinds of head and claims that every phrase has both a
distributional head and a structural head, although he agrees that normally the
same word is both distributional and structural head of a phrase. Finally,
Gisborne's Chapter 9 then challenges Hudson's classification of dependencies.
The diversification of heads (different kinds of dependency) plays a role in
WG as well. Gisborne is in favour of a more fine-grained account of
xviii WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Kensei Sugayama
References
Hudson, R. A. (1971), English Complex Sentences: An Introduction to Systemic Grammar.
Amsterdam: North Holland.
— (1976), Arguments for a Non-transformational Grammar. Chicago: University of
Chicago Press.
— (1984), Word Grammar. Oxford: Blackwell.
— (1990), English Word Grammar. Oxford: Blackwell.
— (2004, July 1-last update), 'Word Grammar', (Word Grammar), Available:
www. phon. ucl. ac. uk/home/dick/wg. htm (Accessed: 18 April 2005).
— (in preparation), Advances in Word Grammar. Oxford: Oxford University Press.
Pollard, C. and Sag, LA. (1987), Information-Based Syntax and Semantics. Stanford:
CSLI.
Schachter, P. (1978), 'Review of Arguments for a Non-Transformational Grammar'.
Language, 17, 348-76.
Sugayama, K. (ed. ) (2003), Studies in Word Grammar. Kobe: Research Institute of
Foreign Studies, KCUFS.
Tesniere, Lucien (1959), Elements de Syntaxe Structurale. Paris: Klincksieck.
Introduction
This page intentionally left blank
1 What is Word Grammar?
RICHARD HUDSON
Abstract
The chapter summarizes the Word Grammar (WG) theory of language structure
under the following headings: 1. A brief overview of the theory; 2. Historical
background; 3. The cognitive network: 3. 1 Language as part of a general network;
3. 2 Labelled links; 3. 3 Modularity; 4. Default inheritance; 5. The language
network; 6. The utterance network; 7. Morphology; 8. Syntax; 9. Semantics; 10.
Processing; and 11. Conclusions.
As the theory's name suggests, the central unit of analysis is the word, which
is central to all kinds of analysis:
• Grammar. Words are the only units of syntax (section 8), as sentence
structure consists entirely of dependencies between individual words; WG
is thus clearly part of the tradition of dependency grammar dating from
Tesniere (1959; Fraser 1994). Phrases are implicit in the dependencies, but
play no part in the grammar. Moreover, words are not only the largest units
of syntax, but also the smallest. In contrast with Chomskyan linguistics,
syntactic structures do not, and cannot, separate stems and inflections, so
WG is an example of morphology-free syntax (Zwicky 1992: 354). Unlike
syntax, morphology (section 7) is based on constituent-structure, and the
two kinds of structure are different in others ways too.
• Semantics. As in other theories words are also the basic lexical units
where sound meets syntax and semantics, but in the absence of phrases,
words also provide the only point of contact between syntax and semantics,
giving a radically 'lexical' semantics. As will appear in section 9, a rather
unexpected effect of basing semantic structure on single words is a kind of
phrase structure in the semantics.
• Situation. We shall see in section 6 that words are the basic units for
contextual analysis (in terms of deictic semantics, discourse or sociolinguistics).
Words, in short, are the nodes that hold the 'language' part of the human
network together. This is illustrated by the word cycled in the sentence / cycled to
UCL, which is diagrammed in Figure 1.
Figure 1
WHAT IS WORD GRAMMAR? 5
As can be seen in this diagram, cycled is the meeting point for ten
relationships which are detailed in Table 1. These relationships are all quite
traditional (syntactic, morphological, semantic, lexical and contextual), and
traditional names are used where they exist, but the diagram uses notation
which is peculiar to WG. It should be easy to imagine how such relationships
can multiply to produce a rich network in which words are related to one
another as well as to other kinds of element including morphemes and various
kinds of meaning. All these elements, including the words themselves, are
'concepts' in the standard sense; thus a WG diagram is an attempt to model a
small part of the total conceptual network of a typical speaker.
2. Historical Background
The theory described in this article is the latest in a family of theories which have
been called 'Word Grammar' since the early 1980s (Hudson 1984). The present
theory is very different in some respects from the earliest one, but the continued
use of the same name is justified because we have preserved some of the most
fundamental ideas - the central place of the word, the idea that language is a
network, the role of default inheritance, the clear separation of syntax and
semantics, the integration of sentence and utterance structure. The theory is still
changing and further changes are already identifiable (Hudson, in preparation).
As in other theories, the changes have been driven by various forces - new
data, new ideas, new alternative theories, new personal interests; and by the
influence of teachers, colleagues and students. The following brief history may
be helpful in showing how the ideas that are now called 'Word Grammar'
developed during my academic life.
The 1960s. My PhD analysis of Beja used the theory being developed by
Halliday (1961) under the name 'Scale-and-Category' grammar, which later
turned into Systemic Functional Grammar (Butler 1985; Halliday 1985). I
spent the next six years working with Halliday, whose brilliantly wide-ranging
analyses impressed me a lot. Under the influence of Chomsky's generative
6 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Most of the work done since the start of WG has applied the theory to English,
but it has also been applied to the following languages: Tunisian Arabic (Chekili
1982); Greek (Tzanidaki 1995, 1996a, b); Italian (Volino 1990); Japanese
(Sugayama 1991, 1992, 1993, 1996; Hiranuma 1999, 2001); Serbo-Croatian
(Camdzic and Hudson 2002); and Polish (Gorayska 1985).
The theory continues to evolve, and at the time of writing a 'Word Grammar
Encyclopedia' which can be downloaded via the WG website (www. phon. u-
cl. ac. uk/home/dick/wg. htm) is updated in alternate years.
The idea is that memory connections provide the basic building blocks through
which our knowledge is represented in memory. For example, you obviously know
your mother's name; this fact is recorded in your memory. The proposal to be
considered is that this memory is literally represented by a memory connection,...
That connection isn't some appendage to the memory. Instead, the connection is the
memory.... all of knowledge is represented via a sprawling network of these
connections, a vast set of associations. (Reisberg 1997: 257-8)
'come' but that its past tense is the irregular came, whereas regular past tenses
are handled by a general rule and not stored in the network. The WG view is
that exceptional and general patterns are indeed different, but that they can
both be accommodated in the same network because it is an 'inheritance
network' in which general patterns and their exceptions are related by default
inheritance (which is discussed in more detail in section 4). To pursue the last
example, both patterns can be expressed in exactly the same prose:
(1) The shape of the past tense of a verb consists of its stem followed by -d.
(2) The shape of the past tense of come consists of came.
The only difference between these rules lies in two places: 'a verb' versus come,
and 'its stem followed by -ed" versus came. Similarly, they can both be
incorporated into the same network, as shown in Figure 2 (where the triangle
once again shows the 'isa' relationship by linking the general concept at its base
to the specific example connected to its apex):
Figure 2
3. 2 Labelled links
It is easy to misunderstand the network view because (in cognitive psychology)
there is a long tradition of 'associative network' theories in which all links have
just the same status: simple 'association'. This is not the WG view, nor is it the
view of any of the other theories mentioned above, because links are
WHAT IS WORD GRAMMAR? 9
Figmre 3
there is a single matrix for each concept, and if two attributes have the same
value this is shown (at least in one notation) by an arc that connects the two
value-slots. But when it comes to the attributes themselves, their labels are
repeated across matrices (or even within a single complex matrix). For example,
the matrix for a raising verb contains within it the matrix for its complement
verb; an arc can show that the two subject slots share the same filler but the only
way to show that these two slots belong to the same (kind of) attribute is to
repeat the label 'subject'.
In a network approach it is possible to show both kinds of identity in the
same way: by means of a single node with multiple 'isa' links. If two words are
both nouns, we show this by an isa link from each to the concept 'noun'; and if
two links are both 'subject' links, we put an isa link from each link to a single
general 'subject' link. Thus labelled links and other notational tricks are just
abbreviations for a more complex diagram with second-order links between
links. These second-order links are illustrated in Figure 4 for car and bicycle, as
well as for the sentence Jo snores.
Figure 4
WHAT IS WORD GRAMMAR? 11
3. 3 Modularity
The view of language as a labelled network has interesting consequences for the
debate about modularity: is there a distinct 'module' of the mind dedicated
exclusively to language (or to some part of language such as syntax or
inflectional morphology)? Presumably not if a module is defined as a separate
'part' of our mind and if the language network is just a small part of a much
larger network. One alternative to this strong version of modularity is no
modularity at all, with the mind viewed as a single undifferentiated whole; this
seems just as wrong as a really strict version of modularity. However there is a
third possibility. If we focus on the links, any such network is inevitably
'modular' in the much weaker (and less controversial) sense that links between
concepts tend to cluster into relatively dense sub-networks separated by
relatively sparse boundary areas.
Perhaps the clearest evidence for some kind of modularity comes from
language pathology, where abilities are impaired selectively. Take the case of
Pure Word Deafness (Airman 1997: 186), for example. Why should a person
be able to speak and read normally, and to hear and classify ordinary noises,
but not be able to understand the speech of other people? In terms of a WG
network, this looks like an inability to follow one particular link-type ('sense') in
one particular direction (from word to sense). Whatever the reason for this
strange disability, at least the WG analysis suggests how it might apply to just this
one aspect of language, while also applying to every single word: what is
damaged is the general relationship 'sense', from which all particular sense
relationships are inherited. A different kind of problem is illustrated by patients
who can name everything except one category - e. g. body-parts or things
typically found indoors (Pinker 1994: 314). Orthodox views on modularity
seem to be of little help in such cases, but a network approach at least explains
how the non-linguistic concepts concerned could form a mental cluster of
closely-linked and mutually defining concepts with a single super-category. It is
easy to imagine reasons why such a cluster of concepts might be impaired
selectively (e. g. that closely related concepts are stored close to each other, so a
single injury could sever all their sense links), but the main point is to have
provided a way of unifying them in preparation for the explanation.
In short, a network with classified relations allows an injury to apply to
specific relation types so that these relations are disabled across the board. The
12 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
approach also allows damage to specific areas of language which form clusters
with strong internal links and weak external links. Any such cluster or shared
linkage defines a kind of 'module' which may be impaired selectively, but the
module need not be innate: it may be 'emergent', a cognitive pattern which
emerges through experience (Karmiloff-Smith 1992; Bates et al. 1998).
4. Default Inheritance
Default inheritance is just a formal version of the logic that linguists have always
used: true generalizations may have exceptions. We allow ourselves to say that
verbs form their past tense by adding -ed to the stem even if some verbs don't,
because the specific provision made for these exceptional cases will
automatically override the general pattern. In short, characteristics of a general
category are 'inherited' by instances of that category only 'by default' - only if
they are not overridden by a known characteristic of the specific case. Common
sense tells us that this is how ordinary inference works, but default inheritance
only works when used sensibly. Although it is widely used in artificial
intelligence, researchers treat it with great caution (Luger and Stubblefield 1993:
386-8). The classic formal treatment is Touretzky (1986).
Inheritance is carried by the 'isa' relation, which is another reason for
considering this relation to be fundamental. For example, because snores isa
'verb' it automatically inherits all the known characteristics of 'verb' (i. e. of 'the
typical verb'), including, for example, the fact that it has a subject; similarly,
because the link between Jo and snores in Jo snores isa 'subject' it inherits the
characteristics of 'subject'. As we have already seen, the notation for 'isa'
consists of a small triangle with a line from its apex to the instance. The base of
the triangle which rests on the general category reminds us that this category is
larger than the instance, but it can also be imagined as the mouth of a hopper
into which information is poured so that it can flow along the link to the
instance.
The mechanism whereby default values are overridden has changed during
the last few years. In EWG, and also in Fraser and Hudson (1992), the
mechanism was 'stipulated overriding', a system peculiar to WG; but since then
this system has been abandoned. WG now uses a conventional system in which
a fact is automatically blocked by any other fact which conflicts and is more
specific. Thus the fact that the past tense of COME is came automatically blocks
the inheritance of the default pattern for past tense verbs. One of the
advantages of a network notation is that this is easy to define formally: we always
prefer the value for 'R of C' (where R is some relationship, possibly complex,
and C is a concept) which is nearest to C (in terms of intervening links). For
example, if we want to find the shape of the past tense of COME, we have a
choice between came and corned, but the route to came is shorter than that to
corned because the latter passes through the concept 'past tense of a verb'. (For
detailed discussions of default inheritance in WG, see Hudson 2000a, 2003b. )
Probably the most important question for any system that uses default
inheritance concerns multiple inheritance, in which one concept inherits
WHAT IS WORD GRAMMAR? 13
SEMANTICS
SYNTAX
MORPHOLOGY
PHONOLOGY
GRAPHOLOGY
Figure 5
Figure 6
utterance network has just the same formal characteristics as the permanent
network. For example, suppose you say to me 'I agree. ' My task, as hearer, is to
segment your utterance into the two words / and agree, and then to classify each
of these as an example of some word in my permanent network (my grammar).
This is possible to the extent that default inheritance can apply smoothly; so, for
example, if my grammar says that / must be the subject of a tensed verb, the
same must be true of this token, though as we shall see below, exceptions can
be tolerated. In short, a WG grammar can generate representations of actual
utterances, warts and all, in contrast with most other kinds of grammar which
generate only idealized utterances or 'sentences'. This blurring of the boundary
between grammar and utterance is very controversial, but it follows inevitably
from the cognitive orientation of WG.
The status of utterances has a number of theoretical consequences both for
the structures generated and for the grammar that generates them. The most
obvious consequence is that word tokens must have different names from the
types of which they are tokens; in our example, the first word must not be
shown as / if this is also used as the name for the word-type in the grammar.
This follows from the fact that identical labels imply identity of concept,
whereas tokens and types are clearly distinct concepts. The WG convention is
to reserve conventional names for types, with tokens labelled 'wl', 'w2' and so
on through the utterance. Thus our example consists of wl and w2, which isa T
and 'AGREE: pres' respectively. This system allows two tokens of the same type
to be distinguished; so in / agree I made a mistake, wl and w3 both isa T. (For
simplicity WG diagrams in this chapter only respect this convention when it is
important to distinguish tokens from types. )
Another consequence of integrating utterances into the grammar is that
word types and tokens must have characteristics such that a token can inherit
them from its type. Obviously the token must have the familiar characteristics
of types - it must belong to a lexeme and a word class, it must have a sense and
a stem, and so on. But the implication goes in the other direction as well: the
type may mention some of the token's characteristics that are normally
excluded from grammar, such as characteristics of the speaker, the addressee
and the situation. This allows a principled account of deictic meaning (e. g. /
refers to the speaker, you to the addressee and now to the time of speaking), as
shown in Figure 1 and Table 1. Perhaps even more importantly, it is possible
to incorporate sociolinguistic information into the grammar, by indicating the
kind of person who is a typical speaker or addressee, or the typical situation of
use.
Treating utterances as part of the grammar has two further effects which are
important for the psycholinguistics of processing and of acquisition. As far as
processing is concerned, the main point is that WG accommodates deviant
input because the link between tokens and types is guided by the rather liberal
'Best Fit Principle' (Hudson 1990: 45ff): assume that the current token isa the
type that provides the best fit with everything that is known. The default
inheritance process which this triggers allows known characteristics of the token
to override those of the type; for example, a misspelled word such as mispelled
WHAT IS WORD GRAMMAK?
can isa its type, just like any other exception, though it will also be shown as a
deviant example. There is no need for the analysis to crash because of an error.
(Of course a WG grammar is not in itself a model of either production or
perception, but simply provides a network of knowledge which the processor
can exploit. ) Turning to learning, the similarity between tokens and types
means that learning can consist of nothing but the permanent storage of tokens
minus their utterance-specific content.
These remarks about utterances are summarized in Figure 7, which
speculates about my mental representation for the (written) 'utterance' Tons
mispelled it. According to this diagram, the grammar supplies two kinds of
utterance-based information about wl:
It also shows that w2 is a deviant token of the type 'MISSPELL: past'. (The
horizontal line below 'parts' is short-hand for a series of lines connecting the
individual letters directly to the morpheme, each with a distinct part name: part
1, part 2 and so on. )
Figure 7
18 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
7. Morphology
As explained earlier, the central role of the word automatically means that the
syntax is 'morphology-free'. Consequently it would be fundamentally against the
spirit of WG to follow transformational analyses in taking Jo snores as Jo 'tense'
snore. A morpheme for tense is not a word in any sense, so it cannot be a
syntactic node. The internal structure of words is handled almost entirely by
morphology. (The exception is the pattern found in clitics, which we return to
at the end of this section. )
The WG theory of inflectional morphology has developed considerably in
the last few years (Creider and Hudson 1998; Hudson 2000a) and is still
evolving. In contrast with the views expressed in EWG, I now distinguish sharply
between words, which are abstract, and forms, which are their concrete (visible
or audible) shapes; so I now accept the distinction between syntactic words and
phonological words (Rosta 1997) in all but terminology. The logic behind this
distinction is simple: if two words can share the same form, the form must be a
unit distinct from both. For example, we must recognize a morpheme {bear}
which is distinct from both the noun and the verb that share it (BEAR noun and
BEARvverb). This means that a word can never be directly related to phonemeserb). This means that a word can never be directly related to phonemes
and letters, in contrast with the EWG account where this was possible (e. g.
Hudson 1990: 90: 'whole of THEM = <them>'). Instead, words are mapped
to forms, and forms to phonemes and letters. A form is the 'shape' of a word,
and a phoneme or letter is a 'pronunciation' or 'spelling' of a form. In
Figure 7, for example, the verb MISSPELL has the form {misspell} as its stem (a
kind of shape), and the spelling of {misspell} is < misspell>.
In traditional terms, syntax, form and phonology define different 'levels of
language'. As in traditional structuralism, their basic units are distinct words,
morphemes and phoneme-type segments; and as in the European tradition,
morphemes combine to define larger units of form which are still distinct from
words. For example, {misspell} is clearly not a single morpheme, but it exists as
a unit of form which might be written {mis+spell} - two morphemes combining
to make a complex form - and similarly for {mis+spell+ed}, the shape of the
past tense of this verb. Notice that in this analysis {... } indicates forms, not
morphemes; morpheme boundaries are shown by '+'.
Where does morphology, as a part of the grammar, fit in? Inflectional
morphology is responsible for any differences between a word's stem - the
shape of its lexeme - and its whole - the complete shape. For example, the
stem of misspelled is {misspell}, so inflectional morphology explains the extra
suffix. Derivational morphology, on the other hand, explains the relations
between the stems of distinct lexemes - in this case, between the lexemes
SPELL and MISSPELL, whereby the stem of one is contained in the stem of
the other. The grammar therefore contains the following 'facts':
Figure 8
20 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
demands of syntax, as in the French example (3) where en would follow deux
(*Paul mange deux en) if it were not attached by cliticization to mange, giving a
single word-form en mange.
8. Syntax
As in most other theories, syntax is the best developed part of WG, which
offers explanations for most of the 'standard' complexities of syntax such as
extraction, raising, control, coordination, gapping and agreement. However the
WG view of syntax is particularly controversial because of its rejection of phrase
structure. WG belongs to the family of 'dependency-based' theories, in which
syntactic structure consists of dependencies between pairs of single words. As
we shall see below, WG also recognizes 'word-strings', but even these are not
the same as conventional phrases.
A syntactic dependency is a relationship between two words that are
connected by a syntactic rule. Every syntactic rule (except for those involved in
coordination) is 'carried' by a dependency, and every dependency carries at
least one rule that applies to both the dependent and its 'parent' (the word on
which it depends). These word-word dependencies form chains which link
every word ultimately to the word which is the head of the phrase or sentence;
consequently the individual links are asymmetrical, with one word depending
on the other for its link to the rest of the sentence. Of course in some cases the
direction of dependency is controversial; in particular, published WG analyses
of noun phrases have taken the determiner as head of the phrase, though this
analysis has been disputed and may turn out to be wrong (Van Langendonck
1994; Hudson 2004). The example in Figure 9 illustrates all these
characteristics of WG syntax.
A dependency analysis has many advantages over one based on phrase
structure. For example, it is easy to relate a verb to a lexically selected
preposition if they are directly connected by a dependency, as in the pair
consists of in Figure 9; but it is much less easy (and natural) to do so if the
preposition is part of a prepositional phrase. Such lexical interdependencies are
commonplace in language, so dependency analysis is particularly well suited to
descriptions which focus on 'constructions' - idiosyncratic patterns not covered
by the most general rules (Holmes and Hudson 2005). A surface dependency
analysis (explained below) can always be translated into a phrase structure by
building a phrase for each word consisting of that word plus the phrases of all
the words that depend on it (e. g. a sentence; of a sentence; and so on); but
22 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
KEY
WORD CLASSES
DEPENDENCY TYPES
Figure 9
Figure 10
This diagram also illustrates the notion 'surface structure' mentioned above.
Each dependency is licenced by the grammar network, but when the result is
structure-sharing, just one of these dependencies is drawn above the words; the
totality of dependencies drawn in this way constitutes the sentence's surface
structure. In principle any of the competing dependencies could be chosen, but
in general only one choice is compatible with the 'geometry' of a well-formed
surface structure, which must be free of 'tangling' (crossing dependencies - i. e.
discontinuous phrases) and 'dangling' (unintegrated words). There are no such
constraints on the non-surface dependencies. (For extensive discussion of how
this kind of analysis can be built into a parsing algorithm, see Hudson 2000c;
for a comparison with phrase-structure analyses of extraction, see Hudson
2003c. )
The second complication is coordination. The basis of coordination is
that conjuncts must share their 'external' dependencies - dependencies (if any)
to words outside the coordination. The structure of the coordination itself (in
terms of 'conjuncts' and 'coordinators') is analyzed in terms of 'word-strings',
simple undifferentiated strings of words whose internal organization is
described in terms of ordinary dependencies. A word string need not be a
phrase, but can consist of two (or more) mutually independent phrases as in the
example of Figure 11, where the coordination and conjuncts are bounded by
brackets: {[... ] [... ]}.
Unreal -words are the WG equivalent of 'empty categories' in other
theories. Until recently I have rejected such categories for lack of persuasive
evidence; for example, my claim has always been that verbs which appeared to
have no subject really didn't have any subject at all. So an imperative (Hurry!)
had no subject, rather than some kind of covert subject. However I am now
convinced that, for at least some languages, this is wrong. The evidence comes
from case-agreement between subjects and predicatives (WG sharers) in
24 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Figure II
languages such as Icelandic and Ancient Greek (Hudson 2003a); and the
conclusion is that some words have no realization (Creider and Hudson, this
volume). In this new analysis, therefore, an imperative verb does have a subject:
the word you. This is the ordinary word you, with its ordinary meaning, but
exceptionally, it is unrealized because this is what imperative verbs require of
their subjects. As Creider and I show, unrealized words may explain a wide
range of syntactic facts.
This discussion of syntax merely sets the scene for many other syntactic
topics, all of which now have reasonably well-motivated WG treatments: word
order, agreement, features, case-selection, 'zero' dependents. The most
important point made is probably the claim that in syntax the network
approach to language and cognition in general leads naturally to dependency
analysis rather than to phrase structure.
9. Semantics
As in any other theory, WG has a compositional semantics in which each word
in a sentence contributes some structure that is stored as its meaning. However,
these meanings are ordinary concepts which, like every other concept, are
defined by a network of links to other concepts. This means that there can be
no division between 'purely linguistic' meaning and 'encyclopedic' meaning.
For instance the lexemes APPLE and PEAR have distinct senses, the ordinary
concepts 'apple' and 'pear', each linked to its known characteristics in the
network of general knowledge. It would be impossible to distinguish them
merely by the labels 'apple' and 'pear' because (as we saw in section 3. 2) labels
on concepts are just optional mnemonics; the true definition of a concept is
provided by its various links to other concepts. The same is true of verb
meanings: for example, the sense of EAT is defined by its relationships to other
concepts such as 'put', 'mouth', 'chew5, 'swallow' and 'food'. The underlying
view of meaning is thus similar to Fillmore's Frame Semantics, in which lexical
meanings are denned in relation to conceptual 'frames' such as the one for
'commercial transaction' which is exploited by the definitions of 'buy', 'sell' and
so on. (See Hudson forthcoming for a WG analysis of commercial transaction
verbs. )
Like everything else in cognition, WG semantic structures form a network
with labelled links like those that are widely used in Artificial Intelligence. As in
JackendofFs Conceptual Semantics (1990), words of all word classes contribute
the same kind of semantic structure, which in WG is divided into 'sense5
(general categories) and 'referent' (the most specific individual or category
WHAT IS WORD GRAMMAR? 25
referred to). The contrast between these two kinds of meaning can be
compared with the contrast in morphology (section 7) between stem and whole:
a word's lexeme provides both its stem and its sense, while its inflection
provides its whole and its referent. For example, the word dogs is defined by a
combination of the lexeme DOG and the inflection 'plural', so it is classified as
'DOG: plural'. Its lexeme defines the sense, which is 'dog', the general concept
of a (typical) dog, while its inflection defines the referent as a set with more than
one member. As in other theories the semantics cannot identify the particular
set or individual which a word refers to on a particular occasion of use, and
which we shall call simply 'set s'; this identification process must be left to the
pragmatics. But the semantics does provide a detailed specification for what that
individual referent might be - in this case, a set, each of whose members is a
dog. One WG notation for the two kinds of meaning parallels that for the two
kinds of word-form: a straight line for the sense and the stem, which are both
retrieved directly from the lexicon, and a curved line for the referent and the
shape, which both have to be discovered by inference. The symmetry of these
relationships can be seen in Figure 12.
Figure 12
The way in which the meanings of the words in a sentence are combined is
guided by the syntax, but the semantic links are provided by the senses
themselves. Figure 13 gives the semantic structure for Dogs barked, where the
link between the word meanings is provided by 'bark', which has an 'agent' link
(often abbreviated 'er' in WG) to its subject's referent. If we call the particular
26 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
act of barking that this utterance refers to 'event-e', the semantic structure must
show that the agent of event-e is set-s. As with nouns, verb inflections contribute
directly to the definition of the referent, but a past-tense inflection does this by
limiting the event's time to some time ('tl') that preceded the moment of
speaking ('now5). Figure 13 shows all these relationships, with the two words
labelled VI' and 'w2'. For the sake of simplicity the diagram does not show
how these word tokens inherit their characteristics from their respective types.
Figure 13
but has not been able to illustrate the analyses that these ideas permit. In the
WG literature there are extensive discussions of lexical semantics, and some
explorations of quantification, definiteness and mood. However it has to be said
that the semantics of WG is much less well researched than its syntax.
10. Processing
The main achievements on processing are a theory of parsing and a theory of
syntactic difficulty; but current research is focused on a general theory of cognitive
processing in which language processing falls out as a particular case (Hudson, in
preparation). In this theory, processing is driven by a combination of spreading
activation, default inheritance and binding, like any other psychological model it
needs to be tested, and one step towards this has been taken by building two
computer systems called WGNet++ (see www. phon. ucl. ac. uk/home/WGNet/
wgnet++. htm) and Babbage (www. babbagenet. org) for experimenting with
complex networks.
The most obvious advantage of WG for a parser, compared with
transformational theories, is the lack of freely-occurring 'invisible' words (in
contrast with the unrealized words discussed above, which can always be
predicted from other realized words such as imperative verbs); but the
dependency basis also helps by allowing each incoming word to be integrated
with the words already processed without the need to build (or rebuild) higher
syntactic nodes.
A very simple algorithm guides the search for dependencies in a way that
guarantees a well-formed surface structure (in the sense defined in section 8):
the current word first tries to 'capture' the nearest non-dependent word as its
dependent, and if successful repeats the operation; then it tries to 'submit' as a
dependent to the nearest word that is not part of its own phrase (or, if
unsuccessful, to the word on which this word depends, and so on recursively up
the dependency chain); and finally it checks for coordination. (More details can
be found in Hudson 2000c. ) The algorithm is illustrated in the following
sequence of 'snapshots' in the parsing of Short sentences make good examples,
where the last word illustrates the algorithm best. The arrows indicate syntactic
dependencies without the usual labels; and it is to be understood that the
semantic structure is being built simultaneously, word by word. The structure
after ': -' is the output of the parser at that point
11. Conclusions
WG addresses questions from a number of different research traditions. As in
formal linguistics, it is concerned with the formal properties of language
structure; but it also shares with cognitive linguistics a focus on how these
structures are embedded in general cognition. Within syntax, it uses
dependencies rather than phrase structure but also recognizes the rich
structures that have been highlighted in the phrase-structure tradition. In
morphology it follows the European tradition which separates morphology
strictly from syntax, but also allows exceptional words which (thanks to
cliticization) contain the forms of smaller words. And so on through other areas
of language. Every theoretical decision is driven by two concerns: staying true to
the facts of language, and providing the simplest possible explanation for these
facts. The search for new insights is still continuing, and more cherished beliefs
may well have to be abandoned; but the most general conclusion so far seems
to be that language is mostly very much like other areas of cognition.
References
Aitchison, Jean (1987), Words in the Mind: An Introduction to the Mental Lexicon.
Oxford: Blackwell.
Altaian, Gerry (1997), The Ascent of Babel: An Exploration of Language, Mind and
Understanding. Oxford: Oxford University Press.
Anderson, John (1971), 'Dependency and grammatical functions'. Foundations of
Language, 7, 30-7.
Bates, Elizabeth, Elman, Jeffrey, Johnson, Mark, Karmiloff-Smith, Annette, Parisi,
WHAT IS WORD GRAMMAR? 29
Theory and Corpus Studies: A Dialogue from 10 ICEHL. Berlin: Mouton de Gruyter,
pp. 53-75.
— (2001), 'The stative/dynamic contrast and argument linking'. Language Sciences, 23,
603-37.
Goldberg, Adele (1995), Constructions: A Construction Grammar Approach to Argument
Structure. Chicago: University of Chicago Press.
Gorayska, Barbara (1985), 'The Semantics and Pragmatics of English and Polish with
Reference to Aspect' (Unpublished doctoral dissertation, UCL).
Halliday, Michael (1961), 'Categories of the theory of grammar'. Word, 17, 241-92.
— (1967-8), 'Notes on transitivity and theme in English'. Journal of Linguistics, 3, 37-
82, 199-244; 4, 179-216.
— (1985), An Introduction to Functional Grammar. London: Arnold.
Hiranuma, So (1999), 'Syntactic difficulty in English and Japanese: A textual study'.
UCL Working Papers in Linguistics, 11, 309-21.
— (2001), 'The Syntactic Difficulty of Japanese Sentences' (Unpublished doctoral
dissertation, UCL).
Holmes, Jasper (2004), 'Lexical Properties of English Verbs' (Unpublished doctoral
dissertation, UCL).
Holmes, Jasper and Hudson, Richard (2005), 'Constructions in Word Grammar', in
Jan-Ola Ostman and Mirjam Fried (eds), Construction Grammars: Cognitive Grounding
and Theoretical Extensions. Amsterdam: Benjamins, pp. 243-72.
Hudson, Richard (1964), 'A Grammatical Analysis of Beja' (Unpublished doctoral
dissertation, University of London).
— (1970), English Complex Sentences: An Introduction to Systemic Grammar/Amsterdam:
Nordi-Holland.
— (1976), Arguments for a Non-transformational Grammar. Chicago: University of
Chicago Press.
— (1980a), Sociolinguistics. Cambridge: Cambridge University Press.
— (1980b), 'Constituency and dependency'. Linguistics, 18, 179-98.
— (1981), 'Panlexicalism'. Journal of Literary Semantics, 10, 67-78.
- (1984), Word Grammar. Oxford: BlackweU.
— (1989), 'Towards a computer-testable Word Grammar of English'. UCL Working
Papers in Linguistics, 1, 321-39.
— (1990), English Word Grammar. Oxford: Blackwell.
— (1992), 'Raising in syntax, semantics and cognition', in Iggy Roca (ed. ), Thematic
Structure: Its Role in Grammar. The Hague: Mouton, pp. 175-98.
— (1993), 'Do we have heads in our minds?', in Greville Corbett, Scott McGlashen and
Norman Fraser (eds), Heads in Grammatical Theory. Cambridge: Cambridge
University Press, pp. 266-91.
— (1995), Word Meaning. London: Roudedge.
— (1996a), Sociolinguistics (2nd edition). Cambridge: Cambridge University Press.
— (1996b), 'The difficulty of (so-called) self-embedded structures'. UCL Working Papers
in Linguistics, 8, 283-314.
— (1997a), 'The rise of auxiliary DO: verb non-raising or category-strengthening?'.
Transactions of the Philological Society, 95, 41-72.
— (1997b), 'Inherent variability and linguistic theory'. Cognitive Linguistics, 8, 73-108.
— (1998), English Grammar. London: Routiedge.
- (2000a), 'VflrowY. Language, 76, 297-323.
— (2000b), 'Gerunds and multiple default inheritance'. UCL Working Papers in
Linguistics, 12, 303-35.
— (2000c), 'Discontinuity'. Traitement Automatique des Langues, 41, 15-56.
WHAT IS WORD GRAMMAR? 31
- (2001), 'Cities in Word Grammar'. UCL Working Papers in Linguistics, 13, 293-4.
— (2003a), 'Case-agreement, PRO and structure-sharing'. Research in Language, 1, 7-33.
— (2003b), 'Mismatches in default inheritance', in Elaine Francis and Laura Michaelis
(eds), Linguistic Mismatch: Scope and Theory. Stanford: CSLI, pp. 269-317.
- (2003c), 'Trouble on the left periphery'. Lingua, 113, 607-42.
— (2004), 'Are determiners heads?'. Functions of Language,, 11, 7-43.
— (forthcoming), 'Buying and selling in Word Grammar', in Jozsef Andor and Peter
Pelyvas (eds), Empirical, Cognitive-Based Studies In The Semantics-Pragmatics
Interface. Oxford: Elsevier.
— (in preparation) Advances in Word Grammar. Oxford: Oxford University Press.
Hudson, Richard and Holmes, Jasper (2000) 'Re-cycling in the Encyclopedia', in Bert
Peeters (ed. ), The Lexicon/Encyclopedia Interface. Oxford: Elsevier, pp. 259-90.
Jackendoff, Ray (1990), Semantic Structures. Cambridge, MA: MIT Press.
Kamiloff-Smith, Annette (1992), Beyond Modularity: A developmental perspective on
cognitive science. Cambridge, Mass.: MIT Press.
Kaplan, Ron and Bresnan, Joan (1982), 'Lexical-functional Grammar: a formal system
for grammatical representation', in Joan Bresnan (ed. ), The Mental Representation of
Grammatical Relations. Cambridge, MA: MIT Press, pp. 173-281.
Kreps, Christian (1997), 'Extraction, Movement and Dependency Theory' (Unpub-
lished doctoral dissertation, UCL).
Lakoff, George (1987), Women, Fire and Dangerous Things: What Categories Reveal
about the Mind. Chicago: University of Chicago Press.
Lamb, Sidney (1966), An Outline of Stratificational Grammar. Washington, DC:
Georgetown University Press.
— (1999), Pathways of the Brain: The Neurocognitive Basis of Language. Amsterdam:
Benjamins.
Langacker, Ronald (1987), Foundations of Cognitive Grammar I. Theoretical Prerequisites.
Stanford: Stanford University Press.
— (1990), Concept, Image and Symbol. The Cognitive Basis of Grammar. Berlin: De Gruyter.
Luger, George and Stubblefield, William (1993), Artificial Intelligence. Structures and
Strategies for Complex Problem Solving. Redwood City, CA: Benjamin/Cummings
Pub. Co.
Lyons, John (1977), Semantics. Cambridge: Cambridge University Press.
McCawley, James (1968), 'Concerning the base component of a transformational
grammar'. Foundations of Language, 4, 243-69.
Pinker, Steven (1994), The Language Instinct. Harmondsworth: Penguin Books.
Pinker, Steven and Prince, Alan (1988), 'On language and connectionism: Analysis of a
Parallel Distributed Processing model of language acquisition'. Cognition, 28, 73-193.
Pollard, Carl and Sag, Ivan (1994), Head-driven Phrase Structure Grammar. Chicago:
University of Chicago Press.
Reisberg, Daniel (1997), Cognition. Exploring the Science of the Mind. New York: W. W.
Norton.
Robinson, Jane (1970), 'Dependency structures and transformational rules'. Language,
46, 259-85.
Rosta, Andrew (1994), 'Dependency and grammatical relations'. UCL Working Papers
in Linguistics, 6, 219-58.
— (1996), 'S-dependency'. UCL Working Papers in Linguistics, 8, 387-421.
— (1997), English Syntax and Word Grammar Theory. Unpublished doctoral dissertation,
UCL, London.
Shaumyan, Olga (1995), Parsing English with Word Grammar. Imperial College London
MSc Thesis.
32 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Abstract
In Ancient Greek a predicative adjective or noun agrees in case with the subject
of its clause, even if the latter is covert. This provides compelling evidence for
'empty' (i. e. covert) elements in syntax, contrary to the tradition in WG theory.
We present an analysis of empty elements which exploits a feature unique to
WG, the separation of 'existence' propositions from propositions dealing with
other properties; an empty word has the property of not existing (or, more
technically, 0 quantity). We contrast this theory with the Chomskyan PRO and
the Head-driven Phrase Structure Grammar (HPSG) 'potential' SUBCAT list
1. Introduction
Case agreement in Ancient Greek1 has attracted a small but varied set of
treatments in the generative tradition (Andrews 1971; Lecarme 1978; Quicoli
1982). In this literature the problems were framed and solved in transforma-
tional frameworks. In the present chapter we wish to consider the data from the
point of view of the problems they pose for a theory of case assignment and
phonologically empty elements in a modern, declarative framework - Word
Grammar (WG; Hudson 1990). We present an analysis of empty elements
which exploits a feature unique to WG, the separation of existence propositions
from propositions dealing with other properties; and we contrast it with earlier
WG analyses in which these 'empty' elements are simply absent and with
Chomskyan analyses in terms of PRO, a specific linguistic item which is always
covert. The proposed analysis is similar in some respects to the one proposed
by Pollard and Sag (1994) for HPSG.
2. The Data
We confine our attention to infinitival constructions. The infinitive in Ancient
Greek is not inflected for person, number or case and hence, when predicate
adjectives and predicate nominals appear as complements of infinitives, it is
necessary to account for the case of these elements. One purpose of this
discussion is to show that traditional grammars are right to explain the case of
predicates in terms of agreement with the subject, but this analysis works most
36 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
naturally if we also assume some kind of 'null' subject for some infinitives. The
examples that support the null subject take the accusative case, and are
discussed in section 2. 1; there are well-known exceptions which are traditionally
described in terms of 'attraction' and which are discussed in section 2. 2.
2. 1. Accusative subjects
Traditional grammars of Greek state that the subject of an infinitive takes the
accusative case.
Examples are usually given of the accusative plus infinitive construction as in
the following:
Figure 1
Figure 2
However, note that the predicative may be accusative even when the accusative
subject is itself absent:
This can also be true even when there is a coreferential element in the higher
clause:
Figure 3
38 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
A partial structure for (6) is presented in Figure 3. Such examples raise urgent
questions about the status of 'understood' subjects. If an understood subject is
simply one which is 'understood' in the semantics but entirely absent from the
syntax, it is hard to explain the case of the predicative in these examples. We
return to these questions below.
When the subject of the infinitive is identical to that of the main verb, it is
normally not expressed:
Examples like (8) are hard to explain without assuming some kind of accusative
subject for the infinitive with which the predicative participle ('taking') can
agree, as hinted at in Figure 4; but of course there is no overt accusative subject.
Figure 4
(Other examples for Homeric Greek in II. 7. 198, 13. 269, 20. 361 - Chantraine
1953: 312. )
The emphasis need not be strong, as the following example, with unstressed
clitic pronoun, shows:
When the infinitive is used in exclamations with an overt subject, the latter
appears in the accusative:
These examples with overt accusative subjects strongly support the traditional
rule that infinitives have accusative subjects, so the question is how to allow this
generalization to extend to infinitives which appear to have no subject at all in
order to explain the accusative cases found on predicatives in infinitival clauses
in examples such as (4) and (5).
2. 2 Non-accusative subjects
Greek provides a number of interesting alternatives to the possibilities given in
section 2. 1. These are traditionally discussed under two headings, although the
process is the same in both cases:
Sehr viele der Verben, die den Infinitive zu sich nehmen, haben daneben noch ein
personliches Objekt bei sich, welches in dem Kasus steht, den das Verb erfordert...
Wenn zu dem Infinitive adjektivische oder substantivische Pradikatsbestimmungen
treten, so stehen dieselben entweder vermittelst einer Attraktion mit dem
personlichen Objekte in gleichem Kasus oder mit Vernachlassigung der Attraktion
im Akkusative (Kuhner and Gerth 1955: 24)3
Wenn aber das Subjekt des regierenden Verbs zugleich auch das Subjekt des
Infinitivs ist, so wird das Subjekt des Infmitivs... weggelassen, und wenn adjektivische
oder substantivische Pradiskatsbestimmungen bei dem Infinitive stehen, so werden
diese vermittelst der Attraktion in den Nominative gesetzt (ibid.: 29)4
40 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Figure 5
This analysis easily explains why the lower nominal predicate 'helpers' has
the same case as this shared nominal, but it does not help with examples where
even the higher nominal is merely implicit, as in (16). (We give an explanation
for examples of this type in section 6. )
(17) I saw him yesterday while he was on his way to the beach.
In (17) the three pronouns share a common set of semantic features (gender
and number) and have a common referent, but occur respectively in the
'objective', 'subjective' and 'possessive' cases (to use the terms of traditional
grammar). So far as we know, semantic co-reference between a nominal and a
pronoun never triggers case agreement in the latter, though it often requires
agreement for number and gender. A further consequence is that the only
possible 'target' for case agreement is a word (or other syntactic element); this
rules out a semantic account of case agreement, however attractive such an
account might be for number and gender.
Thus, faced with examples such as (18=6), where an infinitive has an
accusative predicative but no overt subject, we cannot explain the predicative's
case merely by postulating a subject argument in the semantic structure without
a supporting syntactic subject.
The argument X in the semantic structure 'X becoming a king' cannot by itself
carry the accusative case; there must also be some kind of accusative subject
nominal in the syntactic structure. Nor does a control analysis help, because the
controlling nominal is the pronoun soi, 'to you', which is dative; as expected, its
case has nothing to do with that of coreferential nominals.
The analysis seems to show that a specifically syntactic subject must be
present in order to account for the accusative case seen on the predicate
nominal. We accept this conclusion, though somewhat reluctantly because it
conflicts with the stress on concreteness which we consider an important
principle of Word Grammar. We are sceptical about the proliferation of empty
nodes in Chomskyan theories, and have always felt that the evidence for such
nodes rested heavily on theory-internal assumptions which we did not share. In
contrast, the evidence from case agreement strikes us as very persuasive, so we
now believe that syntactic structure may contain some 'null' elements which are
42 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Here gi is the mutated form of ci, 'dog', whose form shows it to be object rather
than subject even when the optional subject i is omitted. Conversely, however,
subjects are also mutated if they are delayed, as in (20):
Figure 6
The relevant grammar is in Figure 7, which includes the very general default
tendency for words to have a realization.
Figure 7
Figure 8
5. 2 'object pro-drop*
(23) ou dei tois paidotribais enkalein oud' ekballein ek
not necessary the(dat) trainers (dat) to-accuse nor to-banish from
to: n poleo: n
the (gen) cities (gen)
'it is not necessary to accuse the trainers nor to banish them from the cities'
(P. G. 460 d. )
This phenomenon, less common than 'subject pro-drop' but very common in
Greek, is traditionally analyzed under the rubric of 'object-sharing' and has no
agreed modern analysis. Treating the 'omitted' object as unrealized provides a
natural and simple account. Note that the traditional shared object analysis
would incorrectly associate the dative case with the object of ekballein (normally
accusative in this context).
5. 3 Subjects of imperatives
In languages where these are usually absent, such as English, the identity of the
unreal subject is very clear: as we assumed above, it must be the pronoun you
for second-person imperatives, and we for first-person plural ones. This is clear
not only from the meaning but also from the choice of pronoun in the tag
question:
has been challenged (Wierzbicka 1988) but we still find it especially plausible
for examples such as scales (plural) contrasting with balance (singular). The
relevant datum is that the choice between this and these matches the syntactic
number when the complement noun is overt (so this balance but these scales],
but the same choice is made even when there is no overt complement
(26) I need some scales to weigh myself on, but these (*this) are (*is) broken.
If pluralia tantum really are singular in meaning, we cannot explain this choice
in terms of meaning, and the most attractive explanation is that the choice is
forced in the same way as in the overt case, by the presence of an unrealized
example of the noun scales (or trousers or whatever). We might also consider
extending this explanation to another curious fact about the demonstrative
pronouns, which is that the singular form can only refer to a thing:
The explanation would be that only one unrealized noun is possible in the
singular: the noun thing. The analyses that we are suggesting are of course
controversial and may be wrong, but if they are correct then they show that the
unrealized word may be a specific lexical noun rather than a general-purpose
pronoun as in the earlier examples.
This covers the territory of so-called 'VP deletion' but also other kinds of
anaphoric ellipsis:
(28) I don't know whether I'm going to finish my thesis this year, but I may.
(29) I may finish my thesis this year, but I don't know for sure.
(30) OK, you didn't enjoy that book, but here's one which you will.
If will has no syntactic complement at all, the extraction of which in (30) is very
hard to explain; but if its complement is an unrealized enjoy, the rest of the
syntactic structure can be exactly as for here's one which you will enjoy., 6
In all these examples the omitted element is redundant and easy to recover,
so the option of leaving it unsaid obviously helps both the speaker and the
CASE AGREEMENT IN ANCIENT GREEK 49
hearer. The familiar functional pressure to minimize effort thus explains why
the choice between 'realized' and 'unrealized' exists in the grammar. On the
other hand, it does not explain why languages allow it in different places - e. g.
why some languages allow tensed verbs to have a null subject while others do
not This variation must be due to different ways of resolving the conflict
between this functional pressure and others which push in the opposite
direction, such as the pressure to make syntactic relations reliably identifiable
(whether by word order as in English or by inflectional morphology as in
Greek).
• PRO and pro are special pronouns which combine the peculiarity of always
being covert with the equally exceptional property of covering all persons,
numbers and genders. The fact that they are exceptional in two such major
respects should arouse suspicion. In contrast, our unrealized pronouns are
the ordinary pronouns - he, me, us, and so on - which just happen to be
unrealized. Even if we count this as an exceptional feature it is their only
exceptional feature, in contrast with the double exceptionality of PRO and
pro.
• In our account, a pronoun may be realized for emphasis in contexts where
it would normally be unrealized; this accounts in a simple way for the
examples in (9) to (12) where the pronoun is emphatic. If the covert
pronoun is always PRO, why should it always alternate with an ordinary
pronoun?
• Our unrealized words need not be pronouns, unlike PRO and pro. As
explained in the previous section, this allows us to extend the same
explanation to other kinds of unexpressed words, such as unrealized
common nouns acting as complement of a pronoun/determiner such as
this, or virtual complements of verbs such as auxiliary verbs. In other words,
our proposal subsumes null subjects under a much broader analysis which
covers ellipsis in general.
• Unrealized words are identified by the quantity feature which applies
outside language (e. g. to Father Christmas and feet) as well as inside. In
contrast, the difference between PRO or pro and other words is specific to
language, involving (presumably) the absence of a phonological entry. Any
explanation which involves machinery that is available independently of
language is preferable to one which involves special machinery.
• In the standard analysis with PRO and pro the difference between these two
is important because both abstract 'Case' and surface case are supposed to
be impossible for PRO but obligatory for pro; this contrast is also claimed to
correlate with the contrast between subjects of non-finite and finite verbs.
50 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
More recently, PRO has been claimed to have a special 'null' case
(Chomsky and Lasnik 1993). The empirical basis for these claims was
undermined long ago (e. g. Sigurdsson 1991), and our analysis does not
recognize the distinction between PRO and pro. Unrealized pronouns all
take case (or lack it) just like realized ones in the language concerned.
These differences between our proposal and the PRO/pro system all seem to
favour our proposal.
• its referent is the 'active argument' of the verb as defined by the latter's
lexical entry - for instance, with RUN/TREKHO: it is the runner, with
FALL/PIPTO: it is the fuller, with LIKE/PHILEO: it is the liker, and so
on;
• in English, it typically stands before the verb;
• it is the typical antecedent of a reflexive object of the same verb;
• the verb agrees with it;
• in English, it is obligatory if the verb is tensed;
• it is also the verb's object if the verb is passive;
• in Greek, its case is typically nominative.
nothing to bring them all together. For example, if himself is the object of hurt it
is tied anaphorically to the hurter via the 'subject' link, but if there is no subject
this link disappears. And yet the fact is that the anaphoric relations are exactly
the same regardless of whether or not there is an overt subject; for example, the
'understood' subject of hurt in Don't hurt yourself! binds yourself in exactly the
same way as the overt one in You may hurt yourself.
The analysis that we are proposing solves these problems by moving towards
the standard view that every verb does indeed have a subject, whether or not
this is overt. Similar problems face the earlier WG approach in other areas of
grammar, and can be solved in the same way. In section 5 we outlined a range
of phenomena that seem to call for analysis in terms of unrealized words, and
which more traditional WG analyses have treated in terms of dependents that
are simply absent.
Another attempt to handle null subjects without invoking PRO is proposed
by Pollard and Sag (1994: 123-45) in the framework of HPSG. As with the
early WG analysis just described, this proposal applies only where syntactic
structure-sharing is not possible. They propose a structure for 'Equi' verbs such
as the following for try (ibid.: 135):
The infinitive's subject is the italicized 'NP' in its 'SUBCAT' (valency) list This
NP merely indicates the need for a subject, and would normally be 'cancelled'
(satisfied) by unification with an NP outside the VP; for example, in They
worked hard the verb needs a subject, which is provided by they. However at
least the intention of this entry is to prevent the need from being satisfied, so
that the infinitive's subject remains unrealized, as in our proposed WG analysis.
Moreover, this unrealized subject in the SUBCAT list may carry other
characteristics which are imposed both by the infinitive and by try; for example,
the subscripts in the entry for try show that it must be coreferential with the
subject of try - i. e. a sentence such as They try to work hard has only one
meaning, in which they are the workers as well as the try-ers. Most importantly
for the analysis of Greek, the unrealized subject can carry whatever case may be
imposed on it by the infinitive (Henniss 1989). Consequently it can be the
target of predicative case agreement, so Ancient Greek case-agreement would
be no problem.
However there are also significant differences between the two proposals.
• The null NP in HPSG is purely schematic, so all null subjects have the
same syntax (bar any specific syntactic demands imposed by the infinitive).
They are also schematic in their semantics, in spite of the coreference
restriction, because reference is distinct from semantics (e. g. the winner may
be coreferential with someone I met last night, but these phrases obviously
have different semantic structures). In contrast, WG null subjects are
ordinary lexical nouns and pronouns.
• So far as we can see, the HPSG machinery for distinguishing overt and
covert valents does not appear to generalize beyond language; and indeed,
many advocates of HPSG might argue that it should not do so. In contrast,
we showed in section 4 that our proposal does; it can explain the 'non-
occurrence' of Father Christmas in just the same way as that of the subject
of an infinitive.
8. Conclusions
The most important conclusion is that where there is strong empirical
evidence for null elements, they can easily be included even in a 'surfacist'
grammar such as WG. This can be done by exploiting the existing WG
machinery for determining 'quantity', a variable which guides the user in
applying knowledge to experience; for example, one of the properties that we
attribute to Father Christmas is zero quantity - i. e. we expect no tokens in
experience. In these terms, a 'null word' is an ordinary word whose realization
has the quantity 0 - an unrealized word. This (or something like it) is
generally available in cognition both for distinguishing fact and fiction and for
cases where an expected attribute is exceptionally absent, so it comes 'for
free', and it is preferable to inventing special linguistic inaudibilia such as
PRO or pro.
References
Andrews, A. (1971), 'Case agreement of predicate modifiers in Ancient Greek'.
Linguistic Inquiry, 2, 127-51.
Borsley, R. (2005), 'Agreement, mutation and missing NPs in Welsh', Available: http: //
privatewww. essex. ac. uk/~rborsley/Agreement-paper. pdf (Accessed: 19 April 2005).
Chantraine, P. (1953), Grammaire Homerique. Vol 2. Paris: Klincksieck.
Chomsky, N. and Lasnik, H. (1993), 'The theory of principles and parameters', in J.
Jacobs, A. v. Stechow, W. Sternefeld and T. Venneman (eds), Syntax: An International
Handbook of Contemporary Research. Berlin: Walter de Gruyter, pp. 506-69.
Goodwin, W. W. (1930), Greek Grammar (rev. Charles Burton Gulick). Boston: Ginn.
Henniss, K. (1989), '"Covert" subjects and determinate case: evidence from
Malayalam', in J. Fee and K. Hunt (eds), Proceedings of the West Coast Conference
on Formal Linguistics. Stanford: CSLI, pp. 167-75.
Hudson, R. (1984), Word Grammar. Oxford: Blackwell.
— (1990), English Word Grammar. Oxford: Blackwell.
— (2003), 'Case-agreement, PRO and structure sharing'. Research in Language, 1, 7-33.
Keenan, E. L. (1976), 'Towards a universal definition of "subject"', in Charles Li (ed. ),
Subject and Topic. New York: Academic Press, pp. 303-33.
Kiihner, R. and Gerth, B. (1955), Ausfuhrliche Grammatik der griechischen Sprache.
Leverkusen: Gottschaksche Verlagsbuchhandlung.
Lecarme, J. (1978), 'Aspects Syntaxiques des Completives du Grec' (Unpublished
doctoral dissertation, University of Montreal).
Liddell, H. G. and Scott, R. (1971), A Greek-English Lexicon (9th edn, rev. H. Jones and
R. McKenzie, suppl. by E. Barber) Oxford: Clarendon Press.
Pollard, C. J. and Sag, I. A. (1994), Head-Driven Phrase Structure Grammar. Chicago:
University of Chicago Press.
Quicoli, A. C. (1982), The Structure of Complementation. Ghent: Story-Scientia.
Sigurdsson, H. (1991), 'Icelandic case-marked PRO and the licensing of lexical
arguments'. Natural Language & Linguistic Theory, 9, 327-63.
Smyth, H. W. (1956), Greek Grammar (rev. G. Messing). Cambridge, MA: Harvard
University Press.
Wierzbicka, A. (1988), The Semantics of Grammar. Amsterdam: Benjamins.
Notes
1 By Ancient Greek we mean the Greek of early epic poetry ('Homeric Greek') down
to the Attic prose of the 5th and 4th centuries B. C. E.
2 A list of abbreviated references to classical authors can be found at the end of this
paper.
3 Very many of the verbs which take the infinitive also take a personal object which
stands in the case that the verb requires... If the infinitive also has an adjectival or
nominal predicate, this stands in the same case as the personal object by (an)
attraction, or in the absence of attraction, in the accusative.
4 However if the subject of the governing verb is at the same time the subject of the
infinitive, the subject of the infinitive is omitted, and if adjectival or nominal
predicates accompany the infinitive, these are put in the nominative by attraction.
5 This felicitous term is taken from the work of Pollard and Sag (1994), but the
analysis was worked out in full detail for English infinitives (and other constructions)
in Hudson (1990: 235-9).
6 We owe this point to Andrew Rosta.
3 Understood Objects in English and Japanese with
Reference to Eat and Taberu: A Word Grammar
account
KENSEI SUGAYAMA
Abstract
The author argues that there is a semantic difference of the suppressed object
between eat and its Japanese equivalent taberu. Then what kind of semantic structure
would the Japanese verb taberu have? This chapter is an attempt to answer this
question in the framework of Word Grammar (Hudson 1984, 1990, 1998, 2005).
1. Introduction
Unlike. English and perhaps most other European languages, Japanese allows
its transitive verbs to miss out their complements (e. g. subject, object) on the
condition that the speaker assumes that they are known to the addressee.1 This
is instantiated by the contrast in (1) and (2):
Sentences (3), (4) and (5) are colloquial and quite often used in the standard
spoken Japanese. In this sense they are not marked sentences. In (3) only the
subject is left out, while in (4) both the subject and object are left out as shown
in the word-for-word translation. Sentence (5) involves the transitive verb oita, a
past tense form of oku 'put', which corresponds to put in English. Oku is a three-
place predicate which semantically requires three arguments [agent, theme,
place]. These three arguments are mapped syntactically to subject, object and
place adverbial, respectively. Quite interestingly (5) shows that the place
element, which is also considered to be a complement (or adjunct-complement)
of the verb oku, is optional when it is known to the addressee, which is virtually
impossible with its counterpart put in English. Although these complements are
in fact missed out (i. e. unexpressed or ungrammaticalized), the addressee
eventually will come to an appropriate interpretation of each sentence where
unexpressed complements are supplied semantically or pragmatically and they
are no doubt given full interpretation. Why is this possible? A possible answer
comes from the assumption that in the semantic structure of the sentences
above, there has to be a semantic argument which should be, but is not actually,
mapped onto the syntactic structure (i. e. grammaticalized as a syntactic
complement in my terms).
Turning to English, on the other hand, it is possible to leave indefinite
objects suppressed for semantically limited verbs such as eat, drink, read, etc. 3
Thus, following Hudson (2005), the syntactic and semantic structure of John ate
will be something like the one shown in Figure 1.
Figure 1
The links between syntactic and semantic structures in Figure 1 are shown by
the vertical solid and curved lines. The categories enclosed by single quotation
marks (e. g. 'John', 'John ate') are concepts which are part of the sentence's
semantic structure; the numbers are arbitrary. Detailed explanation about
syntactic and semantic dependencies will be given in the next section.
But this kind of semantic structure does not seem to be a viable one for the
Japanese verb taberu, because, as I will argue later, there is a semantic difference
in the semantic feature of the suppressed object between eat and taberu, which
does not seem to be properly reflected in the semantic structure of those two
56 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
verbs in Word Grammar (WG). Then what kind of semantic structure will the
Japanese verb tab em, the Japanese equivalent of eat, have? This chapter is an
attempt to answer this question in the framework of Word Grammar.
The rest of the chapter is organized in the following way. Section 2
introduces Word Grammar and deals with the relevant notions used in WG to
deal with the problem of a covert object. Section 3 discusses the analysis of an
intransitive use of the eat type verbs. Section 4 discusses the Japanese verb
taberu, an equivalent of eat in English. It also discusses the interpretation of
tab em which lacks an overt object, using the syntactic and semantic analysis in
WG. Section 5 offers my own account of how taberu is more adequately
described in the semantic structure in WG.
2. Word Grammar
Before continuing any further, let us first have a brief look at the basic
framework of WG and its main characteristics. WG, which is fully developed
and formalized in Hudson (1984, 1990), subsequently revised by Rosta (1997),
is to be taken as a lexicalist grammatical theory because the word is central -
hence the name of the theory, basically making no reference to any
grammatical unit larger than a word.4 In his recent comparison of WG with
Head-driven Phrase Structure Grammar (HPSG), Hudson (1995b: 4) gives a
list of the common characteristics between the two theoretically different
grammars, some relevant ones of which are repeated here for convenience:
(6) a. both (i. e. WG and HPSG) include a rich semantic structure parallel
with the syntactic structure;
b. both are monostratal;
c. both make use of inheritance in generating structures;
d. neither relies on tree geometry to distinguish grammatical functions;
e. both include contextual information about the utterance event (e. g. the
identities of speaker and hearer) in the linguistic structure.
Figure 2
These facts are self-explanatory except for a few technical expressions: 'A isa B'
means 'A is an instance of B' and [0-1] in front of object means 'at least 0 and
at most 1' (i. e. 0 or one in this particular case). An element in italics is meant to
be the antecedent of the pronoun.
Propositions in (7) partially represent the syntactic and semantic structures of
John ate potatoes diagrammed in Figure 3 with eat replaced by its past tense form
ate.
Figure 3
3. Eat in English
Let us consider those English transitive verbs that optionally appear without
their object Examples of such verbs, among others, include dress, shave, lose,
win, eat and read, as in (8a) to (8f):
While these six verbs will take part in the same syntactic alternation, the
intransitive verbs are obviously interpreted in different ways. In the examples
(8d)-(8f), we can identify three verb classes according to difference in
UNDERSTOOD OBJECTS IN ENGLISH AND JAPANESE 59
For the shave type verbs, the object, if omitted, is interpreted as being coreferent
with the subject or the subject's specific body part. For the win type verbs, the
surface intransitive verb form signals a severe narrowing of the range of possible
referents of the implicit object, roughly speaking, 'a specific game' to be
recoverable from the context. For the eat type verbs, the intransitive form
means a lack of commitment by the speaker to the referent of the object.
With the eat type verbs, the identity of the referent of the unexpressed object
may be non-specific, i. e. literally unknown to the speaker, because the
sentences in (10) do make sense.
(10) a. I saw Oliver eating, but I don't know what he was eating
b. When I peeked into Oliver's room, he was reading; now I wonder what
he was reading
In both sentences in (10), the identity of what was eaten/read is asked in the
second part. This implies that the patient (or eat-<?£/read-^) argument of eat/
read, which may be grammaticalized as the object at surface structure, does not
have to be definite.
There is other evidence that supports the indefmiteness of the suppressed
object of eat. Consider the following dialogues:
In both (11) and (12), speaker B cannot reply to speaker A by using ate without
its object. What (11) and (12) suggest is that eat, when its object is suppressed,
cannot have its null object referring to the element in the previous discourse,
which, as I will explain very shortly, is in fact possible in Japanese. What the
ungrammaticality of the utterances by speaker B indicates is that the understood
object has to be indefinite if eat is used as an intransitive verb.7
Here is another interesting piece of evidence supporting that this claim is
true. Observe the following dialogue:
When speaker A first uses the intransitive eat, it is clear that he/she does not
have a definite object (or the referent of a definite object) in mind, and is just
60 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Figure 4
By now it is clear that this semantic representation is inadequate for eat (ate}.
It is not so difficult to see that the important semantic information of the
suppressed object is missing in Figure 4. In Fred ate, there is no object, which
implies that it should be indefinite. Regrettably, this important semantic
information does not seem to be incorporated in Figure 4. Therefore, my
proposal is that we have to revise the diagram so that it can be enriched with the
semantic information of the unexpressed object and accordingly a more
accurate one will be something like the one in Figure 5.
Figure 5
4. Taberu in Japanese
Let us now turn to a Japanese counterpart of eat, taberu. The picture of taberu,
an equivalent of eat in Japanese, is quite different from that of English eat,
UNDERSTOOD OBJECTS IN ENGLISH AND JAPANESE 61
Figure 6
In the last section I argued that the suppressed object of eat is indefinite
because it cannot refer to its antecedent even when it is available in the
preceding context. Interestingly enough the opposite is true with taberu.
Consider die following dialogues corresponding to (11) and (12), which we
discussed in the last section:
In (18) and (19), the definite object referring to an element in die previous
context is left out in B's utterance. In (19), the subject referring to the speaker is
also missing in B's utterance. These cases show diat the suppressed object of
taberu is definite. In contrast, when the object of taberu is indefinite, there are in
fact cases where it has to be expressed as an indefinite noun as in (20):
These arguments make it very clear that WG should represent the syntactic and
semantic structures of inuga tabeta in (18) as diagrammed in Figure 7.
As I stated in section 2, Hudson (1995b) claims that one of the key
characteristics of WG is taken as 'including contextual information about the
utterance event' (e. g. the identities of a speaker and a hearer) in the linguistic
structure. However, as it stands, the syntactic and semantic structures in Figures
UNDERSTOOD OBJECTS IN ENGLISH AND JAPANESE 63
Figure 7
5. Conclusion
Considering the fact I mentioned above of deletability of definite objects given a
proper context in Japanese, it seems reasonable to add the following rule to the
grammar of Japanese to explain the proper semantic structure of tab em:
(21) Knower of eai-ee of sense of tab em = addressee of it.
Taking into account the arguments above, I conclude that the syntax of missing
complements in Japanese can be given a more satisfactory description by
introducing a parameter of 'default definiteness'. I do not want to enter into
details now but simply suggest that to distinguish between complements and
adjuncts in Japanese one needs another parameter such as 'default
definiteness'. To put it differently, by default the definiteness of a covert
complement is [+defmite] and that of a covert adjunct is [+/-definite] as in (22):
(22) • Covert complement of verb = definite
• Covert adjunct of verb = indefinite
• Knower of referent of complement of verb = addressee of it
References
Allerton, D. J. (1982), Valency and the English Verb. London: Academic Press.
Cote, S. A. (1996), 'Grammatical and Discourse Properties of Null Elements in
English'. (Unpublished doctoral dissertation, University of Pennsylvania).
64 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Fillmore, Ch. J. (1986), 'Pragmatically controlled zero anaphora'. BLS 12, 95-107.
Groefsema, M. (1995), 'Understood arguments: A semantic/pragmatic approach',
Lingua 96, 139-61.
Haegeman, L. (1987a), 'The interpretation of inherent objects in English', Australian
Journal of Linguistics, 7, 223-48.
— (1987b), 'Register variation in English'. Journal of English Linguistics, 20, (2), 230-
48.
Halliday, M. A. K. and Hasan, R. (1976), Cohesion in English. London: Longman.
Hudson, R. A. (1984), Word Grammar. Oxford: Blackwell.
— (1990), English Word Grammar. Oxford: Blackwell.
— (1992), 'Raising in syntax, semantics and cognition', in Rocca I. (ed. ), Thematic
Structure: Its Role in Grammar. Berlin: Mouton de Gruyter, pp. 175-98.
— (1994), 'Word Grammar', in Asher, R. E. (ed. ), The Encyclopedia of Language and
Linguistics. Vol. 9. Oxford: Pergamon Press Ltd, pp. 4990-3.
— (1995a), 'Really bare phrase-structure=dependency structure'. Eigo Go ho Bunpoh
Kenkyu (Studies in English Language Usage and English Language Teaching}, 17, 3-17.
- (1995b), HPSG without PS? Ms.
— (1995c), Word Meaning. London: Routledge.
- (1996, October 28), 'Summary: Watch', ( LINGUIST List 7. 1525), Available: http://
linguisdistorg/issues/7/7-1525. html#?CFID=4038808&CFrOKEN=16874386. (Ac-
cessed: 21 April 2005).
— (1998), English Grammar. London: Routledge.
- (2000), '*! amn't'. Language, 76, (2), 297-323.
— (2005), 'An Encyclopedia of English Grammar and Word Grammar', (Word
Grammar), Available: www. phon. ucl. ac. uk/home/dick/wg. htm. (Accessed: 21 April
2005).
Kilby, D. (1984), Descriptive Syntax and the English Verb. London: Groom Helm.
Lambrecht, K. (1996, October 30), 'Re: 7. 1525, Sum: Watch', (LINGUIST List
7. 1534), Available: http://linguisrlist. org/issues/7/7-1534. html#?CFID=4038808&CF-
TOKEN= 16874386 (Accessed: 21 April 2005).
Langacker, R. W. (1990), Concept, Image and Symbol. Berlin: Mouton de Gruyter.
Larjavaara, M. (1996, November 3), 'Disc: Watch', (LINGUIST List 7. 1552),
Available: http: //linguisdist. org/issues/7/7-1552. html#?CFID=4038808&CFTO-
KEN=16874386 (Accessed: 21 April 2005).
Lehrer, A. (1970), 'Verbs and deletable objects'. Lingua, 25, 227-53.
Levin, B. (1993), English Verb Classes and Alternations. Chicago: University of Chicago
Press.
Massam, D. (1987), 'Middle, tough, and recipe context constructions in English'. NELS
18, 315-32.
— (1992), 'Null objects and non-thematic subjects'. Journal of Linguistics, 28, (1), 115-
37.
Massam, D. and Y. Roberge. (1989), 'Recipe context null objects in English'. Linguistic
Inquiry, 20, 134-9.
Quirk, Randolph, Greenbaum, Sidney, Leech, Geoffrey N. and Svartvik, Jan Lars
(1985), A Comprehensive Grammar of the English Language. Harlow: Longman.
Rispoli, M. (1992), 'Discourse and the acquisition of eat'. Journal of Child Language, 19,
581-95.
Rizzi, L. (1986), 'Null objects in Italian and the theory of pro'. Linguistic Inquiry, 17,
501-57.
Roberge, Y. (1991), 'On the recoverability of null objects', in D. Wanner and D. A.
UNDERSTOOD OBJECTS IN ENGLISH AND JAPANESE 65
Kibbee (eds), New Analyses in Romance Linguistics. Amsterdam: John Benjamins, pp.
299-312.
Rosta, A. (1994), 'Dependency and grammatical relations'. UCL Working Papers in
Linguistics, (6), 219-58.
— (1997), 'English Syntax and Word Grammar Theory'. (Unpublished doctoral
dissertation, University of London).
Sperber, D. and Wilson, D. (1995), Relevance (2nd edn) Oxford: Blackwell.
Sugayama, K. (1993), 'A Word-Grammatic account of complements and adjuncts in
Japanese', in A. Crochetiere, J. -C. Boulanger and C. Ouellon (eds), Acte du XVs
Congres International des Linguistes Vol. 2. Sainte-Foy, Quebec: Les Presses de
1'Universite Laval, pp. 373-76.
— (1994), 'Eigo no "missing objects" ni tsuite (Notes on missing objects in English)',
Eigo Goho Bunpoh Kenkyu (Studies in English Language Usage and English Language
Teaching), 1, 91-104.
— (1999), 'Speculations on unsolved problems in Word Grammar'. Kobe City University
Journal, 50, (7), 5-24.
Thomas, A. L. (1979), 'Ellipsis: the interplay of sentence structure and context'. Lingua,
47, 43-68.
Notes
1 Notice here that Word Grammar, in the framework of which the arguments are
developed, assumes that complements include subject as well as object, contrary to
most of the phrase-structure-based theories. For details about the distinction between
complements and adjuncts in Japanese, see Sugayama (1993, 1999).
2 The following symbols for grammatical markers are used in the gloss: TP (Topic),
Sb (Subject), Ob (Object), Q (Question marker). The symbol 0 is used for a zero
pronominal in examples when necessary.
3 For a complete list of some 50 verbs having this feature, together with references, see
Levin (1993: 33). Lehrer (1970) also gives a similar list.
Unspecified Object Alternation in Levin's terms (Levin, 1993: 33), which applies
to eat, embroider, hum, hunt, fish, iron, knead, knit, mend, milk, mow, nurse, pack, paint,
play, plough, polish, read, recite, sew, sculpt, sing, sketch, sow, study, sweep, teach, type,
sketch, vacuum, wash, weave, whittle, write, etc., has the following pattern:
a. Mike ate the cake.
b. Mike ate. (= Mike ate a meal or something one typically eats. )
4 For some recent advances in WG, see Hudson (1992, 1995a, 1995b, 1998, 2000,
2005) and Rosta (1994, 1997).
5 Notice here that C stands for a complement, rather than a complementizer in the
diagram.
6 No phrases or clauses are assumed in WG except for some constructions (e. g.
coordinate structures).
7 Things are not so straightforward. Surprisingly younger children seem to have a
different grammar in which (9) and (10) are grammatical and are actually used.
Rispoli (1992) looked at the acquisition of Direct Object omissibility with eat in
young children's natural production. In terms of GB, eat is one of the many English
verbs with which the internal argument can be saturated. He found that at an earlier
stage of development children frequently omitted a Direct Object with eat when the
understood object referred to something in the discourse context, as in this exchange
between a parent (P) and a child (C):
66 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
KENSEI SUGAYAMA
Abstract
This chapter is an attempt to characterize the be to construction within a Word
Grammar framework. First (section 2), a concise account of previous studies into the
category of be in the construction is followed by a description of be in Word
Grammar (section 3). Section 4, section 5 and section 6, then, present a
morphological, syntactic and semantic discussion of the be to construction,
respectively. Section 7 gives a detailed discussion of the question whether be to is a
lexical unit or not The analysis is theoretically framed in section 8, where it is shown
how Word Grammar offers a syntactico-semantic approach to the construction.
2. Category of Be
Before we go into our analysis, let us have a brief look at what characteristics the
modal be shares with other auxiliary verbs, e. g. can and have.
The table below presents a modal be, a prototypical modal can and a
prototypical perfective auxiliary have in respect of the 30 criteria used in
Huddleston (1980) to characterize auxiliaries and modals.
Table 1 Characteristics of be, can and have
BE CAN HAVE
(modal) (perfect)
1 Non-catenative use -a - -
2 Inversion + + +
POLARITY
3 Negative forms R + +
4 Precedes not + + +
5 Emphatic positive + + +
STRANDING
6 So/neither tags + + +
7 SV order, verb given + + +/R
8 SV order, verb new - + +/R
9 Complement fronting - + +
10 Relativized complement + + +/R
11 Occurrence with DO - - -
12 Contraction + + +
POSITION OF PREVERBS
13 Precedes never + + +
14 Precedes epistemic adverb + + +
15 Precedes subject quantifier + + +
INDEPENDENCE OF CATENATIVE
16 Temporal discreteness + + R
17 Negative complement + + +
18 Relation with subject - - -
TYPE OF COMPLEMENTATION
19 Base-complement = + -
20 fo-complement + -a -
21 -en complement - - +
22 -ing complement - - -
INFLECTIONAL FORMS
23 3rd Singular + - + •
24 -en form - - -
25 -ing form = - R
26 Base-form - - +
27 Past Tense + + +
28 Unreal mood: protasis + + +
29 Unreal mood: apodosis - + -
30 Unreal mood: tentative - + -
NB: R means that the verb has the given property but under restricted conditions.
THE GRAMMAR OF BE TO 69
Clearly, at the outset we can say that be in the current construction shares quite a
lot of features with a typical modal can. In what follows I concentrate on the
extent to which this claim holds.
This analysis implies that BEto may have inherited characteristics of modal
70 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
The idea of 'contradicting' can be spelt out more precisely, but the idea here
should be clear. In a nutshell, the Inheritance Principle says that a fact about
one concept C can be inherited by any instance C' of C unless it is contradicted
by another specific fact about C'.
In sum, WG analyzes be in the be to construction as an instance of modal
verb and be, allowing it to inherit characteristics from both heads in the model-
instance hierarchy in Figure 1.
4. Morphological Aspects
I claim diat be in (1) should be considered as a modal because it shares most of
the properties with a prototypical modal in morphology, syntax and semantics.
This claim is assured in Figure 1 by the fact that BEto has a multiple head, thus
inheriting the features from both heads, i. e. modal verb and be. However, a
further semantic analysis of the construction shows that the sense of the
construction is derived from the sense of the infinitive clause rather than that of
be. Let us start by having a look at the morphological characteristics of be in the
construction and its similarities with other modals.
4. 1 Like modals
Consider the following examples:
Clearly each of the examples in (2) shows that be in this construction cannot be
infinite. The presence of tense is what be shares with modal verbs.
4. 2 Unlike modals
• Be has a distinct j-form controlled by subject-verb agreement
(4) (Her novel is/They were/I am/She wasfWe are} to win the prize.
5. Syntactic Aspects
The second aspect is related to syntax.
5. 1 Like modals
• Only (tensed) auxiliary verbs allow a dependent not to follow them.
• It must share subject with the following verb (i. e. it is a raising. verb)5
The behaviour of be is the same as that of a typical raising verb seem as in (6).
(7) You are to take this back to the library at once ~ This is to be taken back to the
library at once.6
This feature is critical in that two members of the same category of modal verbs
cannot appear consecutively. (8) shows that might and be to belong in the same
category of modal verbs.
72 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
• It is an operator, i. e. has NICE properties [but Code only if the same verb
given in the previous clause as in (11)]
5. 2 Unlike modals
• taking to -infinitive rather than a bare infinitive as a complement of to
(12) That young boy was *(to) become President of the United States.
Although be to has several different meanings, its basic (or core) meaning can be
stated as follows:
Here the key points are the arrangeability of the event described and the
openness of the agent's commitment to the obligation. The first point is most
easily detected when it occurs with the event that cannot be arranged. Consider
the following examples:
All these examples assert the speaker's high certainty at the speech time of the
event happening in the (near) future.
74 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Related to this usage type is 'future in the past', a case where the speech time
is transferred to some point in the past
(40) And the free world has reacted quickly to this momentous process and must
continue to do so if it is to help and influence events [ICE-GB: S1B-054 #17: 1: B]
(41) the system is totally dependent on employee goodwill if it is to produce good
information. [ICE-GB:W2A-016 #118: 1]
(42) However, in nerves regeneration is essential if there is to be a satisfactory
functional outcome. [ICE-GB: W2A-026 #15: 1]
In the end, there are of course differences between be to and modal verbs.
Still be in (1) shares enough properties with modal verbs to be categorized as a
modal, and although it may be categorized as a modal, the sense of the be to
construction is best considered to be an existence of a situation where the event
is represented by the VP in the infinitive. Modal-like meanings of the
construction are derived from the sense of to rather than that of be, which is
attested in the following section.
• Inversion
In a yes-no question, what moves to the front is not be to but be, which suggests
that be behaves like a modal (operator), with to being an infinitive marker.
(46) He should go
Should he go?
(47) He ought to go.
* Ought to he go?
Ought he to go?
(48) He is to go.
*Is to he go?
Is he to go?
• Unlike ought to and have to, the to doesn't have to be retained in be to when a
VP that follows to is deleted.
• Unlike ought to and have to, there is no phonetic fusion with be to.
Though it is not fully clear whether examples in (56) are zeugmatic or not, the
/o-infinitive may be coordinated with a wide range of conjuncts of different
categories (Warner 1993). My informants however say that they are all right
without a zeugmatic reading. If this is the case, the to-infinitive is an
independent unit in be to and there has to be a syntactic gap between be and to.
THE GRAMMAR OF BE TO 77
(56) He was new to the school and to be debagged the following day.
The old man is an idiot and to be pitied.
You are under military discipline and to take orders only from me.
You are a mere private and not to enter this mess without permission.
He was an intelligent boy and soon to eclipse his fellows.
If this is the case and there is a one-to-one relation between syntax and
semantics as is maintained in WG, the fo-mfmitive is semantically as well as
functionally an independent element in the be-to construction and therefore the
be-to is obviously not a syntactic unit, although there has to be some
grammatical relation between the two elements (i. e. be and fo-infmitive). It
seems that what fo-infinitives in (56) have in common is the function of
predication. Otherwise they cannot be coordinated with those conjuncts before
and in (56). This implies that be is an ordinary predicative be, followed by an
infinitival clause with to of predicative function.
Sentences in (57) and (58) give evidence supporting this predicative function
of the infinitive clause because there is no be found in examples in (57).
Nevertheless, the NPs in (57) and (58) all express predication, although they
are NPs as a whole in syntactic terms:
This predicative analysis of the to-infinitive is also supported by (59a), where the to-
infinitive is a part of a small clause with the preposition with. It is a well-attested fact that
with heads a small clause expressing predication as in (59b). Therefore, the to-infinitive
functions as a predicative in (59a).
Each arrow in this diagram shows a dependency between words and it points
from a head to one of its dependents, but what is most important here is that
there are no phrases or clauses in the sense of constituency grammars. Thus in
terms of dependency, are is the root of the sentence (represented as a bold
arrow) on which you and reading depend as a subject and sharer (a kind of
complement),8 respectively. In turn, a is the head of paper. Grammar depends
on paper, Word depends on Grammar, and so on. Turning to the semantics of
this sentence, 'you', which is a referent of you, is linked as a read-er (i. e. agent)
to a semantic concept 'You read a Word Grammar', an instance of 'read'.9
The curved (vertical) lines point to the referent of a word. '-Er' and '-ee' are
names of semantic relation or thematic roles. A small triangle representing the
THE GRAMMAR OF BE TO 79
(60) b.
Here 'be' is an instance of non-epistemic, because (60b) means that some event
is arranged or planned, without any sense of the speaker's judgement on the
proposition embedded in the utterance (sentence). In both cases, the meaning
(i. e. sense) of 'be' needs a proposition which expresses an activity or event as a
dependent, which is a sense of the verb.
Abstracting away from the technical markers, the diagram in (60c) represents
a WG analysis of the coordinate structure in (56). This diagram schematizes the
very idea that the same predicative function (a dependency relation) holds
between was and the first conjunct new to the school enclosed by square brackets
on the one hand, and between was and the second conjunct to be debagged.... at
the same time.
(60) c.
THE GRAMMAR OF BE TO 81
9. Conclusion
In this chapter, I have shown that a morphological, syntactic and semantic
analysis of be in the be to construction provides evidence for the category of be in
this construction proposed here. Namely, be is an instance of a modal verb in
terms of morphology and syntax, while the sense of the whole construction is
determined by the sense of 'to'. The analysis also explains why be to does not
constitute a lexical unit. Finally, the WG account presented here gives the same
syntactic and semantic structures to the construction, reducing the complexity of
the mapping between the two levels of the structure of the modal-like
expression be to.
References
Aarts, Bas (1992), Small Clauses in English. Berlin: Mouton de Gruyter.
Bybee, John (1994), The Evolution of Grammar. Chicago: The University of Chicago
Press.
Celce-Murcia, Marianne and Larsen-Freeman, Diane (1999 ), The Grammar Book.
Boston, MA: Heinle & Heinle.
Collins Cobuild English Language Dictionary for Advanced Learners. (1995 ), Glasgow:
Harper Collins Publishers.
Collins Cobuild English Language Dictionary for Advanced Learners. (2001 ), Glasgow:
Harper Collins Publishers.
Collins Cobuild Advanced Learner's English Dictionary. (2003 ), Glasgow: Harper Collins
Publishers.
Gazdar, Gerald, Pullum, Geoffrey K. and Sag, Ivan A. (1982), 'Auxiliaries and related
phenomena in a restrictive theory of grammar'. Language, 58, 591-638.
Greenbaum, Sidney, Leech, Geoffrey N. and Svartvik, Jan Lars (eds), (1980), Studies in
English Linguistics for Randolph Quirk. London: Longman.
Huddleston, Rodney (1976), 'Some theoretical issues in the description of the English
verb'. Lingua, 40, 331-383.
— (1980), 'Criteria for auxiliaries and modals', in Greenbaum, Sidney, et al. (eds),
Studies in English Linguistics for Randolph Quirk. London: Longman, pp. 65-78.
Hudson, Richard A. (1984), Word Grammar. Oxford: Blackwell.
— (1990), English Word Grammar. Oxford: Blackwell.
— (1996), A Word Grammar Encyclopedia (Version of 7 October 1996). University
College London.
— (2005, February 17 - last update), 'An Encyclopedia of English Grammar and Word
Grammar', (Word Grammar), Available: www. phon. ucl. ac. uk/home/dicVwg. htm.
(Accessed: 21 April 2005).
Kreider, Charles W. (1998), Introducing English Semantics. London: Roudedge.
Lampert, Gunther and Lampert, Martina (2000), The Conceptual Structure (s) of
Modality. Frankfurt am Main: Peter Lang.
Lyons, John (1977), Semantics. Cambridge: Cambridge University Press.
McCawley, James D. (1988),
The Syntactic Phenomena of English. Chicago: University of Chicago Press.
Napoli, Donna Jo (1989), Predication Theory. Cambridge: Cambridge University Press.
Palmer, Frank Robert (19902), Modality and the English Modals. Harlow: Longman.
— (2001 ), Mood and Modality. Cambridge: Cambridge University Press.
Perkins, Michael. R. (1983), Modal Expressions in English. London: Frances Pinter.
82 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Notes
1 This is a revised and expanded version of my paper of the same title read at the
International Conference 'Modality in Contemporary English' held in Verona, Italy
on 6-8 September 2001. 1 am most grateful to the comments from the audience at
the conference. Remaining errors are however entirely my own. The analysis
reported here was partially supported by grants from the Daiwa Anglo-Japanese
Foundation (Ref: 02/2030). Their support is gratefully acknowledged.
2 Here we shall not take into account similar sentences as in (i), which are considered
to have a different grammatical structure from the one we are concerned with in this
chapter.
(i) My dream is to visit Florence before I die.
3 The idea of construction here is quite a naive one, different from the technical
definition of the one used in Goldberg's Construction Grammar.
4 Palmer (1990: 164), among others, claims that 'is to' is formally a modal verb.
5 Be to can be used in the there construction as in (i).
(i) Regional accents are still acceptable but there is to be a blitz on incorrect
grammar. [COBUILD2]
This suggests that be is a raising verb because there is no semantic relation between
there and 'be to'.
Example (i) is construed as a counter-example to view this construction as one
involving a subject-control as in (ii).
(ii) Mary is [PRO to leave by 5]. - Napoli
6 The possibility sense is found only in the passive, so there is no active counterpart
for These insects are to be found in NSW. [Huddleston 1980: 66; Seppanen]
7 Quite exceptionally, it could be arranged by God or other supernatural beings.
Otherwise it cannot be.
8 Sharer is a grammatical relation in Word Grammar.
9 Here we disregard the tense and aspect of the sentence.
5 Linking in Word Grammar
JASPER HOLMES
Abstract
In this chapter I shall develop an account for the linking of syntactic and semantic
arguments in the Word Grammar (WG) framework. The WG account is shown
to have some of the properties of role-based approaches and some of the
properties of class-based approaches.
1. 1 Introduction
Any description of linguistic semantics must be able to account for the way in
which words and their meanings combine in sentences. Clearly, this
presupposes an account of the regular relationships between syntactic and
semantic structures: a description of the mechanisms involved in linking.
The search for an adequate account of linking has two further motivations: it
makes it possible to explain the syntactic argument-taking properties of words
(and therefore obviates the need for valency lists or other stipulative
representations of subcategorization facts); and it provides a framework for
dealing with words whose argument-taking properties vary regularly with the
word's meaning (many such cases are treated below and in the work of other
writers in the field of lexical semantics including Copestake and Briscoe 1996;
Croft 1990; Goldberg 1995; Lemmens 1998; Levin 1993; Levin and Rappaport
Hovav 1995; Pustejovsky 1995; and Pustejovsky and Boguraev 1996).
Levin and Rappaport Hovav provide yet another reason to seek an account
of argument linking: that it is an intrinsic part of the structure of language. In
their introduction, they make the following claim:
To the extent that the semantic role of an argument is determined by the meaning of
the verb selecting it, the existence of linking regularities supports the idea that verb
meaning is a factor in detennining the syntactic structure of sentences. The striking
similarities in the linking regularities across languages suggest that
they are part of the architecture of language. (1995: 1, my emphasis)
Of course, it is not the meanings of verbs alone that are relevant in determining
semantic structure. It should also be clear that I do not share Levin and
84 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
1. 2 Indirect objects
Figure 2 shows some of the syntactic and semantic structure that needs to be
associated lexically with GIVE: the verb has a subject (s in the diagram), an
object (o in the diagram) and an indirect object (io in the diagram), all of which
are nouns. The sense of the verb, Giving, is an event with an 'er' (the referent of
the subject; the properties of 'ers' and 'ees' are discussed shortly), an 'ee' (the
referent of the object), a recipient (the referent of the indirect object) and a result,
LINKING IN WORD GRAMMAR 85
an example of Having which shares its arguments with its parent The giver has
agentive control over the event, being in possession of the givee beforehand and
willing the transfer of possession. The givee and the recipient are more passive
participants: the former undergoes a change of possession, but nothing else, the
latter simply takes possession of the givee. Being a haver presupposes other
properties (centrally humanity), but those are not shown here.
Figure 2 GIVE
The set of verbs that can take an indirect object, of either kind, is in
principle unlimited in size, since it is possible to extend it in one of two ways.
First, membership is open to new verbs which refer to appropriate activities:
Second, and even more tellingly, existing verbs can be used with indirect
objects, with novel meanings contributed by the semantics of the indirect object
(7) The colonel waggled her his bid with his ears.
(8) Dust me the chops with flour.
Examples like (7) and (8) are acceptable to the extent that the actions they
profile can be construed as having the appropriate semantic properties. For
example a bottle of beer can be construed as having been prepared for
someone if it has been opened for them to drink from, but a door is not
construed as prepared when it has been opened for someone to pass through:
UNKING IN WORD GRAMMAR 87
It is clear from this, and from the fact, noted by Levin (1993: 4-5) with
respect to the middle construction, that speakers make robust judgements
about the meanings of unfamiliar verbs in constructions on the basis of the
construction's meaning (see also (10)), that the meaning of the construction
must be represented in a schematic form in the mind of the language user.
Some syntactic properties of the indirect object (in English) are given by
Hudson (1992). These include the possibility of merger with subject in passive
constructions (13), the obligatoriness of direct objects in indirect object
constructions (14) and its position immediately following the verb (15).
The semantic property common to all indirect objects is that they refer to
havers: in the case of the verbs taking 'dative' indirect objects in (2), the result of
the verb's sense is that the referent of the indirect object comes into possession
of something; in the case of those taking 'benefactive' indirect objects in (3), the
verb profiles an act of creating or preparing something intended to be given to
the referent of the indirect object.
Figure 3 shows the various properties associated with indirect objects. First,
the diagram shows that indirect objects are nouns, and that it is verbs, and
more particularly verbs with objects, that have indirect objects: Ditransitive,
the category of verbs with indirect objects, isa Transitive, the category of verbs
with direct objects. This is enough by itself to represent the fact that the direct
object is obligatory with indirect objects (14), but the object relationship is
nevertheless also shown in the ditransitive structure, since it appears in the
word order rule (indirect objects precede objects). The referent of the object
also appears in the semantic structure, along with that of the indirect object,
since without it the semantic structure cannot be interpreted. I show the two
referents as coarguments of the result of the verb's sense, though the
semantics is worked out more clearly in the discussion of Figure 4. The fact
that indirect objects can merge with subjects in passive constructions is dealt
with in the following section.
Indirect objects may have one of two slightly different semantic structures,
each associated with a separate category of ditransitive verbs. In both, the
referents of the two dependents are 'er' and 'ee' of a Having, but the role of that
Having differs somewhat between the two. The two structures are given in
Figure 4.
UNKING IN WORD GRAMMAR 89
The sense of the verb isa Making, and its result (therefore) isa Being (is a state)
and the argument of that Being is the referent of the direct object: baking a cake
results in that cake's existence, baking a potato results in that potato's being
ready. The Having that connects the referents of the two arguments is the
purpose of the verb's sense: the purpose of the baking of the cake is that it
should belong to her (the referent of the indirect object). This concept is
connected to the sense of the verb by the beneficiary relationship (labelled ben/
fy). Ditransitive/2 (17) has as its sense a Giving event, which straightforwardly
has as its result the Having that connects the referents of the two arguments.
The referent of the indirect object is connected to the sense of the verb by the
recipient relationship.
Once these two semantic structures are established, they can be used in the
treatment of the relationship between the indirect object construction and
constructions with the prepositions TO and FOR. Simply, TO has the same sense
as Ditransitive/2 and FOR the same as Ditransitive/1 (with some differences: see
(18)). This synonymy can, though it need not, be treated as a chance occurrence:
no explanation is necessary for the relationship between constructions with
indirect objects and those with TO. The case of FOR and the difference seen in
(18) certainly supports the idea that the two constructions converge on a single
WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
meaning by chance, since the two meanings are in fact different The use of the
indirect object to refer to the beneficiary of an act of preparation is only possible
where the prepared item is prepared so it can be owned (or consumed) by the
beneficiary; this constraint does not apply to beneficiary FOR.
those verbs that can conform to the properties of the construction can appear in
it: *Skate me a half-pipe/*Run me a mile, etc.
1. 3 Objects
Biber et al. (1999: 126-8) give a number of syntactic properties for English
objects, as follows (the properties assigned to objects and to subjects are all
taken from Biber et al. ', some details may be disputed, but the general point
remains the same):
• The first two syntactic properties refer to the classes of the words at either
end of the object relationship: some verbs (the transitive verbs) lexically
select an object; the objects themselves are generally nouns.
• The third property concerns the form of the object: when it is a pronoun, it
takes the 'accusative' form (what I have above called the default form).
• The fourth property concerns its relative position in the sentence: objects
generally follow their parents, and only a limited set of other dependents of
the parent may intervene (any number of predependents of the object may
intervene) (20). Biber et al. note that indirect objects may come between the
object and the parent (21); this possibility is also open to particles (22).
The parent in an object relationship isa Verb and the dependent isa Noun. In
this way, the object relationship defines a class of transitive verbs (verbs that
have objects). Verbs that only appear in transitive constructions inherit all
properties from this class (DEVOUR isa Transitive, just as ACCORD isa
Ditransitive). The word order properties are represented by the next
relationship: the form of the dependent is the next of that of the parent.
The diagram also shows the category Ditransitive (see Figure 4), where the
word order properties are somewhat different (the form of the object is the next
of that of the indirect object). Also represented in the diagram is the class of
passive verbs (the category Passive). These verbs are defined by their formal
properties: the form of a passive verb consists of its base plus a suitable ending
(not shown). There are two classes of passive verb: one, which also isa
Transitive, in which the subject is merged with the object, and one, which also
isa Ditransitive, in which the subject is merged with the indirect object.
The full lexical structure of the object relationship must also include its
semantic properties. In line with the approach outlined above for indirect
94 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
objects, the semantic properties of the object are related to its syntax through a
specialization of the SSP. Biber et al. (1999: 126-8) also identify a range of
possible semantic relationships that correspond with the object relationship (see
a-g), and the lexical semantic representation of the object relationship should be
general over all of these.
determined in the semantics by the nature of the resulting state: where this state
isa Being, its 'er' is the 'ee' of the verb's sense; where it isa Having, its 'er' is the
recipient of the verb's sense, rather than its 'ee', and the 'ee' of the verb's sense
is the same as the 'ee' of the result (see Figure 4 above).
'LocationaT objects, as in c, do not refer to affected arguments, but to parts
of a path. The example in c defines the beginning and end of the path (on
opposite sides of the river), but other examples may profile the beginning (25a),
middle (25b) or end (25c) of the path.
(25) a. The express jumped the rails, (from Biber et al. (1999: 127))
b. nny vaulted the horse.
c. Elly entered the room.
The set of verbs that can appear with an object of this kind is (naturally) limited
to those that can refer to a motion event and in this sense the 'locative' object is
lexically selected by its parent. Notice also that the verb (often) determines
which part or parts of the path may be profiled by such an object. Because of
this, these arguments must appear in the lexical structures of quite specific
categories (at the level of the lexeme or just above). The relevant categories are
subsumed under Transitive, since the syntactic properties are the same as those
of the affected/effected objects, but it is arguable whether they need to be
collected under a category 'locative object verb'. This category is justified to the
extent that generalizations can be made over the relevant constructions.
There seems to be little semantically in common between locative objects
and affected/effected objects, though there is some relationship. For example,
Dowry's (1991) incremental theme is a property of both kinds of object the
event in both cases is bounded by the theme:
When the sense of the verb is an unbounded event, a measure expression can
be used to define a bounded path: Sammy swam jive miles. It is not entirely clear
that arguments of this sort are indeed objects. Some certainly are not; in 27 the
object of pushed is the (pea) and not the measure expression.
(27) Evans pushed the pea five miles with his nose.
The 'measure' objects are also confined to a limited class of verbs, by which
they are semantically selected (weigh Jive tons, measure Jive furlongs). They also
have little in common semantically with the other types of object, since their
semantics is so heavily constrained by the verb.
'Cognate' objects (28, 29) are also associated with a very small class of verbs
(Levin gives 47 (1993: 95-6), out of a total of 3107 verbs). They have
something in common semantically with effected objects, but the semantics is
constrained by the verb, which may also go so far as to select a particular
lexeme. Levin notes:
96 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Most verbs that take cognate objects do not take a wide range of objects. Often they
only permit a cognate object, although some verbs will take as object anything that is
a hyponym of a cognate object (1993: 96).
The verb and its object refer jointly to a performance of some kind.
'Eventive' objects are confined to an even smaller class, the 'light' verbs. In
these cases the event structure is determined by the verb, but the details of the
semantics are supplied by the noun. In light verb constructions with HAVE, the
object refers to an event (have a bath/meal/billiards match); light DO, in contrast,
refers to an affective/effective event, the precise nature of which is determined
by the semantics of the (affected) object:
by way of one of the subclasses) and nouns that are objects inherit some of their
properties from the category that fills the relevant slot in the structure (perhaps also
by way of one of its subclasses: the diagram does not show inheritance relationships
between the object noun in the most general case and those in the subcases, but
these relationships are nevertheless implicit in the inheritance structure).
1. 4 Subjects
Biber et al. (1999: 123-5) give the following syntactic properties for English
subjects:
• Again, the first two syntactic properties concern the classes of words that
participate in the relationship: verbs have subjects, which are generally
nouns. Any verb may have a subject, so the class of 'subject verbs' is less
constrained than the class of transitive verbs. It is perhaps for this reason
that the semantic roles played by subjects are so much more diverse (see
below). All tensed verbs have subjects, so the class Tensed is shown as a
subset of the subject verbs. (See Figure 10. )
• The 'nominative' form of personal pronouns consists of the five words I,
SHE, HE, WE and THEY which are subcases of the relevant pronouns
that are used only in subject position. (See Figure 10. )
98 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
• The word order properties of subjects are slightly more complicated. Generally
the subject precedes its parent, but some subjects follow their parents and in
many of these cases the referent of the verb is questioned (the construction
forms a yes/no question); these cases are represented in the subclass of subject
verbs Inverted. The word order properties of Wh questions are determined in
part by the lexical properties of the category Wh (schematic over Wh words).
This category is always the extractee (x< in the diagram) of its parent and so
precedes it. Where the Wh word is not the subject of the verb, the verb and
subject are also inverted (the complement of Wh isa Inverted).
• The final syntactic property is more properly semantic in WG: just as there
is overlap between the semantics of the indirect object relationship and that
of the preposition TO, so there is considerable overlap between the
semantics of the subject relationship and that of the preposition BY.
The semantic properties of subjects are explored more fully in the following
section, but some general remarks can be made here. Biber et al. (1999: 123-5)
give the following possible semantic roles for subjects:
• The first three roles (a-c) can be collected together by virtue of the force-
dynamic properties they share: agents, causes and instruments all precede
the event in the force-dynamic chain.
• I argue below that affected subjects (e) are similarly controlled by the force-
dynamic structures of the verbs that take them.
• The semantic roles played by the subjects of stative verbs are chiefly
determined by the lexical (semantic) structure of the individual lexeme,
though some semantic classification is possible (see Figure 13).
• 'Local' and 'eventive' subjects are controlled by the lexical structures of the
verbs that take them.
• Since every verb can have a subject, the number of different semantic roles
open to the referents of subjects is limited only by the number of different
event types denoted by verbs. This can be seen particularly clearly in the
case of 'dummy' subjects (h).
Figure 13 collects together the possible semantic roles associated with the
subject relationship, and relates them symbolically to the syntactic properties
identified above (given schematically in the diagram). The various semantic
types of subject are glossed by the 'er' relationship introduced above. A full
account of this relationship and of the 'ee' relationship linked with objects is
provided in the following section. Four kinds of stative predicate are shown,
covering the three possibilities under (d) and the 'local' subjects in (f). Some of
these semantic classes are dealt with in more detail in following chapters; each
makes different requirements of its 'er'. A class of 'eventive verbs' is also
included; these corefer with their subjects.
A linking rule for objects is given in Figure 15 and in (32). The rule pairs the
syntactic relationship object with the semantic relationship 'ee' (this is the
pattern given above for DO/light, which is followed by most transitive verbs).
The former gathers together the syntactic properties of objects (Figure 7) and
the latter the semantic properties associated with them (Figure 8).
102 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
argument is the recipient of C. The relationships 'er' and 'ee', however, which
are linked to subject and object respectively, are less straightforward, so the
status of the subject and object linking rules is at least open to question. In the
second part of this chapter, I address the outstanding issues, providing more
detailed linking rules for subjects and objects.
2. 1 The framework
In the first part of this chapter I sketched a linking mechanism within the WG
framework, based on generalizations over grammatical relations (specializations
of the Syntax Semantics Principle). The details are fleshed out in this part.
The linking regularities presented above consist of symbolic structures which
link specific syntactic relationships (subject, object, indirect object, etc. ) with
specific semantic relationships ('er', 'ee', recipient, etc. ). The syntactic
relationships are identified by a set of word-level (syntactic, morphological,
phonological, etc. ) properties which, by default, are inherited by all cases of the
dependency: unless otherwise specified, subjects precede their parents and
determine their form, objects follow and permit no intervening codependents,
and so on. The semantic relationships are identified by a set of concept-level
(thematic, force-dynamic, etc. ) properties, which likewise constitute the default
model for the relationship. The syntactic and semantic properties taken
together constitute the lexical structure of the relevant relationship, and can be
seen as a gestalt.
As I argue above, semantic relationships like recipient and result are quite
straightforwardly understood in terms of more complex semantic structures.
The relationships 'er' and 'ee', however, which are linked to subject and object
respectively, are less straightforward. An account is provided here in which the
properties of 'ers' and 'ees' are defined by a hierarchy of event types (notice that
an event type (Having) played a role in the definition of result and recipient).
Since most of the event types are defined by a single exceptional argument
relationship, and since the linking regularities are still stated in terms of single
roles, the WG approach outlined here combines the properties of role-based
and class-based approaches. The linking regularities presented above are
generalizations over the linking properties of all subjects, objects, etc. While
each syntactic dependency always maps onto the same semantic argument, the
exact nature of the role played by that argument is determined by the wider
conceptual structure associated with the parent's sense (as represented partly by
its event type). The distinction between words and constructions is an emergent
property of the network structure.
The categories in the event type hierarchy are defined by their semantic
(conceptual) properties, including force-dynamic properties (but not including
aspectual properties: see Holmes (2005: 176-211)). Many of the event types
function as the senses of words, though some do not. The categories support a
104 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
2. 2 Event types
Figure 17 shows the event type hierarchy. The various types are shown, but
most of their properties are not (they are given in the following diagrams). The
category at the top of the hierarchy is labelled Predicate; this is not an entirely
satisfactory name for this concept, but it has the benefit of subsuming both
states and events. The names of the concepts in the hierarchy are intended to
be the senses of lexical words, and for this reason it is perhaps surprising that no
readily useable term exists for the highest category, though it might be argued
that this concept does not have much use as an element in the normal use of
language. The event type hierarchy should more properly be called the
predicate type hierarchy.
Predicates are divided into states (State) and events (Event), the latter
consisting of a series of (more or less transient) states. The most general
category, Predicate, is shown with a single argument, labelled 'er', and this
association is inherited (implicitly) by the two subclasses. The states are divided
into Being and Having; the latter and some of the former have a second
argument, labelled 'ee'. Further properties of these categories are explored
shortly. The events include processes like Laughing and Yawning as well as
the further categories Becoming and Affecting. The first of these is telic (it has a
result which is a state); the second has an 'ee' as well as an 'er'. Affecting
LINKING IN WORD GRAMMAR 105
Figure 19 Feeling
same as the place of a part of the 'ee'. In the case of In, this part is the interior;
in the case of On, it is the surface. The diagram also shows that Containing and
Supporting are the converses of In and On respectively (if a is in b then b
contains a; if a is on b then b supports a). These facts are integral parts of the
meanings of the prepositions.
Figure 20 At
LINKING IN WORD GRAMMAR 107
Figure 21 Having
Figure 23 Events
LINKING IN WORD GRAMMAR 109
independent action in the clause. ' (1993: 6), noting that this is a simple
extension of the category agent to fit other kinds of subject-linked arguments.
This extension covers verbs referring to changes undergone by their single
argument (unaccusative verbs), whose arguments therefore may have few or no
agentive properties (note however, that some are agents (30c). In fact, the actors
of other one- or two-argument events are also not agents ((28a), (28c), (29b)).
Agency is a property of some actors, determined by the thematic properties
of the event, so the thematic role agent ('the semantic role borne by an NP
which is perceived as the conscious instigator of an action' ibid.: 11) is not called
for. Actor, then, corresponds roughly to Dowty's (1991) proto-agent: it is
defined by properties like volitional involvement, causal instigation etc., but not
all cases share all these properties. Dowty's proto-agent wills the event, is
sentient, causes an event or change of state, moves and has independent
existence; the WG treatment presented here accepts all of these but the fourth,
movement.
Patient ('the semantic role borne by an NP which expresses the entity
undergoing an action' Trask (1993: 202)) is schematic over the second
argument of transitive events. Affecting, which is the most general such event,
subsumes processes (like pushing a pea or patting a dog) and causative events
(like pushing a pea to Peterborough or angering a tourist). The patient is the
affected (or effected) argument, even in some of the transitive processes.
Processes have a temporal profile that consists of a set of repeated events.
These events may themselves be causative (Pushing consists of a set of repeated
causative actions on an object), though they may also be states (Patting consists
of a set of repeated locative states) in which case the patient is the theme of the
state (see below).
Dowty's (1991) proto-patient undergoes a change of state, is an incremental
theme, is causally affected by another participant, does not move and does not
have independent existence. Again, the WG analysis accepts all these but the
fourth, concerning movement The incremental theme is a product of the
aspectual structure of affective events (see Holmes 2005).
States have themes, and some have actors. Actors of states share the
properties of those of non-states. The theme is the argument that the state is
predicated of (theme is also used with similar meaning as the name of a
discourse function, where it contrasts with rheme, as topic does with comment).
Trask gives 'an entity which is in a state or a location or which is undergoing
motion' (1993: 278), a definition which subsumes some patients, as defined
above; Trask also notes that the terms theme and patient are used more or less
interchangeably. However, in the current framework the two are separate:
patients undergo some affective/effective process or change; themes have some
stable property. Locative states also have a landmark: the argument whose
position defines that of the theme.
The above definitions of the thematic roles are given in terms of semantic
properties. For example, an actor wills the event, is sentient, causes an event or
change of state and has independent existence. These semantic properties of
actor are shown in Figure 28.
LINKING IN WORD GRAMMAR 113
Figure 28 Actor
clear there that both words and constructions may select both argument and
participant roles.
Since the participant roles are defined in terms of sets of default properties,
it is possible for more than one argument of a verb's sense to fit the bill for one
or other participant role. This is the case for the verbs SPRAY and LOAD. As
is well known, these two verbs can be used with objects referring to a thing or
substance moved or to the place it is moved to. These two possibilities reflect
two ways of interpreting the roles of the participants (of choosing which
participant best fits the patient model, and is therefore linked to 'ee' and thence
to object).
In these cases the lexical properties of the syntactic relationship (here object)
can be added to those of the verb. Where the two are not in conflict, they are
simply merged. For example, since LOAD does not select either of its non-
subject arguments as an incremental theme, this property is assigned to the
object-linked argument by the semantics of the 'ee' relationship (40) (the
mechanics of this example are discussed in Holmes 2005: 206ff).
When there is a conflict between the lexical properties of the construction and
those of the verb, the construct is (usually) rendered incoherent. The two
examples in (41) are unacceptable because the lexical structure of POUR
specifies that the 'ee' of its sense is a liquid (that is how the manner of pouring is
defined) and that of COVER specifies that the 'ee' of the sense ends up
underneath something. These two requirements clash with the semantics of the
construction.
3. Conclusion
In the first part of this chapter I sketched the linking mechanisms of WG.
Syntactic and semantic associative relationships participate in symbolic
relationships: syntactic dependencies have meanings, which serve to determine
the interpretations of compositional structures, as well as to constrain the
possibilities for composition. Just as the (default) properties of syntactic
associations are given in terms of a network of related concepts and properties
surrounding the dependency class they instantiate, so are the (default)
properties of semantic associations.
In the second part I distinguished two kinds of semantic association:
participant roles, which carry thematic content; and argument roles, which are
determined by the force-dynamic properties of the event class.
LINKING IN WORD GRAMMAR 115
References
Biber, Douglas, Johansson, Stig, Leech, Geoffrey, Conrad, Susan and Finegan, Edward
(1999), Longman Grammar of Spoken and Written English. Harlow, Essex: Longman.
Bresnan, Joan W. (1982), The Mental Representation of Grammatical Relations.
Cambridge, MA: MIT Press.
Chomsky, Noam (1970), 'Remarks on nominalization', in Roderick A. Jacobs and
Peter S. Rosenbaum (eds), Readings in English Transformational Grammar. Waltham
MA: Ginn and Company, pp. 184-221.
— (1981), Lectures on Government and Binding. Dordrecht Foris.
Copestake, Ann and Briscoe, Ted (1996), 'Semi-productive polysemy and sense
extension', in James Pustejovsky and Branimir Boguraev, Lexical Semantics: the
Problem of Polysemy. Oxford: Clarendon Press, pp. 15-68.
Croft, William (1990), 'Possible verbs and the structure of events', in Savas L.
Tsohatzidis (ed. ), Meanings and Prototypes: Studies in Linguistic Categorization.
London: Routledge, pp. 48-73.
Cruse, David A. (1986), Lexical Semantics. Cambridge: Cambridge University Press.
Davis, Anthony R. and Koenig, Jean-Pierre (2000), 'Linking as constraints on word
classes in a hierarchical lexicon'. Language, 76, 56-91.
Dowry, David R. (1991), 'Thematic proto-roles and argument selection'. Language, 67,
547-619.
Goldberg, Adele E. (1995), Constructions: a Construction Grammar Approach to Argument
Structure. Chicago: University of Chicago Press.
— (2002), 'Surface generalizations'. Cognitive Linguistics, 13, 327-56.
Holmes, Jasper W. (2005), 'Lexical Properties of English Verbs' (Unpublished doctoral
dissertation, University of London).
Hudson, Richard A. (1984), Word Grammar. Oxford: Blackwell.
— (1990), English Word Grammar. Oxford: Blackwell.
— (1992), 'So-called "double objects" and grammatical relations'. Language, 68, 251-76.
— (1994), 'Word Grammar', in Ronald Asher (ed. ), The Encyclopedia of Language and
Linguistics. Oxford: Pergamon Press, pp. 4990-93.
— (1999), 'Subject-verb agreement in English'. English Language and Linguistics, 3, 173-207.
— (2004, July 1-last update), 'Word Grammar', (Word Grammar], Available:
www. phon. ucl. ac. uk/home/dicVwg. htm (Accessed: 18 April 2005).
Jackendoff, Ray S. (1990), Semantic Structures. Cambridge, MA: MIT Press.
Lemmens, Maarten (1998), Lexical Perspectives on Transitivity and Ergativity: Causative
Constructions in English. Amsterdam: J. Benjamins.
Levin, Beth (1993), English Verb Classes and Alternations: a Preliminary Investigation.
Chicago: University of Chicago Press.
Levin, Beth and Rappaport Hovav, Malka (1995), Unaccusativity: at the Syntax-Lexical
Semantics Interface. Cambridge, MA: MIT Press.
Pustejovsky, James (1995), The Generative Lexicon. Cambridge, MA: MIT Press.
— (2001), 'Type construction and the logic of concepts', in Pierette Bouillon and
Federica Busa (eds), The Language of Word Meaning. Cambridge: Cambridge
University Press, pp. 91-123.
Pustejovsky, James and Branimir Boguraev (1996), 'Introduction: lexical semantics in
context', in James Pustejovsky and Branimir Boguraev, Lexical Semantics: the
Problem of Polysemy. Oxford: Clarendon Press, pp. 1-14.
Shibatani, Masayoshi (1996), 'Applicatives and benefactives: a cognitive account', in
Masayoshi Shibatani and Sandra A. Thompson (eds), Grammatical Functions: their
Form and Meaning. Oxford: Clarendon Press, pp. 157-94.
116 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Notes
1 The SSP given here turns out not to be able to account for all cases of linking. It is
revised in Holmes (2005: 44).
2 Note that in Biber et al. the VP category subsumes the 'verbal complex' (main verb
and any auxiliaries), but not any complements or other postdependents of the verb.
3 Of course, if a particular speaker knows the words ACTOR and THEME (as
metalinguistic terms), then they must have these relationships in their lexicon, since
they are (or should be!) the meanings of the relevant terms.
6 Word Grammar and Syntactic Code-Mixing
Research
EVA EPPLER
Abstract
This chapter aims to show that WG is preferential over other linguistic theories
for the study of bilingual speech. Constituent-based models have difficulties
accounting for intra-sentential code-mixing because the notions of government
and functional categories are too powerful and rule out naturally occurring
examples. Properties of WG which make this syntactic theory particularly well
suited for code-mixing research are the central role of the word, the dependency
analysis, and several consequences of the view of language as a network which is
integrated with the rest of cognition. A qualitative and quantitative analysis of
because and weil clauses shows that code-mixing patterns can be studied
productively in WG.
1. Introduction
Intra-sententially CODE-MIXED data, i. e. utterances constructed from words
from more than one language, pose an interesting problem for syntactic
research as two grammars interact in one utterance. Based on a German/
English bilingual corpus,11 will show in section 2 of this chapter that constraints
on code-switching formulated within Phrase Structure Grammar frameworks
(Government and Binding, Principles and Parameters, Minimalism) are too
restrictive in that they rule out naturally occurring examples of mixing.
In section 3 I will discuss aspects of WG that make it particularly well suited
for the syntactic analysis of intra-sententially mixed data. WG facilitates the full
syntactic analysis of sizeable corpora and allows us to formulate hypotheses on
code-switching which can subsequently be tested on data. All findings are
supported by quantitative data.
As the word order contrast between German and English is most marked in
subordinate clauses, I focus on examples of this construction type in section 4. 1
will show that code-mixing patterns can be studied productively in terms of
WG: WG rules determining the word order in German/English mixed clauses
hold in relation to my corpus and are supported by evidence from other
corpora. The main section of this chapter focuses on because and weil clauses. A
comparison of the mixed and monolingual clauses reveals that German/English
bilinguals who engage in code-mixing recognize and utilize structural
118 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
(3) *MEL: ich hab(e) gedacht there is going to be a fight. Jenl. cha, line 987
I have thought
WORD GRAMMAR AND SYNTACTIC CODE-MIXING 119
(4) *TRU: der hat iiber faith + healing gesprochen. Jen2. cha, line 2383
he has about spoken
(5) TRU: to buy yourself in means that +... Jenl. cha, lines 977ff
DOR: du kannst dich nochmal einkaufen.
you can yourself once more buy in
and the domain of government was too large. The above formulation of the
government constraint includes the whole maximal projection and thus, for
example, bans switching between verbs and location adverbs, again contrary to
the evidence. Therefore a limited definition of government, involving only the
immediate domain of the lexical head, including its complements but not its
modifiers/adjuncts, was adopted and the Government Constraint was re-
phrased (Muysken 1989) in terms of L-marking:
Muysken and collaborators thus shifted from an early and quite general
definition of government to the more limited definition of Lrmarking in their
formulation of the Government Constraint. L-marking restricts government to
the relation between a lexical head and its immediate complements. Even the
modified version of the government constraint in terms of L-marking is
empirically not borne out, as we see from the following example:
Muysken (2000: 25) identifies two main reasons why the Government
Constraint, even in its revised form, is inadequate. The main reason is that
CATEGORIAL EQUIVALENCE2 undoes the effect of the government restriction.
The Government Constraint is furthermore assumed to insufficiently acknowl-
edge the crucial role functional categories are supposed to play in code-mixing.
Functional categories feature prominently in several approaches to code-
mixing. Joshi, for example, proposes that 'Closed class items (e. g. determiners,
quantifiers, prepositions, possessive, Aux, Tense, helping verbs, etc. ) cannot be
switched' (1985: 194). Myers-Scotton and Jake (1995: 983) assume that in
mixed constituents, all SYSTEM MORPHEMES3 that have grammatical relations
external to their head constituent (i. e. participate in the sentence's thematic
grid) will come from the language that sets the grammatical frame in the unit of
analysis (CP). And Belazi, Rubin and Toribio (1994) propose the Functional
Head Constraint.
Their model is embedded in the principles and parameters approach.
Belazi, Rubin and Toribio (1994) propose to restrict code-mixing by the
120 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
The language feature of the complement F-selected by a functional head, like all
other relevant features, must match the corresponding feature of that functional
head.
(8) *LIL: you don't need to wegwerfen. Jen2. cha, line 2555
throw away
(9) *DOR: sie haben einfach nicht ge#bother-ed. Ibron. cha, lines 1012, 14
they have simply not
WG over other theories which can only generate syntactic structures for
sentences. From the examples quoted so far, it is obvious that the audio data
this study is based on are transcribed as utterances, i. e. units of conversational
structure. For the grammatical analysis, however, I assume that conversational
speech consists of the instantiation of linguistic units, i. e. sentences. In other
words, every conversational utterance is taken to be a token of a particular type
of linguistic unit, the structural features of that unit being defined by the
grammatical rules of either German or English. When using a WG approach to
code-mixed data, one does not have to 'edit' the corpus prior to linguistic
analysis. Any material that cannot be taken as a token of either a German or
English word-form can be left in the texts, but if it cannot be linked to other
elements in the utterance via a relationship of dependency, it is not included in
the syntactic analysis. That is, all the words in a transcribed utterance that are
related to other words by syntactic relationships constitute the sentences the
grammatical analysis is based on. As far as I am aware, WG is the only syntactic
theory that can (and wants to) generate representations of actual utterances, and
facilitates the grammatical analysis of natural speech data without prior editing.
Another consequence of integrating utterances into the grammar is that a
word token must be able to inherit from its type. Obviously the token must
have the typical features of its type - it must belong to a lexeme and a word
class, it must have a sense and a stem, and so on. But the implication goes in the
other direction as well: the type may mention some of the token's
characteristics that are normally excluded from grammar, such as characteristics
of the speaker, the addressee and the situation. For example, we can say that
the speaker is a German/English bilingual and so is the addressee; the situation
thus allows code-mixing. This aspect of WG theory thus allows us to
incorporate sociolinguistic information into the grammar, by indicating the kind
of person who is a typical speaker or addressee, or the typical situation of use.
Treating utterances as part of the grammar has further effects which are
important for the psycholinguistics of processing. The main point here is that
WG accommodates deviant input because the link between tokens and types is
guided by the 'Best Fit Principle' (Hudson 1990: 45fD: assume that the current
token is-a the type that provides the best fit with everything that is known. The
default inheritance process which this triggers allows known characteristics of
the token to override those of the type. Let's take the deviant word /bAS9/ in
the following example:
(lOa) *TRU: xxx and warum waren keine bus(s)e [%pho: bAS9]? JenS. cha, line 331
why were there no buses
El) In English any verb follows its subject but precedes all its other
128 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
dependents. This holds true for main as well as subordinate clauses and
gives rise to SVO order in both clause types.
E2) Subordinators, e. g. because, require a following finite verb as their
complement. A word's complement generally follows it. 14
For German the most relevant rules15 concerning word order in main and
subordinate clauses are:
Gl) A default finite verb follows one of its dependents but precedes all other
dependents. This gives rise to a verb second (V2) word order in German
main clauses.
G2) A finite verb selected by a lexical subordinator/complementizer takes all
its non-verb dependents to the left, i. e. it is a 'late'16 verb.
G3) Subordinators/complementizers, e. g. daft, select a 'late' finite verb as their
complement. 17 According to G2 finite 'late' verbs follow all their non-verb
dependents.
(11) Ich glaube nicht, da|3 wir die Dorit schon gekannt haben
I think not that weDorit already known have
JenS. cha, line 83
The utterance initial main clause displays V2 word order. The finite auxiliary
haben which depends on the subordinates daft, on the other hand, is in clause
final position following all other constituents including non-finite verbs like
gekannt. In English finite verbs in subordinate clauses do not behave differently
from finite verbs in main clauses. Therefore we do not have to override the
default rule El in the 'isa-hierarchy' of grammar rules. Because German finite
verbs depending on a subordinates take a different word order position than
'independent' finite verbs, we need a more specific rule (G2) that overrides the
default rule (Gl) in the cases stated, i. e. finite verbs selected by German
subordinators.
The pre-minimalism constituent based models discussed in section 2 all
have difficulties accounting for mixing between SVO and SOV languages
because of the opposite setting of the branching parameter. I will show in the
next section that this code-mixing pattern can be studied productively in terms
ofWG.
(12) *MEL: ich hab(e) gedacht, there is going to be a fight. Jenl. cha, line 987
I have thought
(13) *MEL: I forgot, dass wir alle wieder eine neue partie angefangen haben.
that we all again a new game started have
Jenl. cha, line 2541
(14) *TRU: die mutter wird ihr gelernt haben, how to keep young.
her mother would her taught have Jenl. cha, line 2016
(15) *DOR: wenn du short hist, you -wouldn't talk.
when you are
*DOR: aber wenn man geld hat, you talk.
but when one money has Jen3, line 581-2
(16) *TRU: er schreibt fuenfzehn, if you leave it in your hand.
he counts fifteen Jen2. cha, line 932
(17) *LIL: das haengt davon ab, what 'nasty' is(t).
that depends on Jen2. cha, line 1062
Note that the null hypothesis is born out in examples (12)-(17) and in the vast
majority of monolingual and mixed dependencies19 in the German-English
corpus. The WG rules determining the word order in main and subordinate
clauses also hold. These findings are furthermore supported by the quantitative
analysis of 1, 350 monolingual and 690 mixed dependency relations in a 2, 000
word monolingual sample corpus and a 7, 000 word code-mixed corpus (see
Eppler 2004).
130 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
(18) DOR: es war unsere [... ] Schuld because man fiihlt sich
it was our fault tone feels
mit den eigenen Leuten wohler.
with the own people happier. Ibron. cha, line 221
Because in the above example can be argued to be a single lexical item inserted
WORD GRAMMAR AND SYNTACTIC CODE-MIXING 131
Because also enters syntactic relations where the word on which it depends is
English (eat) and its dependent is German (schmeckt), as in:
(20) DOR: eat it with der Hand-! because das schmeckt ganz anders.
the hand it tastes very differently
Ibron. cha, line 2214
or vice versa, e. g. because has a German head verb (habe) but an English
complement (know):
(21) MEL: ich hab's nicht einmal gezahlt because I know I'm going to lose.
I have it not even counted
Jenl. cha, line 881
The German subordinator of reason, weil, on the other hand, only enters into
monolingual dependency relations:
(22) DOR: dann ist sie, weil sie so ungliicklich war, doit gestorben.
then has she, because she so unhappy was, there died
Ibron. cha, line 1002
Table 3: Verb position after weil partly based on Uhmann (1998: 98)
Weil Vf V2 Vf V2
Eppler (2004) 34 25 58% 42%
BYU (Vienna) 62 11 85% 15%
Farrar (1998) BYU 1147 517 69% 31%
Schlobinski (1992) 74 22 70% 23%
Uhmann (1998) 24 19 56% 44%
Dittmar (1997) 99 29 77. 3% 22. 7%
Table 3 shows that between 15 per cent and 44 per cent of dependent verbs in
these corpora are not in final position. So weil+V2 word order is not just a
peculiarity of the German spoken by my bilingual informants.
We thus have two problems to solve: 1) the asymmetrical distribution of
because and weil in the corpus; and 2) the word order variation in both mixed
and monolingual causal clauses introduced by because and weil. In the next
section I will suggest possible solutions to these two problems.
4. 2 Possible explanations
(25) Almost jedes moi is Suppe gewen, because mir ban kei
every time is it soup be we have no
Zeit khat fer Supper recht essen.
time had for soup properly to eat
Treffers-Daller (1994: 192-5) discusses (25) and (26) and suggests analyzing the
conjunctions in these two examples as coordinators. For monolingual English
Schleppegrell (1991: 323) argues that 'a characterisation of all because clauses as
subordinate clauses [... ] is inadequate'. The possibility of a paratactic24 function
for because will be discussed in the next section.
Gardner-Chloros's (1991) French/Alsatian data also offer an interesting
example of two Alsatian clauses linked by a French causal conjunction.
specific context that requires dependent verbs to take all their dependents to
the left. As because is an English subordinates which does not specify that its
complement has to be a clause final verb, we get main clause word order (SVO
in monolingual English or V2 in mixed utterances).
Supporting evidence for this interpretation comes from the six instances
where the finite verb follows a dependent other than its subject (cf. examples
23-24 and 27 below).
(27) DOR: I lost because # dreimal gab sie mir drei Konige.
three times gave she me three kings
Jenl. cha, line 817
In the above example the verb is in second position, but the clause is clearly not
SVO. The finite verb is preceded by an adverbial but followed by the subject.
In other words, the clause displays the V2 order expected in German main
clauses.
But how do we know that because and the because-clause, are used in a
restrictive subordinating way in examples (23), (24) and (27)? This question
needs to be addressed because research conducted by, amongst others,
Rutherford (1970), SchleppegreU (1991) and Sweetser (1990), cast doubt on
the characterization of all because-danses as subordinate clauses. Especially in
spoken discourse, because can be used in a variety of non-subordinating and not
strictly causal functions.
Several criteria have been proposed to distinguish between restrictive (i. e.
subordinating25) and non-restrictive because-clauses (Rutherford 1970). In
sentences containing restrictive because clauses yes/no questioning of the whole
sentence is possible; pro-ing with so or neither covers the entire sentence; they
can occur inside a factive nominal; and if another because clause were added,
the two causal clauses would occur in simple conjunction. In semantic terms the
main and the subordinate clause form one complex proposition and the
because-clause provides the cause or reason for the proposition itself. This
causal relationship is one of 'real-world' causality (Sweetser 1990: 81). Chafe
(1984) asserts that restrictive because clauses have a reading that presupposes the
truth of the main clause and asserts only the causal relation between the clauses.
These clauses tend to have a commaless intonational pattern.
I will now apply these characteristics to some of the causal clauses
introduced by because in the corpus cited so far. Utterance (27) passes all of
Rutherford's (1970) syntactic criteria for restrictive because-clauses. The main
and because-clanses form one complex proposition with a reading in which 'her
giving the speaker three kings' is the real world reason for the speaker losing the
game of cards. The truth of the sentence-initial clause is presupposed and the
causal relation between the two clauses is asserted. These properties of (27)
speak for a restrictive analysis. The intonational contour of the utterance,
however, displays a short pause after the conjunction.26 Note furthermore that
the causal clause in (27) contains a pre-posed constituent that triggers inversion,
i. e. a main clause phenomenon (Green 1976). So there are indicators for both
a restrictive/subordinate reading but also syntactic and intonational clues that
136 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
This example passes none of Rutherford's (1970) 'tests'. The intonational drop
before the conjunction which intonationally separates the two clauses also
suggest a non-subordinate analysis for (28). A restrictive reading of the whole
construction is awkward if not unacceptable: feeling relaxed in the company of
fellow compatriots is not the cause or reason for feeling guilty. The non-
restrictive reading in which the because clause provides the reason why the
speaker said 'it was our own fault' is far more plausible. The because clause,
furthermore, indicates an interpretative link between clauses that are several
utterances apart: the last utterance in (28) provides a 'long-distance' reason for
the first utterance in this sequence. Schleppegrell (1991: 333) calls these uses of
because 'broad-scope thematic links'. They can be only identified when a corpus
provides the relevant context for the example. The wider context also identifies
the clause preceding the causal clause as presupposed and thematic. The
information provided in the causal clause is new and asserted.
The analysis so far suggests that because is used in non-restrictive and non-
subordinating functions in code-mixed utterances in my corpus. Without
repeating them, I will now briefly discuss the other examples in which because
introduces a clause with predominantly German lexical items (Examples 19-20
and 23-24). Example (19) is a response to a preceding wh-question and thus an
independent utterance, the information presented in the reply is not
informationally subordinated, it forms the focus of the discourse and provides
new information (Schleppegrell 1991: 31). Example (20) has two intonational
contours. The intonational rise and the verb first order mark the initial clause as
a command or suggestion, i. e. an independent proposition; the following
because clause then represents an elaboration of that proposition. The content
of the causal clause is therefore not presupposed. Example (20) displays all the
characteristics of an 'epistemic' (Sweetser 1990) because, which indicates
'elaboration and continuation in non-subordinating and non-causal contexts'
(Schleppegrell 1991: 323). The because clause in example (23) is preceded by a
short pause, contains a main clause phenomenon (extraction), and is reflexive
WORD GRAMMAR AND SYNTACTIC CODE-MIXING 137
on the previous discourse; finally, the because clause in (24) follows a rising
intonation of the initial tag, and again explicitly mentions the speaker's
knowledge state ('it's true').
We can conclude that those clauses in which because has a German (V2)
verb as its complement, display more characteristics of 'non-restrictive'
(Rutherford 1970) clauses and should therefore be analyzed as paratactic
rather than subordinating. The Word Grammar rules formulated in section 3
still account for the data because if because is not analyzed as a subordinator, the
default rule Gl is not overridden and G2 and G3 do not get activated.
The analysis of the code-mixed data discussed so far indicates that the
predominantly German clauses introduced by because fulfil functions that are
not strictly causal but rather epistemic, broad-scope thematic link, etc. This
distinct usage is also reflected in their structural and intonational patterns. We
can therefore assume that we are dealing with non-restrictive because that is non-
subordinate and thus triggers main clause (V2) word order.
However, we also need to consider the monolingual data. The monolingual
German data from my corpus are more worrying at first sight Like because, well
was traditionally analyzed as a subordinating conjunction with causal meaning
which takes a finite verb as its complement These grammar rules are not
absolutely adhered to by my informants and monolingual speakers of German.
Only 58 per cent of verbs depending on well in the speech of my informants are
in clause final/late' position. Table 3 shows, furthermore, that in corpora of
similar, i. e. southern, varieties of German only 3. 1 -85 per cent (with an average of
approximately 67 per cent) of the subordinate clauses introduced by well are
grammatical according to the rules for monolingual German as stated in section 3.
The recent German literature on well constructions (Giinthner 1993, 1996;
Pasch 1997; Uhmann 1998), however, suggest an explanation for the
monolingual German data and opens up the possibility for an interesting
interpretation of the mixed data. There is agreement among the above named
researchers that a) there is considerable variation in the use of well + V2 or well
+ Vf; b) well + V2 is most frequent in southern German dialects; and c) weil
clauses with verb final placement and weil clauses with main clause (V2) word
order are found to show systematic but not absolute differences. In a nutshell,
the analysis for German weil is similar to the analysis proposed for English
because: there are two types of weil clauses, one strictly subordinating one, and
several non-restrictive paratactic uses. The factor that best seems to account for
the data is the information structure of the construction. If pragmatics and
syntax, which in German is a much clearer indicator than in English, fail to
provide clear criteria as to which type of ^/-construction we are dealing with,
intonation can once again help to disambiguate. Example (29) from my corpus
illustrates epistemic weil + V2:
(29) LIL: sie hat sich gedacht, die [/] die muss doch Wien kennenlernen,
'She thought she needs to get to know Vienna'
weil die eltern sind beide aus Wien.
because parents are both from Vienna
JenS. cha, line 107-8
138 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
early 20th century went through a phase that mirrors present-day English in the
distribution between because and for. In modern Pennsylvania German, well
does not seem to be able to function as subordinates in its own right any longer
and it has to be backed up by another complementizer to trigger verb final
placement. This supports rule G2 (section 3) which implicitly proposes a
subordinate feature on lexical complementizers. Modern PG seems to have lost
this feature and therefore needs to be 'backed up' by another subordinates to
trigger verb final word order.
Dann in modern PG, on the other hand, after having gone through the stage
of fer (<Engl. for), is eventually replaced by because, as in my data. This
development not only backs up the speculation voiced in section 4. 1. 2, i. e. that
the discrepancy between my and Clyne's (1973) German-English corpora might
be due to prolonged language contact, but also the qualitative analysis presented
in section 4. 2. 2.
The WG stipulation of a subordinate feature on German complementizers
(Rule G3) is furthermore supported by data from another language contact
situation with word order contrasts in subordinate clauses: French Dutch
contact in Brussels. The most frequently borrowed subordinates in Brussels
Dutch is tandis que. Treffers-Daller (1994: 191) observes that the Dutch
equivalent of tandis que, terwijl, is rarely used in her corpus. In those cases that
do occur, the Dutch conjunction is followed by the Dutch complementizer dat.
like well in Pennsylvania German, Brussels Dutch terwijl may also have lost
the subordinate feature and require an 'obvious' complementizer to trigger verb
final placement.
I do not claim that the present work illuminates theories of language structure
but it confronts a linguistic theory, Word Grammar, with statistical data, and
shows that this theory of language structure can be successfully and
illuminatingly used for the analysis of monolingual and code-mixed construc-
tions. The WG formulation of the null hypothesis is born out with just a
handful of exceptions, and the WG rules determining word order in
monolingual German or English and code-mixed clauses also hold.
The investigation of word order in subordinate clauses, furthermore, shows
that the null hypotheses seems to be correct even in cases where we would
expect restricitions on code-switching due to surface word order differences
between the two grammars involved in mixing. The quantitative analysis of
monolingual and code-mixed because and well clauses revealed that a) the core
group of informants favour the English causal conjunction because over German
weil or denn; the use of well and denn are restricted to monolingual German
contexts, and because is also used to introduce mixed utterances; b) the word
order in weil clauses varies between verb final, as required in subordinate
clauses, and verb second, the main clause order; the coordinating conjunction
denn only occurs once and with main clause order, as expected; mixed clauses
introduced by because invariably have verb second structure. Independent
research on the syntactic, intonational, semantic and pragmatic properties of
monolingual because and weil clauses has shown that these properties cluster to
form two main types of causal clauses: restrictive and non-restrictive (Rutherford
1970). The qualitative analysis of the monolingual causal clauses in the corpus
revealed that they also fall into these two types and that the mixed utterances
introduced by because predominantly have the grammatical properties of non-
restrictive clauses. Thus Boumans' (1998: 121) hypothesis that 'foreign
conjunctions do not trigger verb-final in German clauses simply because they
are used in functions that require main clause order' could be verified. The
quantitative analysis of because and weil clauses has furthermore demonstrated
how frequency distributions of a specific grammatical pattern in monolingual
speech data can be combined with our knowledge about syntactic and pragmatic
properties of grammars to handle frequency in bilingual data (Muysken 2000).
The WG analysis of German (and Dutch) lexical subordinators having a
'subordinate' feature which triggers verb final placement was furthermore
supported by data from two other language contact situations (Pennsylvania
German and Brussels Dutch) in which certain subordinators seem to have lost
this feature and therefore to require 'backing up' from overt complementizers.
WORD GRAMMAR AND SYNTACTIC CODE-MIXING 141
References
Belazi, H. M., Rubin, E. J. and Toribio, A. J. (1994), 'Code switching and X-bar theory:
The functional head constraint5. Linguistic Inquiry, 25, 221-37.
Bentahila, A. and Davies, E. E. (1983), 'The syntax of Arabic - French code -
switching'. Lingua, 59, 301-30.
Bolle, J. (1995), 'Mengelmoes: Saranan and Dutch language contact', in Papers from the
Summer School Code-switching and Language Contact. Ljouwerl/Leeuwarden: Fryske
Akademie, pp. 290-4.
Boumans, L. (1998), The Syntax of Codeswitching: Analysing Moroccan ArabiclDutch
Conversations. Tilburg: Tilburg University Press.
Chafe, W. L. (1984), 'How people use adverbial clauses'. Berkeley Linguistics Society, 10,
437-49.
Chomsky, N. (1981), Lectures on Government and Binding. Dordrecht: Foris.
Clyne, M. G. (1973), 'Thirty years later: Some observations on "Refugee German" in
Melbourne', in H. Scholler and J. Reidy (eds) Lexicography and Dialect Geography,
Festgabefor Hans Kurath. Wiesbaden: Steiner, pp. 96-106.
— (1987), 'Constraints on code-switching: how universal are they?' Linguistics, 25, 739-
64.
DiSciullo, A-M., Muysken P. and Singh, R. (1986), 'Government and Code-Mixing'.
Journal of Linguistics, 22, 1-24.
Eppler, E. (1999), 'Word order in German-English mixed discourse'. UCL Working
Papers in Linguistics, 11, 285-308.
— (2004), '... Because dem Computer brauchst' es ja nicht zeigen': Because + German
main clause word order'. International Journal of Bilingualism, 8, 127-43.
— German/English LIDES database <http: Htalkbank. org/datalLIDES/Eppler. zip>.
Gardner-Chloros, P. (1991), Language Selection and Switching in Strasbourg. Oxford:
Clarendon Press.
Green, G. M. (1976), 'Main clause phenomena in subordinate clauses'. Language, 52,
382-97.
Grosjean, F. (1995), 'A psycholinguistic approach to codeswitching', in P. Muysken and
L. Milroy (eds), One Speaker, Two Languages. Cambridge: Cambridge University
Press, pp. 259-75.
Gumperz, J. J. (1982), Discourse Strategies. Cambridge: Cambridge University Press.
Gunther, S. (1993), '... Weil-man kann es ja wissenschafuich untersuchen'- Diskur-
spragmatische Aspekte der Wortstellung in weil-Satzen'. Linguistische Berichte, 143,
37-59.
— (1996), 'From subordination to coordination?'. Pragmatics, 6, 323-56.
Hudson, R. A. (1980), Sociolinguistics. Cambridge: Cambridge University Press.
— (1997), 'Inherent variability and linguistic theory'. Cognitive Linguistics, 8, 73-108.
— (1990), English Word Grammar. Oxford: Blackwell.
— (2000), 'Grammar without functional categories', in R. Borsley (ed. ), The Nature and
Function of Syntactic Categories. New York: Academic Press, pp. 7-35.
Joshi, A. K. (1985), 'Processing of sentences with intrasentential code-switching', in L.
Dowry, L. Kartunnen and A. M. Zwicky (eds), Natural Language Parsing.
Cambridge: Cambridge University Press, pp. 190-205.
Lehmann, Ch. (1988), 'Towards a typology of clause linkage' in J. Haiman and S.
Thompson (eds), Clause combining in grammar and discourse. Amsterdam/
Philadelphia: John Benjamins, pp. 181-226.
Louden, M. L. (2003), 'Subordinate clause structure in Pennsylvania German'. FGLSj
SGL Joint Meeting. London, 2003.
142 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Notes
1 The corpus was collected in 1993 from German-speaking Jewish refugees residing
in London. All transcripts are available on <http: //talkbank. ord/data/LIDES/
Eppler. zipX
2 Categorial equivalence is 'when the switched element has the same status in the
two languages, is morphologically encapsulated, shielded off by a functional
element from the matrix language, or could belong to either language' (Muysken
2000: 31).
3 Myers-Scotton and Jake (1995: 985) define system morphemes as morphemes that
do not participate in the thematic structure of a sentence, i. e. they are specified as [-
theta-role assigner/receiver]. A second feature characteristic of 'most' system
morphemes is the feature [+ Qantification]. A morpheme has a plus setting for
quantification within the Matrix Language Frame model, if it restricts possible
referents of a lexical category. Myers-Scotton and Jake (1995: 985) give tense and
aspect as examples for [+ QJ. Tense and aspect restrict the possible reference of
predicates (i. e. verbs and adjectives). Prototypical system morphemes are inflections
and most function words.
4 The WG approach of incorporating different syntactic properties of WORDS in
isa-hierarchies seems more economical and convincing.
5 The Free Morpheme Constraint (Sankoff and Poplack 1981: 5) prohibits switching
between a bound morpheme (pre- or suffix) and a lexical form unless the latter has
been phonologically integrated into the language of the bound morpheme.
6 Note the similarity of this corollary with the WG null hypothesis this study is based
on.
7 Constituency analysis is applied only to coordinate structures.
8 This system implies that code-mixing ought to be less frequent among typologically
quite different language pairs.
9 According to the theory of markedness (Scotton 1990: 90), speakers know that for a
particular conventionalized exchange in their community, a certain code choice will
be the unmarked realization of an expected rights and obligations set between
participants. They also know that other possible choices are more or less marked
because they are indexical of other than the expected rights and obligations set.
10 Smooth code-switches are unmarked by false starts, hesitations, lengthy pauses, etc.;
flagged switches are accompanied by discourse markers and other editing
phenomena (Poplack 1980).
11 A database is particularly important for studies of codes that do not have 'native'
speakers who can provide fairly reliable grammaticality judgements. A corpus is also
an essential test for the constraints on and models of code-mixing.
12 An alternative analysis of this example would be that it is ambiguous, i. e. it conforms
to two different models. The stem conforms to the English phonological model and
the suffix conforms to the German plural suffix; i. e. it is a morphologically
integrated borrowed stem.
13 The figures for individual cases need not be the same; cases of lexical diffusion
would seem to suggest the contrary (Hudson 1980: 168ff). And presumably the
entrenchment value for the general rule in such cases could be different from all the
individual rules.
14 Default inheritance rules apply to the few English constructions in which the
complement comes before the head.
15 These rules are not intended to cover scrambling, double infinitive constructions
and other well-known word order intricacies of German.
144 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
16 The term 'late' was chosen instead of 'final' because finite dependent auxiliaries in
double infinitive constructions can be followed by their non-finite dependents; cf.
endnote 15.
17 Support for this analysis comes from the fact that German subordinate clauses
lacking a subordinator/complementizer are V2 (or verb initial). Cf.:
Sie sagte, sie kennen Doris vs. Sie sagte, daft sie Doris kennen
She said they know Doris She said that they Doris know
According to G3, it is only subordinators/complementizers that select 'late' finite
verbs. So if a verb depends directly on another verb (kennen directly depending on
sagte and not daff) the default rule need not be overridden.
18 Exceptions to this rule are extraposition and double-infinitive constructions.
19 The null hypothesis is violated in five tokens of two construction types: word-order
violations of objects and negatives (see Eppler 2004).
20 The data this study is based on are transcribed in the LIDES (Language Interaction
Data Exchange) system. More information on the transcription system can be found
on <www. ling. lancs. ac. uk/staff/mark/lipps/ >.
21 See for example Clyne (1987), Gardner-Chloros (1984), Salmons (1990), Treffers-
Daller (1994).
22 Example (24) is an incomplete subordinate clauses. This does not effect the analysis
because the word order position of the relevant finite dependent verb is clear.
23 Since all my informants are from Vienna, I used only examples from the ten
Viennese informants for the Brigham Young Corpus (BYU) corpus. Farrar (1998)
counted all occurrences of weil in the speakers of southern German dialects from
the BYU corpus. Schlobinski's (1992) data are standard Bavarian; and the Uhmann
(1998) corpus is 'alemannisch-bairisch'.
24 Lehmann (1988) suggests that for clauses that are linked in a relationship of
sociation rather than dependency, 'paratixis' is a more appropriate term than
'coordination'.
25 Two clauses (X and Y) have been defined as being in a subordination relationship
'if X and Y form an endocentric construction with Y as the head' (Lehmann 1988:
182).
26 Note that in the English literature, Rutherford (1970) and Thome (1986), the
comma intonation is assumed to precede the conjunction. Schleppegrell (1991:
333) mentions the possibility of because followed by a pause.
7 Word Grammar Surface Structures and HPSG
Order Domains*
TAKAFUMI MAEKAWA
Abstract
In this chapter, we look at three different approaches to the asymmetries between
main and embedded clauses with respect to the elements in the left periphery of
a clause: the dependency-based approach within Word Grammar (Hudson
2003), the Constructional Head-driven Phrase Structure Grammar (HPSG)
approach along the lines of Ginzburg and Sag (2000), and the Linearization
HPSG analysis by Chung and Kim (2003). We argue that the approaches within
WG and the Constructional HPSG have some problems in dealing with the
relevant facts, but that Linearization HPSG provides a straightforward accounta of
them. This conclusion suggests that linear order should be independent to a
considerable extent from combinatorial structure, such as dependency or phrase
structure.
1. Introduction
There are two ways to represent the relationship between individual words:
DEPENDENCY STRUCTURE and PHRASE STRUCTURE. The former is a pure
representation of word-word relationships while the latter includes additional
information that words are combined to form constituents. If all work can be
done just by means of the relationship between individual words, phrase
structure is redundant and hence dependency structure is preferable to it. It
would therefore be worth considering whether all work can really be done with
just dependencies. We will look from this perspective at certain linear order
asymmetries between main clauses and subordinate clauses. One example of
such asymmetries can be seen in the contrast of (1) and (2). The former shows
that a topic can precede a fronted wA-element in a main clause:
It is clear that main clauses are different from subordinate clauses with respect
to the possibility of topicalization. It has been noted by a number of researchers
that elements occurring in the left periphery of the clause, such as interrogative
and relative pronouns, topic and focused elements, show such linear order
asymmetries (see Haegeman 2000; Rizzi 1997; and works cited therein).
The purpose of this overview chapter is to take a critical look at the current
treatment of such asymmetries within the frameworks of WORD GRAMMAR
(wo) and HEAD-DRIVEN PHASE STRUCTURE GRAMMAR (HPSG),
and ask how they should be represented in the grammar. We compare the WG
approach developed in Hudson (2003; see also Hudson 1995, 1999) with the two
relatively recent versions of HPSG: what can be called CONSTRUCTIONAL HPSG in
which grammars include hierarchies of phrase types (Sag 1997; and Ginzburg and
Sag 2000), and so-called LINEARIZATION-BASED HPSG (or LINEARIZA-
TION HPSG), in which linear order is independent to a considerable extent
from phrase structure and is analysed in terms of a separate level of 'ORDER
DOMAINS' (Pollard et al. 1994; Reape 1994; Kathol 2000, etc. ).1 It will be argued
that trie WG and the Construction HPSG approaches have some problems, but
that Linearization HPSG can provide a straightforward account of the facts.
The organization of this chapter is as follows. In the next section we consider
how a WG approach might accommodate the asymmetry between main and
subordinate wA-interrogatives. Section 3 then looks at a Construction HPSG
analysis along the lines of Ginzburg and Sag (2000). In section 4 we shall
outline a Linearization HPSG analysis developed by Chung and Kim (2003).
In the final section, we offer some concluding remarks.
(3)
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 147
The fact that it is the subject of the two verbs is indicated by the two arrows
labelled V (subject). The arrow labelled V indicates that raining is a 'SHARER'
of was. This is so named since it shares the subject with the parent verb. In a
notation adopted here, the dependencies that do not provide landmarks are
drawn below the words. Therefore, one of the V arrows, the one from raining
to it, is drawn below the words. We thus pick out a sub-set of total
dependencies of a sentence and draw them above the words. This sub-set is
called SURFACE STRUCTURE. Word-order rules are applied to it, and determine
the positioning of a word in relation to its landmark or landmarks. Thus, the
surface structure is the dependencies which are relevant for determining word
order. A word-order rule specifies that a subject normally precedes its
landmark, and another rule specifies that a sharer normally follows its
landmark, as illustrated by the representation in (3).
Among the rules that control the surface structure, THE NO-TANGLING
PRINCIPLE is the most important for our purpose: dependency arrows in the
surface structure must not tangle.2 This principle excludes the ungrammatical
sentence (4b):
(5)
(6)
148 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
On the other hand, Hudson (1990: 361-82; 2003) argues that the verb is a
complement of the wh-pronoun and thus depends on it
(7)
(8) a. Pat I know he's invited a friend. Jo: Oh, who [has he invited] ?
b. I know he's invited a friend, but I'm not sure who [he's invited].
Second, the pronoun is what is selected by the higher verb. In (9) wonder and
sure require a subordinate interrogative clause as their complement For a
clause to be subordinate interrogative, the presence of either a wh-pronoun, or
whether or if is required.
Third, the pronoun selects the verb's characteristics such as finiteness and
whether or not it is inverted. (10) illustrates that why selects a finite or infinite
verb as its complement, but when only selects a finite verb:
(11) indicates that why selects an inverted verb as its complement whereas how
come selects a non-inverted verb:
(12) illustrates that what, who and when select a to-infinitive, but why does not:
Hudson (2003) argues that all of these phenomena are easily accounted for if
the 2£>/z-pronoun is a parent of the next verb. In the framework of WG,
therefore, there is no reason to rule out any of (6) and (7); the sentence is
syntactically ambiguous. Thus, in What happened? what and happened depend on
each other, and the dependency structure may be either of (13a) and (13b):
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 149
(13) a.
b.
The preposed adjunct now would otherwise follow its parent need as in (14b),
but it precedes it. This situation is represented in WG by adding an extra
dependency 'EXTRACTEE' to now.
(15)
The arrow from need to now is labelled 'x <, > a', which means 'an adjunct
which would normally be to the right of its parent (" > a") but which in this case
is also an extractee ("x>")'. Thus the adjunct now is to the left of the parent
verb need.
With this background in mind, let us now turn to the asymmetry between
main and subordinate clauses in question: adverb-preposing is not possible in
subordinate interrogatives although it is possible in main interrogatives.
As stated above, a w/z-pronoun and its parent are mutually dependent. In (16a)
150 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
do is the complement of what whereas what is the extractee of do. Thus, the
dependency structure for (16a) would be either of (17a) and (17b). In the
former, what is the parent, and the dependency labelled 'c' from what to do is
put in surface structure. In the latter, however, do is the parent and the
dependency labelled 'x<' from do to what is put in surface structure. The
preposed adjunct now is labelled 'x <, > a', and precedes its parent do. As the
diagram shows, the 'x <, > a' arrow from do to now tangles with the vertical
arrow in (17a). Thus, it violates the No Tangling Principle. On the other hand,
there is no tangling in (17b), so it is the only correct WG analysis of (16a).
(17) a.
b.
(18)
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 151
Thus, WG can capture the linear order asymmetries of the main and
subordinate clauses in terms of dependencies in surface structure and general
principles on dependencies.
Although the WG analysis looks successful in accommodating the
asymmetry between the main and subordinate clauses, there are some
weaknesses. As surveyed above, the WG approach states that adjunct prepbsing
is possible out of main w/z-interrogatives because the preposed adjunct avoids
violation of the No-Tangling Principle due to the fact that it is a co-dependent of
the ^/z-element (Hudson 2003: 636). The argument along these lines would
suggest that extraction is allowed as long as it does not violate the No-Tangling
Principle. However, there are cases in which extraction out of embedded wh-
interrogative is excluded although it does not violate the No-Tangling Principle.
The data comes from the SUBJECT-AUXILIARY INVERSION (SAI) structures
illustrated by (19). 3
(19) Under no circumstances would I go into the office during the vacation.
Here the preposed SAI operator precedes the wA-element. Note that the
situation is completely on a parallel with the case of the adjunct preposing like
(16a), which is repeated here.
As we have seen, (22) is grammatical because it does not violate the No-
152 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
On the other hand, the same permutation of w/z-element and a topic is allowed
in subordinate clauses, as in (25):
Here we have yet another asymmetry between a main and a subordinate clause.
Our observation in (23)-(25) indicates that the word order which is grammatical
in subordinate clauses is ungrammatical in main clauses. Note that the No-
Tangling Principle cannot exclude the ungrammatical cases since they are all in
main clauses. Therefore, the only option we have is to specify word-order rules
to exclude ill-formed cases. Now the same problem as (21) and (20) arises
again. Such word-order rules would have to state that it is applied to a main
clause but not to a subordinate clause. However, it is impossible for WG rules
to refer to a clause since WG does not recognize any unit larger than a word.
Thus, we cannot adopt a word-order rule approach, either.
We have pointed out that the No-Tangling principle is not effective enough
to accommodate the cases of preposing of an SAI operator, another asymmetry
between main and subordinate w/z-interrogatives. Recall that the most important
assumption for a WG approach is that the wA-pronoun is the parent of the
subordinate wA-interrogatives. We should note that this assumption itself is not
without problem. Consider examples in (26a) and (26b); the former is the one
cited by Hudson himself as a problematic data for his analysis (Hudson 1990:
365): 4
In WG treatment of wh-pYonoun, which and who are not only the subject of
have and shot, respectively, but also the subject of is. The verb should agree in
number with its subject, so have/shot and is should both agree with which/who.
Which in (26a) should share its plurality with students since the former is a
determiner of the latter; who in (26b) should share its plurality with themselves
since the former is the antecedent of the latter. This does not explain the
morphology of the copula verb in both sentences, which requires the singular
subject. This analysis would predict sentences like the following to be
grammatical:
The copular verb is are, not is, agreeing with its subject which in (a) and who in
(b). These sentences are, however, ungrammatical. Thus, the assumption that
the w^-pronoun is the parent of the subordinate interrogatives has a weakness.
We should also note that there are some cases where an extractee is allowed
to precede the complementizer. The following examples are from Ross (1986):
(28) a. Handsome though Dick is, I'm still going to marry Herman,
b. The more that you eat, the less that you want.
In (28a), the first clause is the subordinate clause, and the adjective handsome, a
complement of is, is in front of the complementizer though. In (28b) the more,
which is an object of eat and want, is followed by the complementizer that.5 It
would be natural to assume the fronted elements in these examples to be an
extractee in WG's terms; but if so, the dependency arrow from the verb to the
extractee would tangle with the vertical arrow to the complementizer, and hence
the resulting structure in (29) violates the No-Tangling Principle. 6
(29)
(30)
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 155
(31)
Phrases are classified along two dimensions: clausality and headedness. The
clausality dimension distinguishes various kinds of clauses. Clauses are subject
to the constraint that they convey a message. Core clauses are one subtype,
which is defined not to be modifiers and headed by finite verbal forms or the
156 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
auxiliary to. The headedness dimension classifies phrases on the basis of their
head-dependent properties, i. e. whether they are headed or not, what kind of
daughters they have, etc. A general property of headed phrases (hd-ph} is the
presence of head daughter, and this phrasal type is constrained as follows:
(33) hd-fill-ph:
This constraint requires the following properties. First, the head daughter must
be a verbal projection. Second, one member of the head daughter's SLASH set
is identified with the LOCAL value of the filler daughter. Third, other elements
that might be in the head daughter's SLASH must constitute the SLASH value
of the mother. Ginzburg and Sag (2000) treat the topicalization constructions as
a subtype of hd-fill-ph, and posit a type topicalization-clause (top-cl). It is also
assumed to be a subtype of core-cl. The type top-cl is subject to the construction-
particular constraint which takes the following form:
(34) top-cl:
(35) a. Problems of this sort, our analysis would never account for.
b. * She subtly suggested [problems of this sort, our analysis would never
account for].
(Ginzburg and Sag 2000: 50)
Topicalization of such a clause is ruled out by (34). The filler daughter of the
topicalized clause is constrained to be [WH {}], the effect of which is to prevent
any w/z-words from appearing as the filler or an element contained within the
filler. The constraints introduced above are unified to characterize the
topicalized clause constructions.
Given the above constraints, a sentence with a preposed adjunct will have
something like the following structure (Bouma et al. 2001; Kim and Sag 2002):
(36)
As noted above, certain types of adverbial phrases are selected by the verbal
head and listed in the ARG-ST list, along with true arguments. Thus, adjunct-
preposing and standard cases of topicalization can be given a unified treatment.
The ARG-ST of the verb visit thus contains an adverbial element, whose synsem
is specified as a gap-ss. Gap-ss, a subtype of synsem, is specified to have a
nonempty value for the feature SLASH. Its LOG value corresponds to its
SLASH value, as indicated by the shared value [1]. The ARGUMENT REALI-
ZATION PRINCIPLE ensures that all arguments, except for a gap-ss, are realized
on the appropriate valence list (i. e. SUBJ(ECT), COMP(LEMENT)S or
SP(ECIFIER), and hence are selected by a head. Note that in (36) the gap-ss in
the ARG-ST list of visit does not appear in a COMPS list The nonempty
158 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
SLASH value is incorporated into the verb's SLASH value.7 The verb's
SLASH value is projected upwards in a syntactic tree from the head daughter to
mother, due to the GHFP. The termination of this transmission, which is
effected by subtypes of the hd-fill-ph constructions, occurs at an appropriate
point higher in the tree: a dislocated constituent as specified as [LOG [1]]
combines with the head that has the property specified in the constraint for hd-
fill-ph in (33).
Now we can consider how this approach might accommodate the asymmetry
between main and subordinate w/z-interrogatives. The data observed in the last
section can be summarized as (37):
(37) Distribution of SAJ operator, ro/z-element and topic (Based on Chung and Kim
2003)
Main clause Embedded clause
*
TOP<WH ok (16a) (16b)
WH<TOP * (24) ok (25)
* *
OP<WH (21) (20)
*
WH<OP (23a) ok (23b)
We will begin with the asymmetry in terms of the interaction of a topic and a
wA-element. The relevant data is repeated here for convenience with the labels
and brackets added:
SI is composed of the topic filler and the clausal head, S2. S2 of the two
sentences in (38) is of the type ns-wh-int-cl. What we need to do is to check
compatibility of a clause of the type top-d and that of the type ns-wh-int-cl, with
the latter being the head of the former. We saw above that clauses of the type
top-d are constrained by various constraints; the unification of the constraints is
represented as follows:
(39)
Of note here is that the LOG value of the mother [2] is shared with that of the
head, due to the GHFP (32). This means that the head daughter, in this case a
clause of the type ns-wh-int-d, should have a finite verb as its head and its 1C
value is +. According to the hierarchy in (31), a clause of this type is
characterized as unification of core-d, int-d, hd-ph, hd-fill-ph, wh-int-ph and ns-
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 159
(40)
The shared value between the features 1C and INV(ERTED) guarantees that if
the clause of this type is inverted ([INV +]) then its 1C value is -+, that is, it
appears in a main clause; if it is uninverted ([INV — ]) then it should be in an
embedded clause ([1C — ]). The S2 of (38a), the head daughter of the whole
clause, is inverted, so its INV value is +, and hence its 1C value is +. This
satisfies the requirement stated in (39) that the head daughter of a topicalization
construction is an independent clause.
The S2 in (38b) is an instance of ns-wh-int-cl as in the previous case, but it is
not inverted (i. e. [INV — ]) in this case. The S2 should then be specified as [1C
— ] due to constraint (40). As we saw above, the head daughter of a
topicalization construction should be [1C +]. This is the reason why the
embedded interrogative does not allow topicalization. Under Ginzburg and
Sag's (2000) analysis, the asymmetry between main and subordinate wh-
interrogatives in terms of adjunct preposing is thus due to the conflict of the
requirement from the topicalization constructions and the embedded
interrogative constructions: the former requires [1C +] while the latter is
specified as [1C — ].
We will move on to the data problematic to a WG approach. Let us first
consider how Ginzburg and Sag's (2000) approach might deal with the
asymmetry in terms of the order WH <TOPIC. As we observed in (24), the
WH < TOPIC order is ungrammatical in main clauses. The data is repeated
in (41), with the square brackets and the labels added:
(41) a. * [si To whom, [§2 a book like this, would you give?]]
b. * [si For what kind of jobs [52 during the vacation would you give into the
office?]]
(42) a. the man (si to whom, [52 liberty, we could never grant]]
b. ?I wonder [Si to whom [S2 this book, Bill should give. ]]
c. I was wondering [51 for which job, [32 during the vacation, I should give
into the office. ]]
160 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
It is not clear exactly what sort of constraints preposed operators must satisfy in
Ginzburg and Sag's (2000) system, but it is clear that the S2 in (43a, b) and the SI
in (44a, b) are clauses of the type ns-wh-int-cl. Therefore, they should at least
satisfy constraint (40). As we saw above, this constraint guarantees that the clause
of this type is inverted ([INV +]) if it is in a main clause ([1C +]) and that it is
uninverted ([INV —]) if it is in an embedded clause ([1C — ]). All the
occurrences of ns-wh-int-cl in (43) and (44) are inverted, so they all should be
independent ([1C +]), and that means they cannot appear in subordinate clauses.
This correctly predicts that (43b) is ungrammatical, but it is problematic to (44b);
we have here an example of the clause of the type ns-wh-int-cl, which appears in a
subordinate clause ([1C — ]), but is inverted ([INV +]). Nothing in Ginzburg and
Sag's (2000) constraints rules out the (a) examples in (43) and (44).
It seems, then, that an approach to the asymmetry between main and
subordinate wMnterrogatives within the framework of Ginzburg and Sag (2000)
has some problems.
In this framework, word order is determined not at the level of the local tree,
but in a separate level of 'order domains', an ordered list of elements that
contain at least phonological and categorical information (see, e. g. Pollard et al.
1994; Reape 1994; and Kathol 2000). The list can include elements from
several local trees. Order domains are given as the value of the attribute
DOM(AIN). At each level of syntactic combination, the order domain of the
mother category is computed from the order domains of the daughter
constituents. The domain elements of a daughter may be COMPACTED to form a
single element in the order domain of the mother or they may just become
elements in the mother's order domain. In the latter case the mother has more
domain elements than the daughters. For example, let us consider the following
representation for the sentence Is the girl coming? (Borsley and Kathol 2000):
(45)
The VP is coming has two daughters and its domain contains two elements, one
for is and one for coming. The top S node also has two daughters, but its order
domain contains three elements. This is because the VP's domain elements
have just become elements in the S's order domain, whereas those of the NP
are compacted into one single domain element, which ensures the continuity of
the NP. Discontinuity is allowed if the domain elements are not compacted: is
and coming are discontinuous in the order domain of the S.
The notable feature of Chung and Kim's (2003) analysis is that each element
of a clausal order domain is uniquely marked for the region that it belongs to
(Borsley and Kathol 2000; Kathol 2000, 2002; Perm 1999). The positional
assignment is determined by the following constructional constraints:
a
(46) -
b.
c.
162 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
(47) Distribution of English left peripheral elements (Chung and Kim 2003)
Marker field Topic field Focus field
1 2 3
(48) states that the elements in position 1 should precede those in 2, which
should in turn precede those in 3.
Now let us consider how this approach might accommodate the asymmetry
between main and subordinate w/t-interrogatives. The summary of the relevant
data given in (37) is repeated here in (49):
(50)
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 163
This order domain does not violate the Topological Linear Precedence
Constraint in (48), and hence accounts for the grammaticality of (16a), repeated
here for convenience:
(52)
(52) violates the Topological Linear Precedence Constraint since its DOM
element marked 2 precedes that marked 1. This explains the ungrammaticality
of (16b).
Thus, Chung and Kim's (2003) approach can accommodate the asymmetry
between main and embedded clauses with respect to a topic and a wA-element.
The fact that WH < TOP is excluded from the main clauses is accounted for
along the same lines. This linear order leads to the following order domain:
(54)
Here, the element with 3 precedes that with 2, which violates (48), which
accounts for the ungrammaticality of the sentences in (24).
For embedded clauses, on the other hand, (46a) and (46c) require that a topic
should be in 2 and a wA-phrase in position 1, respectively. The resulting order
is (56):
(56)
164 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Constraint (46b) states that w/z-elements and operators are both assigned to
position 1 in main clauses. This accounts for the ungrammaticality of the WH
< OP and the OP < WH orders in main clauses: the competition for a single
position between these two elements entails that they cannot co-occur.
The OP < WH order is correctly excluded since it entails 3 < 2, which violates
(48).
5 Concluding Remarks
In this chapter, we have looked at three different approaches to the
asymmetries between main and embedded clauses with respect to the elements
in the left periphery of a clause. We compared the dependency-based
approach developed within WG (Hudson 2003) with the Constructional
HPSG approach along the lines of Ginzburg and Sag (2000), and the
Linearization HPSG analysis by Chung and Kim (2003), and argued that the
approaches within WG and the Constructional HPSG have some problems in
dealing with the relevant facts, but that Linearization HPSG provides a
straightforward account of them.
As we discussed at the outset of this chapter, dependency structure is simpler
than phrase structure in that the former only includes information on the
relationship between individual words, but the latter involves additional
information about constituency. Other things being equal, simpler representa-
tions are preferable to more complex representations. This might lead to the
conclusion that WG is potentially superior to HPSG. We have shown,
however, that both the dependency-based analysis in WG and the constituency-
based analysis in Constructional HPSG are not satisfactory in accounting for the
linear order facts. These two frameworks follow the traditional distinction
between the rules for word order and the rules defining the combinations of
elements.9 We should note, however, that the rules for word order are applied
to local trees in Constructional HPSG and to dependency arrows in WG.
Sisters must be adjacent in Constructional HPSG whereas in WG the parent
and its dependent can only be separated by elements that directly or indirectly
depend on one of them. This means that the linear order is still closely tied to
the combinatorial structure. That these frameworks cannot accommodate
certain linear order facts suggests that neither dependency structure nor phrase
structure is appropriate as the locus of linear representation. We saw above that
the linearization HPSG analysis gives a satisfactory account of linear order of
elements in the left periphery. This conclusion suggests that we need to
separate linear order from combinatorial mechanisms more radically than the
above traditional separation of the rules.
References
Abeille, Anne and Godard, Daniele (1997), 'The syntax of French negative adverbs',
in Danielle Forget, Paul Hirschbuhler, France Martineau, and Maria L. Rivero
(eds), Negation and Polarity: Syntax and Semantics. Amsterdam: John Benjamins, pp.
1-17.
Baltin, Mark (1982), 'A landing site for movement rules'. Linguistic Inquiry, 13, 1-38.
Borsley, Robert D. (2004), 'An approach to English comparative correlatives', in Stefan
Muller (ed. ), Proceedings of the HPSG04 Conference. Stanford: CSLI Publications, pp.
70-92.
Borsley, Robert D. and Kathol, Andreas (2000), 'Breton as a V2 language'. Linguistics,
38, 665-710.
Borsley, Robert D. and Przepiorkowski, Adam (eds), Slavic in Head-Driven Phrase
Structure Grammar. Stanford: CSLI Publications.
166 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Bouma, Gosse, Malouf, Rob and Sag, Ivan A. (2001), 'Satisfying constraints on
extraction and adjunction'. Natural Language and Linguistic Theory, 19, 1-65.
Chomsky, Noam and Lasnik, Howard (1977), 'Filters and control'. Linguistic Inquiry, 8,
425-504.
Chung, Chan and Kim, Jong-Bok (2003), 'Capturing word order asymmetries in English
left-peripheral constructions: A domain-based approach', in Stefan Miiller (ed. ),
Proceedings of the 10th International Conference on Head-Driven Phrase Structure
Grammar. Stanford: CSLI Publications, pp. 68-87.
Ginzburg, Jonathan and Sag, Ivan A. (2000), Interrogative Investigations. Stanford: CSLI
Publications.
Haegeman, liliane (2000), 'Inversion, non-adjacent inversion and adjuncts in CP'.
Transaction of the Philological Society, 98, 121-60.
Hudson, Richard A. (1990), English Word Grammar. Oxford: Blackwell.
— (1995), 'HPSG without PS?'. Available: www. phon. ucl. ac. uk/home/dick/unpub. htm.
(Accessed: 21 April 2005).
— (1999), 'Discontinuity'. Available: www. phon. ucl. ac. uk/home/dick/disconthtm. (Ac-
cessed: 21 April 2005).
- (2003), 'Trouble on the left periphery'. Lingua, 113, 607-42.
— (2005, Feburuary 17-last update), 'An Encyclopedia of English Grammar and Word
Grammar', (Word Grammar). Available: www. phon. ucl. ac. uk/home/dick/wg. htm.
(Accessed: 21 April 2005).
Kathol, Andreas (2000), Linear Syntax. Oxford: Oxford University Press.
— (2002), 'Linearization-based approach to inversion and verb-second phenomena in
English', in Proceedings of the 2002 LSK International Summer Conference Volume II:
Workshops on Complex Predicates, Inversion, and 0 T Phonology, pp. 223-34.
Kim, Jong-Bok and Sag, Ivan A. (2002), 'Negation without head-movement'. Natural
Language and Linguistic Theory, 20, 339-412.
Koizumi, Masatoshi (1995), 'Phrase Structure in Minimalist Syntax'. (Unpublished
doctoral dissertation, MIT),
van Noord, Gertjan and Bouma, Gosse (1994), 'Adjuncts and the processing of lexical
rules', in Fifteenth International Conference on Computational Linguistics (COLING
'94), pp. 250-6.
Perm, Gerald (1999), 'Linearization and WH-extraction in HPSG', in R. D. Borsley and
A. Przepiorkowski (eds) Slavic in Head-Driven Phrase Structure Grammar. Stanford:
CSLI Publications, pp. 149-82.
Pollard, Carl, Kasper, Robert and Levine, Robert (1994), Studies in Constituent Ordering:
towards a Theory of Linearization in Head-driven Phrase Structure Grammar. Research
Proposal to the National Science Foundation, Ohio State University.
Pollard, Carl and Sag, Ivan A. (1994), Head-Driven Phrase Structure Grammar. Chicago:
University of Chicago Press.
Przepiorkowski, Adam (1999a), 'On complements and Adjuncts in Polish', R. D.
Borsley and A. Przepiorkowski (eds) Slavic in Head-Driven Phrase Structure
Grammar. Stanford: CSLI Publications, pp. 183-210.
Przepiorkowski, Adam (1999b), 'On case assignment and "adjuncts as complements"',
in Gert Webelhuth, Jean-Pierre Koenig and Andreas Kathol (eds), Lexical and
Constructional Aspects of Linguistic Explanation. Stanford: CSLI Publications, pp.
231-45.
Reape, Michael (1994), 'Domain union and word order variation in German', in John
Nerbonne, Klaus Netter and Carl J. Pollard, (eds), German in Head-Driven Phrase
Structure Grammar. Stanford: CSLI Publications, pp. 151-98.
Rizzi, Luigi (1997), 'On the fine structure of the left-periphery', in Liliane Haegeman
WG SURFACE STRUCTURES AND HPSG ORDER DOMAINS 167
Notes
* I would like to thank Bob Borsley and Kensei Sugayama for their helpful
comments. Any errors are those of the author.
1 For comparison of WG with an earlier version of HPSG (Pollard and Sag 1994),
see Hudson (1995).
2 In the current version of WG (Hudson 2005), the No-Tangling Principle has been
replaced with ORDER CONCORD, whose effects are essentially the same as for its
predecessor. In this chapter we will refer to the No-Tangling Principle.
3 The examples in the rest of this section are cited from Haegeman (2000) unless
otherwise indicated.
4 (26b) was provided for me by Bob Borsley (p. c. )
5 (28b) is not acceptable to some speakers (Borsley 2004).
6 The data in (28) could be accommodated in WG if we assumed a dependency
relation between the complementizer and the extractee (Borsley (p. c. ); and
Sugayama (p. c. )). Needless to say, however, an argument along these lines would
need to clarify the nature of this apparently ad hoc grammatical relation.
7 This amalgamation of the SLASH values is due to the SLASH-Amalgamation
Constraint (Ginzburg and Sag 2000: 169):
(i)
8 See Kathol (2002) for an alternative analysis for English clausal domains.
9 In constituency-based grammars such as HPSG, these two rule-types are LINEAR
PRECEDENCE RULES and IMMEDIATE DOMINANCE RULES.
This page intentionally left blank
Part II
Abstract
Heads of phrases are standardly diagnosed by both structural and distributional
criteria. This chapter argues that these criteria often conflict and that the notion
'head of a phrase' is in fact a conflation of two wholly distinct notions, 'structural
head' (SH) and 'distributional head' (DH). The SH is the root of the phrase and
is diagnosed by structural criteria (mainly, word order and ellipsis),. Additionally,
the distribution of the phrase may be conditioned by one or more words in the
phrase: these are DHs. The SH is often a DH, but there are many English
constructions in which a DH is not the SH and is instead a word subordinated
within the phrase. The chapter discusses a variety of these constructions,
including: that-dauses; pied-piping; degree words; attributive adjectives; determi-
ners; just, only, even; not, almost, never, all but; the type-of construction;
coordination; correlatives; adjuncts; subjects; empty categories.
1. Introduction
The central contention of this chapter is that a number of constructions in
English oblige us to recognize that the distribution of a phrase may be
determined by a word subordinated within the phrase, rather than by, as
commonly taken for granted, the structural head of a phrase - i. e. the highest
lexical node in the phrase. By a phrase's 'distribution' is meant the range of
environments - positions - in which it can occur. An example of such a
construction is pied-piping (discussed in section 8), as in (la). The root of the
phrase in the midst of which throng of admirers is in, but it is by virtue of
containing which that it occupies its position before the inverted auxiliary, for
(la) alternates with (Ib), but not with (2a-2b).
(1) a. In the midst of which throng of admirers was she finally located?
b. Which throng of admirers was she finally located in the midst of?
(2) a. *In the midst of this throng of admirers was she finally located,
b. *This throng of admirers was she finally located in the midst of.
The head of a phrase is normally understood to be denned, and hence
diagnosed, by both structural and distributional criteria. But the notion 'head of a
phrase' is in fact a conflation of two wholly distinct notions: the distributional, or
'external', head, and the structural, or 'internal', head. These two types of head are
172 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
explained in sections 2-3. Although I use the term 'phrase' in a mostly theory-
neutral way, it is important to realize that it doesn't entail the Phrase Structure
Grammar notion that phrases are nonlexical nodes. In Word Grammar (WG),
which is the grammatical model that serves as a framework for the discussion of
grammatical analysis in this chapter, all nodes are lexical, and WG defines a
phrase as a word plus all the words that are subordinate to it1 (A word's
'subordinates' are its 'descendants' in the syntactic tree, the nodes below it; its
'superordinates' are its 'ancestors', the nodes above it) The words in a sentence
comprise all the nodes of a tree, and every subtree of the sentence tree is a phrase.
2. Structural Heads
A phrase's structural head (henceforth 'SH') is, as stated above, to be defined
as the highest lexical node in the phrase. In a model such as Word Grammar,
in which all nodes are lexical, the SH is, therefore, the root of the phrase's
tree structure. For determining which word is the root of a phrase, the
principal diagnostic is word order. Take the phrase eat chocolate: if chocolate is
the root then there cannot be a dependency between eat and a word that
follows chocolate; and if eat is the root then there cannot be a dependency
between chocolate and a word that precedes eat. The test shows that eat is the
root of eat chocolate:
(3) a. *Do Belgian eat chocolate. ['Do eat Belgian chocolate. ']
b. *Do your eat chocolate. ['Do eat your chocolate. ']
(4) a. Eat chocolate today,
b. Don't eat chocolate.
These restrictions follow from the general (and probably inviolable) principle of
grammar that requires phrases to be continuous: no parts of a phrase can be
separated from one another by an element that is not itself contained within the
phrase. The principle is discussed further in section 15. Diagrammatically, the
principle can conveniently be captured as a prohibition against a branch in the
syntactic tree structure crossing another, as in (5).
(5)
3. Distributional Heads
The distribution of a word (or phrase) is the range of grammatical environments
it can occur in. In the broadest sense, this includes a word's co-occurrence with
both its dependents, e. g. the fact that eat can occur with an object (eat chocolate),
and its regent, e. g. the fact that eat can be complement of an auxiliary (will eat).
(The term 'regent' is used in this chapter as the converse of 'dependent'. ) But in
the narrower and more usual sense employed here, a word's distribution
essentially concerns what it can be a dependent of. 'Distribution' in the latter
sense contrasts with 'Valency' (or 'Selection'), which concerns what a word can
STRUCTURAL AND DISTRIBUTIONAL HEADS 173
be regent of. As a first approximation, we can therefore say that the distribution
of X is the product of rules that such and such a regent permits or requires a
dependent belonging to a category that X belongs to. But the topic of this
chapter is such that instead of that first approximation we need, at least
pretheoretically, to formulate this in terms of the notion 'distributional head':
when a word permits or requires a dependent of category X, it permits or
requires a dependent that is a phrase whose distributional head is of category X.
Models of syntax have generally held that something is a distributional head
(henceforth 'DH') if and only if it is a SH - in other words, that a phrase has
just one sort of head, and that this single head determines both the structural
and the distributional properties of the phrase. But my first aim is to show that
the two sorts of head must be distinguished. Normally the two sorts of head
coincide, so that one word is both SH and DH of a phrase - i. e. that the root of
a phrase determines its distribution. This is generally known as 'endocentricity'.
But a fair number of constructions in English suggest that that norm cannot be
exceptionless. (And, as we will see later, once we acknowledge that the norm
has exceptions, there is reason to question whether it is in fact even much of a
norm at all. ) In these constructions, the SH is not the DH. This is exocentricity.
But the constructions involve a very particular kind of exocentricity: in them,
the DH is subordinate to the SH. That is, the distribution of the phrase is
determined not by the root of the phrase but by a word subordinated more or
less deeply within the phrase. I will call this phenomenon 'hypocentricity', since
the DH is below the SH in the tree.
Although the notion 'structural head', defined as the root of a phrase, has a
role in the formal analysis of hypocentricity, the notion 'distributional head'
does not, and is purely descriptive. This is because it turns out that a phrase
may have many distributional heads. This can be illustrated as follows. Section
11 argues that in 'determiner phrases' (i. e. 'noun phrases' in the traditional
sense), the determiner is SH and the noun is DH. This is illustrated in (6a),
where, as in subsequent examples, small capitals indicate the SH and italics the
DH. And section 8 argues that in pied-piping in wh-relative clauses, the wh-
word is DH, so SH and DH are as indicated in (6b-6c). But in (6c) the locus of
the DH also follows the pattern of (6a), giving (6d), where there is one SH, the,
and two DHs, news and which.
(6) a. [THE news] had just reached us
b. [NEWS of which] had just reached us
c. [THE news of which] had just reached us
d. [THE news of which] had just reached us
In the formal analysis of hypocentricity introduced in section 4 and presented
in full in section 6, a phrase's DHs are defined relationally relative to the SH.
So, although I said in section 1, in framing the discussion of hypocentricity,
that the notion 'head of phrase' is a conflation of two sorts of phrasal head, the
structural and the distributional head, it would be more accurate to say that the
traditional notion 'head of a phrase' remains valid, but that it is defined by
structural criteria, as the phrase root, and, contrary, to what is usually thought,
174 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
4. 77m*-Clauses
In a ^/-clause, the SH is that, which explains why it must be at the extreme left
edge of the clause. But the DH of the thai-clause is the finite complement of
that. The evidence for this is that the clausal complement of certain verbs, such
as require and demand, must be subjunctive.2 So the DH is the subjunctive word;
it is the presence of the subjunctive word that satisfies the selectional
requirements of require I demand:
(7) a. I require [THAT she be/*is here on time].
b. I demand [(THAT) she give/*gives an immediate apology].
A satisfactory analysis of this phenomenon is provided in Rosta (1994, 1997)
(from whose terminology I deviate in this chapter without further comment).
That is defined (in its lexical entry) as 'surrogate' of its complement, the finite
word. As a general rule, every word is also surrogate of itself; so the finite word
is surrogate of itself. Require /demand select for a complement that is surrogate of
a finite word. Since the surrogates of a finite word are itself and that (if it is
complement of that), the selectional requirements of require/demand can be
satisfied by that or by a finite word. Surrogacy accounts for some hypocentric
constructions, but not all. We return to this point in section 6.
5. Extent Operators
I adopt 'extent operator' as an ad hoc term to cover such items as all but, more
than, other than, almost, not, never, which do not necessarily form a natural
grammatical class but do at least have certain shared properties that warrant
their being discussed together here. Reasons will be given why, when an extent
operator modifies a predicative word, as in (8a-8f), or a number word, as in
(9a-9f), the extent operator appears to be SH and the number or predicative
word to be DH.
(8) a. She had [ALL but expired}.
b. My rent has [MORE than doubled].
c. She was [OTHER than proud of herself].
STRUCTURAL AND DISTRIBUTIONAL HEADS 175
The second kind of evidence for the identification of the SH comes from
certain of the extent operators' more core variants that have a nominal
complement, notably all but JVP, more than JVP and over/under JVP. It is easy to
demonstrate that in these variants, all, more and over /under is both SH and DH.
For instance, in (12a) we have me rather than / and are rather than am because
the subject is all rather than me/L4'
(12) a. All but me/*I are/*am to come.
b. More than me/*I are/*am here.
c. Over/Under us/*we seems/*seem unsuitable for the storage of
radioactive waste.
The third and most telling sort of evidence comes from ellipsis. Ellipsis
involves the deletion of the phonological content of some syntactic structure,
and it seems to operate rather as if (the phonology of) a branch of the syntactic
tree were snipped off. Thus if the phonological content of one node is deleted,
then so must be the phonological content of all nodes subordinate to it. 5 So, if
we have established that a branch links two nodes, X and Y, and X's phonology
remains when Y's is deleted, it must follow that Y is subordinate to X. And we
find that with certain extent operators, including not, nonstandard never
(meaning 'not' rather than 'nowhen') and almost, their phonology can remain
when the phonology of the DH is deleted. (Words whose phonology is deleted
are shown in subscript. )
(13) a. %I would prefer that you be so not
mde.
b. %I know you want to do it, but try to not
c. Would I do it? I wouldn't not do it
d. We'll make him not
e. I know it's unmanly to flinch, but how can you stand there and not fi^ch?
f. You can't go out without knickers - not go out without knickers and still
stay decent
g. %She never stoie your cigarette lighter- ['She didn't']
h. %Did she do it? No, but she almost
(13g) is dialectal, and I have also marked (13a, b, h) as subject to variant
judgements, because some speakers reject them, but all of (13a-13h) are
acceptable for some speakers, and that is what matters here. The conclusion is
that the deleted DH is subordinate to the extent operator, which is therefore
the SH.6
We have established, then, that the internal structure of these phrases is as
shown in (14a-14b). This raises two questions. The first, which is addressed in
section 6, concerns the structure of (15a-15b): how can structures (14a-14b) be
reconciled with the fact that it is perished that satisfies the selectional
requirements of had?
almost perished
STRUCTURAL AND DISTRIBUTIONAL HEADS 177
b.
* I know almost she perished. ['She almost perished. ']
It seems, then, that (16a-16b) must involve something along the lines of
obligatory 'leftwards-extraposition' of the subject; the subject moves from its
ordinary position and ends up as a subordinate of the extent operator, as
diagrammed in (19a-19b).7 We return to this matter in section 17. 2, where a
far more satisfactory solution is provided.
SUBJECT
SUBJECT
But as (21a-21c) show, the complement of clausal that can be a finite word, or
an extent operator, but not another that. The same pattern holds for
complements of extent operators, (22a-22f). (Structurally, almost in (21c) and
(22d) occurs in the position where the complement of but and that is expected.
For this reason, I conclude that almost (rather than perished] is indeed the
complement of but and that. And, as pointed out in section 5, almost is the
semantic head in almost perished: it means 'an event near to being an event of
perishing'. ) Hence it cannot be the case that the selectional requirements of that
or of extent operators are such that any surrogate of a finite word will satisfy
them.
(20-22) show that there are two types of hypocentric phrase. In one type, the
SH is surrogate of the DH, and the SH can be that, an extent operator, or the
DH, the finite word. In the other type, the SH can be an extent operator or the
DH, but not that. To capture this pattern, which, as we will see in later sections,
generalizes across many diverse constructions, we need to posit a subtype of of
the Surrogate relation, which I will call 'Proxy'. So, whereas require/demand
select for a complement that is surrogate of a subjunctive word, that selects for a
complement that is proxy of a finite word. Likewise for extent operators: almost
and but (in all but} select for a complement that is proxy of a finite word (or of
whatever other sorts of word extent operators can modify).
In general, the format for selectional rules will be not (23a), but rather (23b-
23c). Rules of type (23a) seem to be surprisingly scarce: I am currently aware of
only one instance, which is discussed in section 16 (in examples (73a-73c)).
STRUCTURAL AND DISTRIBUTIONAL HEADS 179
(26a) fails to account for why the subjunct must be at the extreme edge of the
focused phrase. With the subjunct as SH, as in (25a), the ungrammaticality of
(27a) is predicted by the crossing branches. But no such prediction is made
with structure (26a), as in (27b):
8. Pied-piping
The roots of the bracketed phrases in (33a-33b) are under and on. But the
phrases occupy their position before the inverted auxiliary by virtue of
containing a negative element, no, or an interrogative wh-word, which. The root
of the bracketed phrase in (33c) is on or should, depending on which is
subordinate to which in one's analysis, but the DH is which: the rule for finite
wh-relative clauses is that they contain a subject or topic phrase that contains a
relative wh-word. In contrast to certain other hypocentric constructions, the
semantic head of the phrase appears to be the DH, the wh-word, since
semantically it is the relationship of binding or equation that connects the
relative clause to the modificand.
(33) a. [UNDER no circumstances] would she consent.
b. [ON the corner of which streets] should we meet?
c. streets [on the corner of which we should meet]
As with the inside-out interrogative construction discussed in section 13, the SH
in pied-piping is surrogate of the DH, and - in relative clause pied-piping at
least - the surrogate relation is long-distance, i. e. there is no upper limit to the
number of nodes that can be on the path up the tree from DH to SH.
9. Degree Words
In the phrases bracketed in (34a-34d), the degree words (too, more, as) modify
the adjectives that are the distributional head. If the DH were also the SH, we
would expect complements of the degree word to appear on the same side of
the adjective as the degree word does, as in the ungrammatical (35a-35d), in
order to avoid the sort of crossing branches diagrammed in (36).
A standard solution to this problem would have the adjective as SH and the
complement of the degree word obligatorily extraposed. But as far as I am
aware, this solution is motivated solely by the lack of any mechanism for
182 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
To rule out (39), the extrapositional analysis would have to posit that than him
first extraposes to become some sort of postdependent of sophisticated, in line
with the extraposition rule that applies to degree words, and then extraposes
again to become a postdependent of the noun, in line with the extraposition
rule that applies to attributive adjectives.
(39) *a more sophisticated than him person
The suggested analysis is that attributive adjectives are not adjuncts. Rather,
they take the proxy of a noun as their complement, and the adjective is proxy of
its complement. Further evidence for this analysis comes from ellipsis:
(40) She chose an easy puzzle and he chose a difficult puzzle
this booklthat book fails to capture the analogy with the adjective phrases this big I
that big. But section 9 has argued that in this big/that £fg, the SH is this /that, and
big is DH, so the analogy is captured.
There are other examples that can be used to make the point made by (41a-
41b), but none that are quite so convincing. The verb crane appears to require
X's neck as its object, but it's hard to prove that this is not merely the consequence
of the verb's meaning, which specifies that the cranee is the neck of the craner. A
somewhat more convincing example is (46): wreak for many speakers requires an
object whose DH is havoc, and no synonym will suffice in its stead.
(46) The storm will wreak the usual havoc/%devastation.
The same appears to hold for cognate objects:
(47) a. She smiled her usual smile/*grin.
b. She slept a deep and placid sleep/* slumber/* somnolence/*kip.
But the contrast in (48a-48c) shows havoc and cognate objects to be unlike way
adverbials, and makes it hard to maintain that the presence of havoc and smile in
(48b-48c) is a syntactic requirement. 8
(48) a. *She did it something that fell short of a wholly sensible way.
b. The storm will wreak something that falls short of outright havoc.
c. She smiled something that fell short of the sweet smile we had come to
expect from her.
The relationship between determiner and noun is analogous to that between
clausal that and finite word. Just as multiple that is ungrammatical (See that
(*that) she does), so are multiple determiners: the (*the) book, or, more
plausibly, *a my book, fa book of mine'. Clausal that takes the proxy of a finite
word as its complement, and is surrogate of its complement Likewise, the
determiner takes the proxy of a common noun as its complement, and is
surrogate of its complement
This analysis predicts that (49) should be ungrammatical. I am not sure that
that prediction is correct, though.
(49) ?She did it the opposite of a sensible way.
(62) a. She always chooses [nobody CAN ever guess which item from the menu].
f. The drug makes you [you CAN never be sure how virile].
(64) a. The teacher will give every child in the class a gold star.
b. *The teacher will every child in the class give a gold star.
c. *The teacher will give a gold star every child in the class.
d. Every child in the class was given a gold star.
e. Also given a gold star were all the children in the class.
But the < e > analysis allows us to do away with the distinction between branch
and nonbranch dependencies: with the sole exception of Binder, all
dependencies are branches. The syntactic structure of a sentence is just a
tree with labelled branches, supplemented by nonbranch relations of types
Binder, Surrogate and Proxy. (Even the branch labels are potentially
redundant, given that a branch is distinguished from its siblings by its position. )
Thus, the whole apparatus of syntactic structure can be significantly simplified,
for the price of merely one extra lexical item among thousands.
On the assumption that unbound <e> is interpreted as 'something/
someone', we are then in a position to posit structures for (65a-65d)14 that yield
the meaning that the sentences actually have. Furthermore, the presence of
< e > in (65c-65d) provides a way to capture the fact that even though there is
no overt or deleted object of keep or subject of alive, semantically the object of
keep is still understood to be the subject of alive. ((65c) is the structure one
would have if unbound < e > is added to otherwise orthodox WG. (65d) is the
structure I am proposing. )
STRUCTURAL AND DISTRIBUTIONAL HEADS 189
d. ... but need'st not strive officiously to keep <e> <e> alive.
BINDER
The main snag with < e > has to do with the phenomenon of connectivity,
whereby traces have to have the categorial properties of their binder, i. e. of what
they're traces of. An adjective leaves an adjectival trace, a noun leaves a nominal
trace, and so forth. This is so that the trace can satisfy the categorial selectional
requirements imposed on the position the trace occupies. For example, the
subject-raising verb wax requires an adjectival complement. So if <e> is
Section 11 explains why which is surrogate of items/place I body/way. Section 9
wroth.
If we had to introduce invisible words of every conceivable word class, and add
rules requiring them to agree in word class with their binder, then this would be
very much the opposite of a simplification to the grammar. But the Proxy
relation provides a simple solution, if <e> is proxy of its binder. The
selectional requirements of wax are that it takes a complement that is surrogate
of an adjective, and this requirement is satisfied in (66), since (i) how is binder of
<e> and hence <e> is proxy of how; (ii) being a degree word, how is
surrogate of wroth; and (iii) <e> is therefore surrogate of wroth.
In all the other cases discussed in this chapter where the Proxy relation is to
be found, it occurs in a hypocentric phrase, where the SH is proxy of the DH,
which is subordinate to the SH. But this clearly does not apply to the proxy
relation holding between <e> and its binder. The conclusion to be drawn
from this is that rather than the Surrogate and Proxy relations being merely
convenient ways to formalize hypocentricity, they are in fact fundamental, and
hypocentricity is merely a convenient label for phrases whose root is surrogate
of one of its subordinates.
15. Coordination
Since its beginnings, WG has analyzed coordination as exceptional to major
and otherwise exceptionless principles. The first exception is that whereas the
rest of syntax consists solely of dependencies between lexical nodes, i. e.
190 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
(67) Give {< [Sophy] [roses] > and < [Edgar] [tulips] >}.
(69) * Give students tulips of linguistics. ['Give students of linguistics tulips. ']
16. Correlatives
Another instance of hypocentriciry in coordination arises with correlatives (both,
either, neither). The correlative's position at the extreme edge of the phrase
follows if it is SH. The conjunction is complement of the correlative, and the
correlative is proxy of the conjunction.
Correlatives are one of the very rare instances, mentioned in section 6, of words
whose complement is their complement target rather than a proxy of their
target. This can be seen from the ungrammaticality of (73a) in contrast to (73b-
73c). In (73a), the complement of both is or, which is proxy of each and: this is
ungrammatical, because the complement of both must be and.
in the sentence tree. But in this section I will argue that branches are
unlabelled, and that there is no distinction between branches and
dependencies; and so-called 'dependency types' are in fact lexical items in
their own right. Thus, instead of X being subject of Y, there is a word, an
instance of the lexical item < PREDICATION >, that has two dependents, X (the
subject) and Y (the predicate). These words that take over the job of
grammatical relations ('GRs'), I will call 'GR-words'. GR-words belong to a class
of function word characterized, in part, by phonologylessness.
This proposals is relevant to this chapter for two reasons. First, the phrases
whose root is a GR-word are strongly or weakly hypocentric. And second, many
of the other analyses made elsewhere in the chapter converge on the GR-word
analysis as a more or less inescapable conclusion.
17. 1 Adjuncts
Semantically, adjuncts are the converse of complements, in that whereas X is a
semantic argument of Y when X is a complement of Y, Y is a semantic
argument of X when X is an adjunct of Y. For instance, in She snoozed during the
interval, the snoozing is the theme argument of 'during'. A natural logical
corollary of the fact that the modificand is an argument of the modifier is that
modification is recursive: after one modifier has been added to the modificand,
another modifier (of the same type or another type) can always be added. This
is, of course, because a predicate's argument 'place' ('attribute') can have only
one 'filler' ('value'), but the filler of one argument place can also be the filler of
many others. We would therefore predict that recursibility is a default property
of adjunction. One cannot rule out, a priori, the possibility of special rules
prohibiting adjunct recursion in certain constructions, but it is hard to imagine
what systemic or functional motivation there could be for such a prohibition. So
the null hypothesis is therefore that all adjuncts are recursible. 17
If adjuncts were simply dependents of the word they modify, then the
principle of Branch Uniqueness ought to make them irrecursible. I propose
instead that Adjunct is not a dependency but rather a GR-word. Rather than X
being adjunct of Y, X and Y are dependents of an Adjunction GR-word; X is
the modifier dependent and Y is the modificand dependent. Adjunction is a
word class; the words it contains are the different kinds of adjuncts, such as
< manner-adverbial >, < depictive >, and so forth. Adjunction phrases are
hypocentric: the adjunction (the SH) is proxy of first dependent, the
modificand (the DH). This can be seen from (74), where it is dozed that
satisfies the requirement of had for a past participle as its complement target:
the relationship between her going to bed and her being agitated is that the
former occurs during the latter.
A further merit of adjunctions is that they explain what Hudson (1990) calls
'semantic phrasing'. For example, (76a-76b) are not synonymous. (76a) says
that what happens provocatively is her undressing slowly, while (76b) says that
what happens slowly is her undressing provocatively. This nuance of meaning is
reflected directly in the structure, (77a-77b).
b. work clothes bag ['work bag for clothes', 'clothes bag for work']
But then we find that this analysis falls foul of word order evidence. The
structure given to (82a) fails to rule out (82b):
(82) a.
revenge kitchen implement attack ['revenge attack with kitchen implement']
b.
(83) a.
b.
* <n+n> <n+n> <n+n> kitchen revenge implement attack
['revenge attack with kitchen implement']
17. 2 Subjects
Conjoined predicates, as in (88a), present a problem. If the structure is as in
(88b), then No Crossing is violated. If the structure is as in (88c) or (88d), then
some kind of rule of leftwards extraposition of subjects is required:
As we saw in section 5, exactly the same problem arises with extent operators:
But under the GR-word analysis, the problem evaporates. The GR-word
<predication > has two dependents, 18 the first corresponding to the subject and
the second corresponding to the predicate19:
196 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
That <predicatiori> is not proxy of its second dependent can be seen from the
fact that one <predicatiori> cannot be second dependent of another, i. e. that
multiple subjects cannot occur.
As it stands, the analysis makes it look coincidental that it is only the second
('predicate') dependent of < predication > and not the first ('subject') that has
DH-like properties. Therefore the grammar should perhaps formally accord
the second dependent in this construction a special status. Let us therefore call
< predication > the 'guardian' of its second dependent, the metaphor being that
the second dependent is a legal minor and its intercourse with its
superordinates must always be mediated by its guardian. And let us add rules
((93a-93b); (93b) replaces (24c)):
(93) a. If X is surrogate of Y, then X is guardian of Y.
b. If X is guardian of Y, and Y is guardian of Z, then X is guardian of Z.
In this case, the complement of auxiliary have must be a <predication > that is
guardian of a past participle, and the complement of wax must be a
<predication > that is surrogate of an adjective.
Like Subject and Predicate, both the topic and the 'comment' phrases can be a
coordination.
(95) a. White chocolate, she keeps on giving me and I can't help gorging
myself on.
b. Both white chocolate and Cheshire cheese, I can't help gorging myself
on.
As with <predication >, these facts motivate a GR-word for the topic-comment
structure, its first dependent being the topic and its second the comment. The
second dependent of this GR-word is a <predication > that is guardian of a
verb or auxiliary.
< Topic—comment > can therefore be equated with finiteness: a finite clause is
one that contains < Topic-comment >, which we could equally well call
<finite >.
I leave for future investigation issues about the relationship between
<finite> and mood and tense, about verbal inflection, and about whether
mood exists as a grammatical category in English.20 At any rate, it is clear that
<finite> phrases are at most weakly hypocentric. If Indicativity and
Subjunctivity are subtypes of <finite >, then <finite> phrases are not
hypocentric at all, since know can select for a surrogate of < indicative >, require
for a surrogate of < subjunctive > and insist for a surrogate of <finite >. If, on
the other hand, the mood distinctions are located lower down within the
<finite> phrase, then know /require /insist will select for a surrogate of
<Jinite>, but know and require will further stipulate that their complement
must be guardian of wherever the appropriate mood distinction is located.
17. 4 Complements
An inescapable corollary of the proposed analysis of coordination is that
conjuncts are complete phrases.21 In this section I sketch how this must work,
though the sketch is of a solution strategy rather than an analysis worked
through in detail.
198 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
The answer has to be that there is an extra GR-word present, whose function is
like that of an X' node, uniting into a single phrase its two separate dependents,
V and O. This gives the structures in (99a-99c).
(99) a. She will [[<X'> eat cheese] and [<X'> drink wine]],
Similarly, (lOOa) shows that Sophy is a complete phrase in (lOOd), and (lOOb)
shows that Sophy is a complete phrase in (lOOd). But (lOOc) shows also that
Sophy roses is a complete phrase in (lOOd).
(101)a. She will <X > give <ditransitive> Sophy and Edgar roses.
She will <X> give <ditransitive> Sophy roses and <ditransitive> Edgar tulips,
(102a) shows that roses today is aphrase. Today is an adjunct, but not roses:
it is not the roses that occur today but rather their being given. Today must be
an adjunct of a GR-word that marks the second object, as in (102b). (103a)
therefore has structure (103b).
(102) a. She will give Sophy roses today and tulips tomorrow.
b.
She will <X'> give <ditr> Sophy <adj> <GR> roses today and <adj> <GR> tulips tomorrow.
With the exception of < transitive >, the GR-words involved in comple-
mentation would be guardians rather than surrogates or proxies, since the GR-
words are not freely omissible in the way that surrogates and proxies are. As for
the kind of hypocentricity, if any, involved with < X' >, I leave this for future
investigation.
18. Conclusion
This chapter has demonstrated the existence - and indeed the prevalence - of
hypocentricity, the syntactic phenomenon whereby the distribution of a phrase
is determined not by the root of the phrase but by a word subordinate to the
phrase root. Hypocentricity comes in different 'strengths'. In the strongest form
of hypocentricity, a phrase with a given SH is in free variation with a version of
the phrase with the SH absent. These are the hypocentric constructions that
involve the Proxy relation. The instances discussed in this chapter involve (i)
'extent operators' like almost, not and all but, (ii) 'focusing subjuncts' like even;
(iii) attributive adjectives; (iv) the type-of construction; (v) coordinating
conjunctions; (vi) correlatives like both; and (vii) adjunctions, which are the
invisible words that link adjuncts to their modificands. Apart from coordina-
tion, these could all be called 'modifier constructions'. In addition it has been
suggested that invisible bound variables are proxy of their binder, even though,
exceptionally, the binder would not be a subordinate of its proxy.
In hypocentricy of 'intermediate strength', a phrase with a given SH is in
distributional alternation with a version of the phrase with the SH absent, but
the variation is limited to certain environments. These are the hypocentric
constructions that involve the Surrogate relation. The instances discussed in this
chapter involve (i) clausal that; (ii) inside-out interrogative clauses, which behave
like clausal determiners; (iii) pied-piping; (iv) degree words; and (v)
determiners.
In the weakest form of hypocentricity, there is no distributional alternation,
but the DH is nevertheless sensitive to material external to the hypocentric
200 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
phrase. These are the hypocentric constructions that involve the Guardian
relation. The instances discussed in this chapter involve the invisible GR-words
<predication >, which is the root of the subject+predicate construction, and
< finite >, which is the root of the topic+comment construction, and various
other GR-words that form the structural basis of complementation.
The relations Proxy and Surrogate are initially motivated as mechanisms that
provide an analysis for constructions that cannot otherwise be satisfactorily
handled by WG. Once this mechanism is admitted, it opens the way - or the
floodgates - for a series of increasingly radical (and increasingly sketchy and
programmatic) analyses of coordination and of grammatical relations, which
aim to simplify WG by drastically reducing the range of devices from which
syntactic structure is constituted, while still remaining consistent with WG's
basic tenets. The devices that are done away with are (i) exceptions to the
principle of Node Lexicality, i. e. nonlexical phrasal nodes, which orthodox
WG uses for coordination; (ii) exceptions to the No Crossing principle barring
crossing branches in the sentence tree; (iii) dependencies that are not associated
with branches in the sentence tree; and, perhaps, (iv) dependency types tout
court. In their most extreme form, these changes result in a syntactic structure
consisting of nothing but words linked by unlabelled branches forming a tangle-
free tree, supplemented by Binder, Proxy, Surrogate and Guardian relations.
While I believe the necessity for Proxy and Surrogate relations is demonstrated
fairly securely by the earlier sections of the chapter, their extended application
in the analysis of coordination, empty variables and grammatical relations is of a
far more speculative nature. But my aim in discussing these analyses in this
chapter has been to point out how they are possible within a WG model and
why they are potentially desirable.
References
Cormack, Annabel and Breheny, Richard (1994), 'Projections for functional categories'.
UCL Working Papers in Linguistics, 6, 35-62.
Hudson, Richard (1976), 'Conjunction reduction, gapping and right node raising'.
Language, 52, 535-62.
— (1990), English Word Grammar. Oxford: Blackwell.
— (2004), 'Are determiners heads?'. Functions of Language 11, 7-42.
Jaworska, Ewa (1986), 'Prepositional phrases as subjects and objects'. Journal of
Linguistics, 22, 355-74.
Payne, John (1993), 'The headedness of noun phrases: Slaying the nominal hydra', in
Greville G. Corbett, Norman M. Fraser and Scott McGlashan (eds), Heads in
Grammatical Theory. Cambridge: Cambridge University Press, pp. 114-39.
Quirk, Randall, Greenbaum, Sidney, Leech, Geoffrey N. and Svartvik, Jan Lars (1985),
A Comprehensive Grammar of the English Language. Harlow: Longman.
Rosta, Andrew (1994), 'Dependency and grammatical relations'. UCL Working Papers
in Linguistics, 6, 219-58.
— (1997), 'English Syntax and Word Grammar Theory'. (Unpublished doctoral
dissertation, University of London).
Van Langendonck, Willy (1994), 'Determiners as Heads?'. Cognitive Linguistics, 5,
243-59.
STRUCTURAL AND DISTRIBUTIONAL HEADS 201
Notes
1 In standard Phrase Structure Grammar, lexical nodes are terminal and nonlexical
nodes are nonterminal. If a nonterminal node is defined as one that contains others,
then in WG all nodes are terminal. But this terminology is a bit misleading, since in
a WG tree structure, terminal nodes are ones that have no subordinates. Hence it is
more perspicuous to define WG as maintaining that all nodes are lexical.
2 This problem posed by require was pointed out by Payne (1993); cf. also Cormack
and Breheny (1994).
3 See Jaworska (1986) for examples of such nonpredicative prepositions.
4 These judgements are for conservative Standard English. Admittedly there is the
famous line 'The boy stood on the burning deck, whence all but he had fled'
(Felicia Hemans, 'Casabianca'), but that could be a solecism induced by the
register, hypercorrectively, and by the disconcerting unfamiliarity of the word-
sequence him had in contrast to he had. It is also true that for many speakers of
contemporary English, the rules for the incidence of personal pronouns' subjective
forms, especially in less colloquial registers, seem to be pretty much of the make-
them-up-as-you-go-along or when-in-any-doubt-use-the-subjective-form sort.
5 It has to be admitted that this claim flies in the face of a certain amount of evidence
to the contrary, notably determiner-complement ellipsis, as in (i), and
pseudogapping, as in (ii).
(i) Do students of Lit tend to be brighter than those students °f Lang?
(ii) She will do her best to bring the food, as will he a0 his best to bring me wine.
6 (13f) raises its own analytical curiosities, which I won't investigate further here.
Logically, the structure is 'not [[go out without knickers] and [still stay decent]]';
that is, the ellipsis is of one conjunct Egregiously unexpected though such a
phenomenon is, I find that the analogous structure in (i) is, also surprisingly,
acceptable.
(i) Nobody likes to complain, but she should [[COmpiaiJ and [be the happier for
it]].
7 The arrows below the words represent dependencies that don't form branches in
the tree structure. (See section 14 on the eradication of such dependencies. )
8 More generally, I would maintain that open class lexemes are invisible to syntax
and hence that selectional rules cannot refer to them. Only word classes are visible
to syntax and can be involved in selectional rules. WG has always held that
lexemes are word classes, but my contention is that this is true only of closed class
lexemes, in that closed class lexemes are word classes that are associated with
particular phonological forms, whereas open class lexemes are morphological
stems (which is why processes of derivational morphology, which output stems, can
output only open class lexemes). Way (and a few other similar words, such as
place) would be a subclass of Common Noun.
9 More precisely, the rule is that by default, nouns receive a mass interpretation, but
the complement of certain determiners, such as an, must be proxy of a noun that
receives a count interpretation.
10 I don't know why (56c) is grammatical. Cake has a count interpretation, so ought to
be complement target of a determiner, but if it is complement target of these then it
ought to agree in number with these.
11 I suggest that the reason is that common nouns take plural inflection only when
complement of a plural determiner. (This supposes that bare plurals are
complement of a phonologically invisible plural an. } In (55b) there is no plural
determiner present to trigger the plural inflection on cakes.
202 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
(v) know that she is/ was/ *be/ *were/ goes/ *go/ went mad
(vi) She would, *is/ %was/ *?be/ %were/ *goes/ *go/ *went she mad.
(vii) She would, is/ %was/ *?be/ %were/ *goes/ *go/ *went you to.
(viii) I would prefer that you not be/%be/*is/*is.
(ix) She almost *be/*be/is/%is.
21 In determining what sorts of phrase can be coordinated, it is important to factor
out the extraneous but distorting effects of Right Node Raising-type operations,
which delete the phonology of part of one conjunct. See Rosta (1997).
9 Factoring Out the Subject Dependency
NIKOLAS GISBORNE
Abstract
This chapter offers a revision to the English Word Grammar (EWG) model by
factoring out different kinds of dependency. This is because the information
encoded in the EWG model of dependencies is not organized at the appropriate
level of granularity. It is not enough to say, for example, that by default the
referent of a subject is the agent of the event denoted by the verb.
1. Introduction
The English Word Grammar model treats dependencies as asymmetrical
syntactic relations (Hudson 1990: 105-8), where the critical information is the
asymmetry and the relative ordering of head and dependent. Hudson (1990:
120-1) goes on to treat grammatical relations as a particular subclass of
dependency relation, and to identify certain semantic roles as being
prototypically linked to certain grammatical relations. For the English Word
Grammar model, therefore, dependencies are labelled asymmetrical syntactic
relations which are also triples of semantic information, syntactic relation
information and word-order information bound together by default inheritance.
The theory is syntactically minimalist: all syntactic phenomena are analyzed in
terms of dependency relations and the categorization of the words that the
dependencies relate to.
The result is a highly restrictive model of grammar, where all relationships
are strictly local. It differs from other lexicalist frameworks, such as Lexical
Functional Grammar (LFG), in that there are not different domains of structure
which represent different kinds of information. All grammatically relevant
information for WG is read off the lexicon and the dependency information.
Within this theory of grammar, Hudson (1990) uses an inventory of
dependencies which is pretty much what you find making up the set of
grammatical relations in both traditional grammar and classical transformational
grammar.
In this chapter, I offer a revision to the English Word Grammar model by
factoring out different kinds of subject dependency. This is because the
information encoded in the EWG model is not organized at the appropriate level
of granularity. It is not enough to say, for example, that by default the referent of a
subject is the agent of the event denoted by the verb. This is because there are at
FACTORING OUT THE SUBJECT DEPENDENCY 205
least three kinds of subject in English: subjects triggered by finiteness (on the
grounds that English is not a 'pro-drop' language); subjects triggered by
predicative complementation; and 'thematic' or 'lexical' subjects such as the
subjects of gerunds and other inherently predicating expressions. Subjects
triggered by finiteness are not required to be in any kind of semantic relationship
with the event denoted by the verb. Similar observations about the non-
uniformity of the subject relationship are found in McCloskey (1997, 2001).
In this chapter, therefore, I review the inventory of dependencies in Word
Grammar, and establish a more fine-grained account of subjecthood than the
model of Hudson (1990) envisages. I focus on data introduced in Bresnan
(1994). Bresnan (1994) explores locative inversion, shown in (1), and shows
that in inverted sentences like (Ib), the subject properties are split between the
italicized PP and the emboldened NP:
Bresnan's (1994) account explains the split subject properties in terms of the
parallel architecture of LFG where grammatical information is handled in terms
of a-structure, f-structure and c-structure. I show that the revised Word
Grammar account can capture the same kind of data as LFG within a more
parsimonious ontology.
The chapter is organized into 5 sections. In section 2, 1 discuss the different
dimensions, of subjecthood, and explore the different properties that subjects
have been claimed to display since Keenan (1976). In section 3, I lay out the
data that needs to be discussed (drawn from Bresnan 1994), and explain the
problems that this data presents. In section 4, I present the refined view of
subjecthood that this chapter argues for, and show how it accounts for the data.
The final section, section 5, presents the conclusions and some prospects for
future research.
2. Dimensions of Subjecthood
Subject properties have been gathered up in several different places - for
example, Keenan (1976), Keenan and Comrie (1977), and Andrews (1985).
Subjects have been shown to have diverse properties across languages, and it
has been shown that not every subject property is always displayed by all
subjects in a given language. It is this observation that drives Falk's (2004: 1)
claim that 'a truly explanatory theory of subjecthood has yet to be constructed'.
In this section, I itemize and exemplify some of the major features of
subjecthood, which are generally held to apply crosslinguistically, and present
three diagnostics which apply parochially to English. I have relied on the
presentation of these properties in Falk (2004: 2-5), where they are usefully
gathered together. Not all of the subject properties laid out here are directly
relevant to the analysis of the split-subject phenomena found in locative
inversion, but they are relevant to the broader conclusions about subjecthood
that this case study takes us to, and which are laid out in section 5.
206 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
In (2a), the subject is also the agent of the action. In order for the subject not to
have to be the agent, passive voice is available as in (2b). Voice phenomena are
devices for re-arranging the arguments of the verb so that the agent no longer
has to be presented as the subject Of course, it is not always the case that
subjects are agents, because there are verbs that do not have agentive subjects,
as in (3), but many linguists follow Jackendoff (1990) in assuming a hierarchy of
semantic roles, where the most agent-like is always the one which links to the
subject.
From this, we can say that the glass in (3b) is shown to be the subject by the
diagnostics in (4).
The addressee of an imperative is a subject In the following examples, the
addressee has the status of the subject, irrespective of its semantic role.
(5) a. Go away!
b. Be miserable, see if I care!
From the imperative examples, we can see that it is also possible for subjects to
be covert.2
One widely noted diagnostic is to do with anaphora. There is a subject-
object asymmetry, which becomes evident when subject and object are co-
referential. In the case of co-reference, it is the object which is expressed as a
(reflexive)3 pronoun. This is shown in (6):
In (7a), the doctor is shared between persuaded and to see Peter, because to is the
xcomp of persuade. The passivization facts in (7b) show us that the doctor in (7a)
is the subject of to (and see). The passivization facts in (7c) show us that the doctor
is an argument that is shared with persuaded, because it is also the object of
persuaded. The ungrammatical (7d) shows that Peter cannot be the object of
persuaded and of see at the same time. Therefore, the property of being sharable
with a higher predicate is a property of subjects, not other arguments.
The relationship between (8a) and (8b) shows that in (8b), Jane is the subject of
both seems and to like. The example in (8c) shows that the passive subject of (to
be) liked can also be shared with seems. The ungrammatical (8d) shows that it is
not possible to exploit the object of like as the shared subject of seems.
Falk (2004) claims that the subject is the only argument which can be shared
in coordination. The examples in (9) show that a subject can be shared by two
conjoined verbs, but not an object:
However, this observation is not quite right. In Right Node Raising, the object
of the second conjunct can be shared, as in Jane kissed, and Cassandra hugged,
Peter. Right Node Raising needs to be treated as a special construction type
because, among other things, it comes with particular intonation - indicated
here by the commas - which is not a necessary part of the argument sharing in
(9a). But, it is also the case that it is possible to say Cassandra peeled and ate a
grape. Here, both the object and the subject are shared by the conjoined verbs.
There are two remaining properties of subjects which can be stated very
generally. The first is that in many languages the subject is obligatory, as it is in
208 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
English (except in the case of imperatives). This observation gives rise to the
Projection Principle of Chomsky (1981), and its later incarnation as the
Extended Projection Principle (EPP). The second fact is that subjects are
usually discourse topics.
In the next section, I identify some subject properties that are found
parochially in English.
The example in (lla) gives the basic declarative sentence, (lib) shows that it is
impossible to extract a subject after that, even though (lie) shows that it is
possible to extract a subject out of a finite complement clause when there is no
that, and (lid) shows that it is possible to extract other arguments out of a finite
complement clause, like the object.
FACTORING OUT THE SUBJECT DEPENDENCY 209
2. 3 Subject-verb agreement
Subject-verb agreement is not universally found as a subject property; however,
it is not a parochial property of English either. As a phenomenon, agreement is
complex - some languages have agreement that works across a range of
dimensions, whereas English only shows agreement in terms of number
(Hudson 1999). Although English has subject-verb agreement, which is very
common in Semitic, Bantu and Indo-European languages, some other
languages have no agreement morphology at all - for example the modern
Scandinavian languages and the Sinitic languages. English subject-verb
agreement is shown in (12).5
The examples in (12a-12b) show that the number feature of the finite verb co-
varies with the number of the subject. If the subject is plural, so is the verb: girls
triggers like. The example in (12c) shows that a plural subject requires a plural
verb, and the example in (12d) shows that English does not have agreement
with objects, so a plural object cannot rescue a plural verb that has a singular
subject. The agreement phenomena of English are significant in the discussion
of locative inversion that follows.
This, then, completes the review of subject properties. A number of these
properties were exploited by Bresnan in her (1994) article, which is discussed
in the next section, but before we turn to section 3, I shall just summarize the
subject properties in three bullet-point lists here:
• English subjects resist extraction. (In other languages, subjects are often
more extractable than other arguments).
Agreement
• subject-verb agreement: subjects agree with their verb.
The issue, at least in as much as the locative inversion data constitute a problem for
a story of subjecthood, is to do with which of these subject properties belong
together. In the next section, I look at the locative inversion data presented in
Bresnan (1994), and then in section 4, 1 look at these subject properties in the light
of Bresnan's findings about the arguments in the locative inversion construction.
As Bresnan (1994: 75) puts it, 'locative inversion involves the preposing of a
locative PP and the postposing of the subject NP after the verb. The positions
of the locative and subject arguments are inverted without changing the
semantic role structure of the verb. ' Bresnan (1994: 75-6) sets out the limits of
locative inversion, excluding other kinds of inversion around be from the
discussion, and limiting the phenomenon to examples like those in (15).
The examples in (15) can be included in the set of locative inversion data
because the inverted VPs involve a locative PP, and the verbs GOME and LIE
number among the verbs which support locative inversion. Bresnan goes on to
demonstrate that the verbs which allow locative inversion are unaccusative -
thus the grammaticality difference between (16a) and (16b) - or passive (but
with the BY phrase suppressed), as in (17):
From these data, Bresnan concludes that locative inversion 'can occur just in
case the subject can be interpreted as the argument of which the location,
change of location or direction expressed by the locative argument is
predicated' (1994: 80) - to put this another way, the subject must be a 'theme'
in the terms of Jackendoff (1990). This is consistent with Bresnan's account of
unaccusativity, where it is claimed that unaccusativity is not a syntactic
phenomenon, but one where the unaccusative subject's referent is always the
theme of the sense of the verb.
The final aspect of the grammar of locative inversion is that the locative PP is
always an argument of the verb, not an adjunct. The argument/adjunct
distinction is hard to draw in the case of locative expressions, and I do not want
to get bogged down in the debate, but the evidence that Bresnan (1994: 82-3)
brings to bear on the issue is compelling enough. She shows that adjuncts can
be preposed before the subjects of questions, although arguments cannot, and
she uses the so-anaphora test to show that adjuncts can be excluded from the
interpretation of so-anaphora, whereas locative arguments cannot.
To summarize, locative inversion:
There are other facts that apply in the treatment of locative inversion. These
are:
• presentational focus;
• sentential negation;
• other subject properties.
Presentation focus is not strictly syntactic. I shall not return to this. Sentential
negation is more important. Bresnan gives the examples in (18). The
significance is that in (18a), sentential negation is not possible, whereas in (18b),
constituent negation of the postverbal NP is possible:
Bresnan (1994: 88) quotes Aissen (1975: 9) as saying that this restriction is due
to the way in which the locative expression sets a backdrop for a scene.
Negating the main clause undermines this discourse function, whereas
contrastive negation on the postverbal NP does not have such an effect
Bresnan (1994: 88) on the other hand, contrasting English with Chichewa,
argues that sentential negation in Chichewa excludes the subject, so the
restriction comes down to a statement about the scope of negation.
212 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Agreement
In the case of agreement, we see that the locative PP does not agree with the
finite verb.
(20) a. On the corner stood a woman fcp who was standing near another woman],
b. On the corner stood a woman [0 standing near another woman] cp.
Note that the locative PP cannot control the participle in the participial relative.
(21) a. She stood on the corner [cp on which was standing another woman] cp.
b. *She stood on the corner [0 standing another woman].
Subject-raising
However, English does allow apparent subject-raising of locative PPs as in (22).
Bresnan (1994: 96) observes that only subjects can be raised in English. She
compares die two examples in (23) as evidence of this:
phrase which is the subject of a predicate like seem is also the subject of the
xcomp of seem.
Tag questions
The argument from tag questions is a negative one: the claim is that the NP
theme cannot be the subject by this diagnostic. In English tag-questions, a
declarative clause expressing a statement is followed by an auxiliary verb and a
pronoun which expresses a questioning of the prepositional content of the
main clause. Examples are given in (24). The pronoun must agree with the
subject of the main clause.
As Bresnan points out, tags are in general unacceptable with locative inversion.
The examples in (25) show this:
The example in (25a) is less unacceptable than that in (25b). Bresnan (1994:
97) quotes Bowers (1976: 237) who gives the example in (26) and argues
that this shows that the postposed NP in locative inversions cannot be the
subject.
The claim is that in (26) there is coreferential with [i]n the garden, which
indicates, if anything, that in the garden is a more likely candidate for subject
status than the postposed NP. 8 Bresnan also quotes *A man arrived didn't onej
he? - an example from Gueron (1980: 661) to show that tags are in general
difficult to establish with locatives even when they do not involve inversion.
However, this example does seem to be set up to make the situation more,
rather than less, problematic: replace a man by the man, and the problem of
pronoun choice vanishes. With appropriate context, as in the train arrived at 3,
didn't it? a tag question is fine with locative inversion.
The tag-question data are difficult to interpret, therefore; it seems that the
best solution is to put them on one side as inconclusive.
b. *It's in these villages that we all believe that can be found the best
examples of this cuisine.
Nonsubject constituents are unaffected by this restriction, as we can see by
comparing extraction of the uninverted locatives:
(28) a. It's in these villages that we all believe the finest examples of this cuisine can
be found .
b. It's in these villages that we all believe that the finest examples of this cuisine
can be found .
Only subjects show the effect.
(29) a. It's this cuisine that we all believe can be found in these villages.
b. *It's this cuisine that we all believe that can be found in these villages.
(30) 'subject gaps at the top level of one coordinate constituent cannot occur with, any
other kind of gap in the other coordinate constituent. '
(31) a. She's someone that loves cooking and hates jogging.
b. She's someone that cooking amuses and jogging bores .
(3la) has two subject gaps; (31b) has two non-subject gaps.
(33) a. That's the old graveyard, in which is buried a pirate and is likely to be
buried a treasure, [subject-subject]
b. That's the old graveyard, in which workers are digging and a treasure is
likely to be buried . [nonsubject-nonsubject]
(34) a. PPThat's the old graveyard, in which workers are digging and is likely
to be buried a treasure, [nonsubject-subject]
b. That's the old graveyard, in which workers are digging and they say is
buried a treasure, [nonsubject-embedded subject]
The crucial point here is that the examples in (33) show that 'the inverted
locative PPs show the extraction patterning of subjects' (Bresnan 1994: 98). As
Bresnan points out, (34a) is fine with there in the subject gap 'which channels
the extraction to a nonsubject argument (the oblique)', which argues in favour
of a subject treatment of the locative PP.
FACTORING OUT THE SUBJECT DEPENDENCY 215
As these examples show, the locative PPs are not able to appear as the subjects
of auxiliary do in closed interrogatives. Moreover, as we can see in (36), they
cannot occur as subjects in open interrogatives, either:
The examples in (36) show that the locative PP cannot appear as the subject in
(36a-36b) although the theme NP can in (36c-36d) in non-inverted examples.
However, Bresnan (1994: 102), in a section arguing against a null expletive
subject analysis, provides the following data, which make the situation reported
here more complex:
The examples in (37) show that when a subject itself is questioned, as in (37a),
which is the interrogative correlate of a portrait of the artist hung on the wall,
subject-inversion is not triggered. In fact, as (37b) shows, auxiliaries cannot
occur. We see the same facts with locatives:
The examples in (38) correlate to on the wall hung a portrait of the artist. In these
examples, on which wall behaves just like a subject in a subject-interrogative.
The evidence from Bresnan's paper, which I have reviewed in this section,
shows that several subject properties may occur on either the locative PP or the
postposed subject NP. Only three properties are not able to occur on the PP:
these are agreement, being the subject of a participial relative, and an inversion
in non-subject interrogatives. The tag question data appears to favour a subject
analysis of the locative PP rather than the postposed NP.
In the next section, I set out to accommodate those facts within Word
Grammar. As I stated in the introduction, these facts are problematic for the
model put forward in Hudson (1990), because there is only a single account of
subjects in that theory. In the next section, I review some of the dimensions of
subjecthood, in the light of the discussion in section 2. 1 go on to argue that we
need to split subjects into three kinds - lexical subjects, syntactic subjects and
morphosyntactic subjects - and that this division can take account of the pattern
of data reported in section 3.
Lexical properties
• Subjects are typically the dependent which expresses the Agent argument in
the active voice.
• (The sole arguments of intransitives typically show (other) subject
properties. )
FACTORING OUT THE SUBJECT DEPENDENCY 217
Syntactic properties
• The subject can be the shared argument in coordination.
• The subject is the only argument of an xcomp which may be shared in a
predicative complementation structure (in both raising and control).
• English subjects resist extraction. (In other languages subjects are often
more extractable than other arguments).
• Auxiliary inversion fails with a subject interrogative.
• Subjects are usually the discourse topic.
Morphosyntactic properties
• (The addressee of an imperative is a subject. )
• The subject is often obligatory.
• Subject-verb agreement.
As we can see, the lexical properties are those properties which are primarily to
do with the mapping of semantic roles to grammatical functions. The first two
items listed as lexical properties concern the mapping of semantic roles in
particular in nominative-accusative languages (like English) rather than
absolutive-ergative languages (like West Greenlandic). I have the second two
items which are listed as lexical properties in brackets, because these are shared
with other parts of the grammar: the identification of the subject of imperatives
is not only a lexical property. What makes it a 'lexical' subject is that the subject
of an imperative picks out the same semantic role as the first two criteria - the
linking facts apply here as well. But this subject-criterion could also be
morphosyntactic, because the imperative is a mood, and mood is a
morphosyntactic feature. Excepting certain well-known construction types, it
is only the imperative mood that permits subjects not to be represented by an
overt noun or noun phrase in English.
The syntactic properties are those that have to do with the grammatical
phenomena that are commonly called 'movement' or 'deletion'. The first three
can be subsumed under the descriptive generalization that the subject is the
element which can be an argument of more than one predicate. The English
extraction data are at odds with the more typical extraction data, but they still
show that subjects can be identified by their extraction properties. I have
included the non-inversion of subject interrogatives as a syntactic fact (rather
than a morphosyntactic one) on the grounds that subject-inversion is a word
order rule, and is therefore syntactic. For this reason, the fact that PP locatives
behave like other subjects shows that they have the same syntactic properties as
other subjects: they resist subject-inversion in interrogatives.
The final observation listed under syntactic subjects is arguably not even
grammatical - but there are constructional interactions between topic and
syntactic structure and, indeed, focus and syntactic structure. Again, there is a
descriptive generalization to be captured, that subjects generally are topics.14
The morphosyntactic properties of subjects tend to be linguistically specific.
218 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
Tenseless languages will not show any morphosyntactic subject properties, and
as Hudson (1999) shows, such properties are in decay in English. I have already
discussed the issue to do with the imperative. I have put the obligatory criterion
here, because it is related to agreement: in languages which have a highly
developed agreement morphology in both the verbal conjugation system and
the nominal declension system it is possible for subjects to be omitted. English
has obligatory subjects - which can be expletive - which appear to be obligatory
because of the impoverished inflectional morphology. 15
It is possible, on the basis of this discussion, to make some general
predictions about what might be found cross-linguistically. The lexical
properties of subjects will vary according to whether the language is nominative
or ergative. The morphosyntactic properties of subjects will vary according to
whether the language has a rich inflectional system, an impoverished
inflectional system, or no inflectional system. And the syntactic properties of
subjects should be relatively consistent across languages: to the extent that there
is variation in the syntactic properties of subjects, it should be attributable to the
interaction between this dimension of subjecthood and one of the other
dimensions.
From the point of view of Word Grammar, these observations about
subjects are only salient if factored-out subjects can do two things: exist in
mismatch structures, so that no single grammatical relation can be held to
obtain between a verb and another element in a clause; and help capture
descriptive facts better than simply treating subject-of as a single unitary relation.
If we return to the data presented in Table 1, we can see that all of the
subject properties that are found on the inverted PP are syntactic subject
properties in that they are all subject properties that are relevant to the ability of
a single noun or noun phrase to be construed as an argument of more than one
predicate. The properties that were found on the locative PP were:
• subject raising;
• tag questions;
• subject extraction;
• extraction from coordinated constituents;
• subject interrogatives.
Excepting the tag-question data (which are arguably morphosyntactic, and which
were, in any case, moot) all of these subject properties are properties that are
related to the ability of a single subject entity to be an argument of more than
one predicate. What we find is that locative PPs can behave like syntactic
subjects, and that when they do behave like syntactic subjects, the NP theme
argument of the verb cannot behave like a syntactic argument.
This is half of an argument that subject properties are split. The other half of
the argument is that the morphosyntactic properties are found on the NP. The
properties that were found on the NP in English, but not on the preposed PP,
were:
FACTORING OUT THE SUBJECT DEPENDENCY 219
• agreement;
• subject of participial relative.
The examples in (35) and (36a-36b) show that the locative PP in locative
inversion cannot undergo subject-inversion. I think that the crucial thing here is
that this is not a syntactic property of subjecthood, but a morphosyntactic one. 16
It is not a syntactic one, because the syntactic constraints were generally
concerned with the ability of a single phrase to occur as an argument of more
than one predicate. The restriction in (35) and (36) is different; in fact, it is
attributable to the PP's lack of morphosyntactic properties. I take it that an
auxiliary cannot invert with a phrase that it cannot agree in number with. 17
However, the examples in (37) and (38), which I repeat here, show that the
locative PP behaves like a subject with respect to subject interrogatives:
In both examples in (39), Jane is the 'er' of'run'. However, in neither example
can Jane be thought of as the morphosyntactic subject, because there is no
agreement: the infinitive does not have a feature-value 'number'.
The examples in (40) are even more acute: the gerund running also has no
value for number, but it does have a subject - the pronoun me in (40a) and the
pronoun my in (40b). Again, in neither case can this be thought of as
morphosyntactic subjecthood.
Additionally, the evidence from gerunds shows that lexical subjects have to be
distinguished from syntactic subjects: in (40a), me is the head of running, and in
(40b), my is the head of running. From these examples, we have to conclude
that there are cases where lexical subjecthood has to be distinguished from
syntactic subjecthood and from morphosyntactic subjecthood.
Another example, although a negative one, comes from weather IT. The
example in (41) shows that weather IT can be simultaneously a morphosyntactic
and a syntactic subject, even though it cannot be a lexical subject given that the
verb rain does not have any semantic arguments.
The example in (41) is also important because it shows that the property of
222 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
being a syntactic subject is not co-extensive with being a topic. Both Bresnan
(1994) and Falk (2004) argue that syntactic subjects are topics of some kind, but
an expletive pronoun is not a candidate for topichood.
5. Conclusions
In this section, I argue that the treatment of the data presented in this chapter
handles the facts and the data more satisfactorily than the mismatch account of
Bresnan (1994), and that it is more compatible with other general assumptions
about the architecture of grammar that Word Grammar adopts. On the basis of
the locative inversion evidence, I have made a distinction between
morphosyntactic and syntactic subjects, and on the basis of further evidence
from other constructions have made a further distinction which separates lexical
subjects out from the other kinds of subject.
Bresnan also argues for a three-way distinction, but in her case the
factorization of subjecthood is over three of the domains of structure that LFG
recognizes: a-structure, f-structure and c-structure. She effectively argues that the
locative PPs can only be construed as subjects because they are also topics. The
problem with this account is that it treats 'topic' as fundamentally syntactic,
located in f-structure, when it is clear (a) that subjects need not be topics; and (b)
that some subjects cannot be topics. Furthermore, we have seen that the
properties of subjects itemized in section 4 do not require there to be a separate
dimension of topichood - it is simply the case that some subjects are syntactic
rather than morphosyntactic.
In some senses, the different approaches between this chapter and Bresnan
(1994) are due to underlying assumptions that the two models have, which
make them different from each other. LFG does not permit there to be a
mapping of more than one f-structure relation between two elements; Word
Grammar does not distinguish between argument structure and the instantiated
dependencies in a given construction. But it is also the case that the WG
account espoused here allows the theory of subjects to be elaborated so that it
can account for a wide range of differences in the spectrum of subject
properties.
There are some obvious avenues for future research: for example, both
West Greenlandic and Mandarin are tenseless. For this reason, Mandarin has
been argued not to have the subject and object dependencies that are witnessed
in other languages. However, while Mandarin has long-distance reflexives, West
Greenlandic does not. One salient difference is that Mandarin is a nominative
language while West Greenlandic is an ergative language, and so the question is
begged whether these facts are attributable to differences in lexical subjects in
these languages.
Certainly more research is required on the cross-linguistic typology of
dependencies. Meanwhile, it is clear that the English Word Grammar model
needs to be revised, to admit at least three different kinds of subject.
FACTORING OUT THE SUBJECT DEPENDENCY 223
References
Aissen, J. (1975), 'Presentational-£/z<?re insertion: a cyclic root transformation'. Chicago
Linguistics Society, 11, 1-14.
Anderson, J. M. (1997), A Notional Theory of Syntactic Categories. Cambridge:
Cambridge University Press.
Andrews, A. (1985), 'The major functions of the noun phrase', in T. Shopen (ed. ),
Language Typology and Syntactic Description, vol. 1: Clause Structure. Cambridge:
Cambridge University Press, pp. 62-154.
Bowers, J. S. (1976), 'On surface structure grammatical relations and the structure-
preserving hypothesis'. Linguistic Analysis, 2, 584-6.
Bresnan, J. W. (1994), 'Locative inversion and the architecture of grammar'. Language,
70, 72-131.
Chomsky, N. (1981), Lectures on Government and Binding. Dordrecht: Foris.
Dixon, R. M. W. (1994), Ergativity. Cambridge: Cambridge University Press.
Falk, Y. (2004), 'Explaining subjecthood' (Unpublished manuscript, Hebrew University
of Jerusalem).
Gueron, J. (1980), 'The syntax and semantics of PP-extraposition'. Linguistic Inquiry,
11, 637-78.
Hudson, R. A. (1990), English Word Grammar. Oxford: Blackwell.
— (1999), 'Subject-verb agreement in English'. English Language and Linguistics, 3, 173-
207.
Jackendoff, R. S. (1990), Semantic Structures. Cambridge, MA: MIT Press.
Keenan, E. L. (1976), 'Towards a universal definition of "subject"', in Charles Li (ed. ),
Subject and Topic. New York: Academic Press, pp. 303-33.
Keenan, E. L. and Connie, B. (1977), 'Noun phrase accessibility and universal
grammar'. Linguistic Inquiry, 8, 63-99.
McCloskey, J. (1997), 'Subjecthood and subject position', in L. Haegeman (ed. ), Elements
of Grammar: Handbook of Generative Syntax. Dordrecht Kluwer, pp. 197-235.
— (2001), 'The distribution of subject properties in Irish', in W. D. Davies and S.
Dubinsky (eds), Objects and Other Subjects. Dordrech: Kluwer, pp. 157-92.
Manning, C. (1996), Ergativity. Stanford, CA: CSLI Publications.
Notes
1 Unless I explicitly state otherwise, the examples in section 1 and section 3 (where I
present Bresnan's locative inversion data) are taken from Bresnan's (1994) paper,
and the grammaticality judgements are hers. I have, however, silently amended
Bresnan's spelling to British English norms.
2 This is a pre-formal statement, and I do not intend it to commit me to any particular
theoretical position.
3 English makes a distinction between disjoint pronouns - the forms him, her, me and
so forth, and anaphoric pronouns like himself, herself, myself and so forth. Not all
languages make this distinction, and English has not always made the same
distinction in the course of its history.
4 The underscore represents the subject position for to. I do not mean by this
representation to suggest that there is actual movement - like Hudson (1990) I
reject a movement account The representation is intended to be pre-formal, and is
borrowed from Bresnan (1994), whose examples I borrow in section 3 - and in
borrowing some of these examples, I import the representation.
5 I adopt the analysis of subject-verb agreement presented in Hudson (1999), which
224 WORD GRAMMAR: PERSPECTIVES ON LANGUAGE STRUCTURE
KENSEI SUGAYAMA