Cognitive Linguistics Croft Cruse (2004)
Cognitive Linguistics Croft Cruse (2004)
Cognitive Linguistics Croft Cruse (2004)
Cognitive Linguistics
Cognitive Linguistics argues that language is governed by general cognitive prin-
ciples, rather than by a special-purpose language module. This introductory text-
book surveys the field of cognitive linguistics as a distinct area of study, presenting
its theoretical foundations and the arguments supporting it. Clearly organized and
accessibly written, it provides a useful introduction to the relationship between
language and cognitive processing in the human brain. It covers the main top-
ics likely to be encountered in a course or seminar, and provides a synthesis of
study and research in this fast-growing field of linguistics. The authors begin by
explaining the conceptual structures and cognitive processes governing linguis-
tic representation and behavior, and go on to explore cognitive approaches to
lexical semantics, as well as syntactic representation and analysis, focusing on
the closely related frameworks of cognitive grammar and construction grammar.
This much-needed introduction will be welcomed by students in linguistics and
cognitive science.
General editors: s . r . a n d e r s o n, j. b r e s na n, b. c o m r i e ,
w. d r e s s l e r , c . e w e n, r . h u d d l e s t o n, r . l a s s ,
d. lightfoot, j. lyons, p. h. matthews, r. posner,
s. romaine, n. v. smith, n. vincent
In this series:
p. h . m at t h e w s Morphology Second edition
b. c o m r i e Aspect
r . m . k e m p s o n Semantic Theory
t. bynon Historical Linguistics
j. allwood, l.-g. anderson and ö. dahl Logic in Linguistics
d. b. f r y The Physics of Speech
r. a. hudson Sociolinguistics Second edition
a . j. e l l i o t Child Language
p. h . m at t h e w s Syntax
a. redford Transformational Syntax
l . b au e r English Word-Formation
s. c. levinson Pragmatics
g. brown and g . y u l e Discourse Analysis
r . h u d d l e s t o n Introduction to the Grammar of English
r . l a s s Phonology
a. comrie Tense
w. k l e i n Second Language Acquisition
a. j. woods, p. fletcher and a. hughes Statistics in Language Studies
d. a. cruse Lexical Semantics
a. radford Transformational Grammar
m . g a r m a n Psycholinguistics
g . g . c o r b e t t Gender
h. j. giegerich English Phonology
r . c a n n Formal Semantics
j. l ave r Principles of Phonetics
f. r . pa l m e r Grammatical Roles and Relations
m. a. jones Foundations of French Syntax
a. radford Syntactic Theory and the Structure of English: A Minimalist Approach
r . d. va n va l i n, j r, and r . j. l a p o l l a Syntax: Structure, Meaning and Function
a. duranti Linguistic Anthropology
a . c r u t t e n d e n Intonation Second edition
j. k . c h a m b e r s and p. trudgill Dialectology Second edition
c. lyons Definiteness
r . k ag e r Optimality Theory
j. a. holm An Introduction to Pidgins and Creoles
c . g . c o r b e t t Number
c . j. e w e n and h. van der hulst The Phonological Structure of Words
f. r . pa l m e r Mood and Modality Second edition
b. j. b l a k e Case Second edition
e . g u s s m a n Phonology: Analysis and Theory
m. yip Tone
w. c r o f t Typology and Universals Second edition
f. c o u l m a s Writing Systems: an Introduction to their Linguistic Analysis
p. j. hopper and e . c . t r au g o t t Grammaticalization Second edition
l. white Second Language Acquisition and Universal Grammar
i. plag Word-formation in English
w. c r o f t and d. a . c r u s e Cognitive Linguistics
a. siewierska Person
a. radford Minimalist Syntax: Exploring the Structure of English
d. b üring Binding Theory
Cognitive Linguistics
W I L L I A M C RO F T
University of Manchester
and
D . A L A N C RU S E
University of Manchester
Cambridge, New York, Melbourne, Madrid, Cape Town, Singapore, São Paulo
Cambridge University Press has no responsibility for the persistence or accuracy of s
for external or third-party internet websites referred to in this publication, and does not
guarantee that any content on such websites is, or will remain, accurate or appropriate.
Contents
vii
viii Contents
3.4.2 Deixis 59
3.4.3 Subjectivity 62
3.5 Constitution/Gestalt 63
3.5.1 Structural schematization 63
3.5.2 Force dynamics 66
3.5.3 Relationality (entity/interconnection) 67
3.6 Conclusion 69
8 Metaphor 193
8.1 Figurative language 193
8.2 The conceptual theory of metaphor 194
8.2.1 Introduction 194
8.2.2 Issues in the conceptual theory of metaphor 198
8.3 Novel metaphor 204
8.3.1 The life history of a metaphor 204
8.3.2 How do we recognize metaphors? 206
8.3.3 Blending Theory and novel metaphors 207
8.3.4 Context sensitivity 209
8.3.5 Asymmetry of vehicle and target 210
x Contents
References 330
Author index 344
Subject index 347
Figures
xii
Tables
xiii
Preface
xv
1
Introduction: what is
cognitive linguistics?
1
2 Introduction
species apparently has. However, innateness of cognitive abilities has not been a
chief concern of cognitive linguists, who are more concerned with demonstrating
the role of general cognitive abilities in language.
The hypothesis that language is not an autonomous cognitive faculty has had two
major implications for cognitive linguistic research. Much cognitive linguistic re-
search has been devoted to elucidating conceptual structure and cognitive abilities
as they are seen to apply to language, in the effort to demonstrate that language can
be adequately modeled using just these general conceptual structures and cognitive
abilities. Part I of this book is devoted to explicating cognitive linguistic models
of cognitive structure and abilities (see also chapter 11).
Second, cognitive linguists appeal at least in principle to models in cognitive psy-
chology, in particular models of memory, perception, attention and categorization.
Psychological models of memory have inspired linguistic models of the organi-
zation of linguistic knowledge into frames/domains (chapter 2), and grammatical
knowledge in networks linked by taxonomic and other relations (see chapters
10–11 in Part III). Psychological models of attention and perception, especially
Gestalt psychology, have led to the explication of many conceptualization pro-
cesses in semantics (chapter 3, and see also the next paragraph). Finally, psycho-
logical models of categorization, in particular prototypes and graded centrality,
and more recent models of category structure, have had perhaps the greatest influ-
ence on both semantic and grammatical category analysis in cognitive linguistics
(chapter 3; see, e.g., Lakoff 1987, Taylor 1989[1997]).
The second major hypothesis of the cognitive linguistic approach is embodied
in Langacker’s slogan ‘grammar is conceptualization.’ This slogan refers to a more
specific hypothesis about conceptual structure, namely that conceptual structure
cannot be reduced to a simple truth-conditional correspondence with the world. A
major aspect of human cognitive ability is the conceptualization of the experience
to be communicated (and also the conceptualization of the linguistic knowledge
we possess). A major theme of the chapters in Part I of this book is that all aspects
of conceptual structure are subject to construal, including the structure of cate-
gories (chapter 4) and the organization of knowledge (i.e., conceptual structures;
chapter 2). In particular, it is argued that grammatical inflections and grammatical
constructions play a major role in construing the experience to be communicated
in specific ways (chapter 3). Part II of this book also explores and defends the
conceptualization hypothesis for a wide range of lexical semantic phenomena, in-
cluding topics widely discussed in cognitive linguistics (polysemy and metaphor)
and lexical semantic topics that have not generally been examined by cognitive
linguists (namely lexical relations such as antonymy, meronomy and hyponymy).
The third major hypothesis of the cognitive linguistic approach is that knowl-
edge of language emerges from language use. That is, categories and structures
4 Introduction
in semantics, syntax, morphology and phonology are built up from our cogni-
tion of specific utterances on specific occasions of use. This inductive process of
abstraction and schematization does not lose the conventionalized subtleties and
differences found among even highly specific grammatical constructions and word
meanings.
As we noted above, this hypothesis is a response to approaches to syntax and
semantics in which highly general and abstract schemas and categories, sometimes
claimed to be innately given, are assumed to govern the organization of linguistic
knowledge, and apparently idiosyncratic or anomalous patterns are relegated to
the periphery. Instead, cognitive linguists argue that the detailed analysis of subtle
variations in syntactic behavior and semantic interpretation give rise to a different
model of grammatical representation that accommodates idiosyncratic as well as
highly general patterns of linguistic behavior (see, e.g., the arguments in chapter 9).
In semantics, this model is manifested in Fillmore’s semantics of understanding
(chapter 2), and Cruse’s dynamic construal approach to categorization (chapter 4
and Part II; see also Croft 2000:99–114). In syntax, this hypothesis has given rise
directly to construction grammar as a new theory of syntax, and the usage-based
model, developed in greatest detail for morphology and phonology. These models
of syntax and morphology are described in Part III of this book.
PA RT I
A conceptual approach to
linguistic analysis
2
Frames, domains, spaces: the
organization of conceptual structure
1 We follow the practice of Fillmore (1982a) and Langacker (1987) in using lower-case italics to
represent the word form, and capitals to represent the concept underlying the word meaning.
7
8 A conceptual approach to linguistic analysis
The need for another means for organizing concepts has been felt by researchers
in cognitive psychology and artificial intelligence as well as in various branches
of linguistics, and has led to a variety of similar proposals, each typically with its
own name. Among these names are: frame, schema, script, global pattern, pseudo-
text, cognitive model, experiential gestalt, base, scene (Fillmore 1985:223, n. 4).
The most influential version of this proposal in cognitive linguistics has been the
model of frame semantics developed by Fillmore. We present Fillmore’s theory
and arguments in this section, and turn to extensions of Fillmore’s ideas by other
cognitive linguists in later sections.2
Fillmore views frames not as an additional means for organizing concepts, but as
a fundamental rethinking of the goals of linguistic semantics. Fillmore describes his
frame semantic model as a model of the semantics of understanding, in contrast
to a truth-conditional semantics: the full, rich understanding that a speaker intends
to convey in a text and that a hearer constructs for that text. Fillmore argues that in
the analysis of linguistic meaning, understanding is the primary data; truth-value
judgments and judgments of semantic relations such as synonymy and implication
are derivative and theory-driven (Fillmore 1985:235). Fillmore’s frame semantics
brings linguistic semantics back to that primary data and does not exclude any of
it from consideration.
Fillmore uses a tool metaphor to describe the understanding process (Fillmore
1982a:112): a speaker produces words and constructions in a text as tools for a
particular activity, namely to evoke a particular understanding; the hearer’s task
is to figure out the activity those tools were intended for, namely to invoke that
understanding. That is, words and constructions evoke an understanding, or more
specifically a frame; a hearer invokes a frame upon hearing an utterance in order
to understand it.
Fillmore uses a wide range of examples to demonstrate that there are signifi-
cant phenomena in linguistic semantics that cannot easily be captured in a model
of structural semantics, semantic features and/or truth-conditional semantics. We
survey his arguments here.
The analysis of semantic features is often justified on the basis of lexical sets
that appear to be analyzable in terms of a simple set of features. For example,
the lexical set in (1) can be analyzed in terms of the features [MALE/FEMALE],
[ADULT/YOUNG], and [UNMARRIED]:
2 The basic sources for Fillmore’s ideas are Fillmore 1975, 1977 (an expanded version of the first
paper), 1982a, 1985, 1986. Unfortunately all of these are difficult to access.
Frames, domains, spaces 9
Yet our understanding of these concepts is more complex than this paradigm of
feature constrasts implies. The relation between man/boy and woman/girl is not the
same: for many people, the term girl is used for female humans at a significantly
higher age than the term boy is used for male humans (Fillmore 1982a:126).
Moreover, the attitudes towards the sexes that this linguistic behavior is assumed
to evoke has led to changes in the relationship and hypercorrection such that
the term woman is attested as being applied even to an eight-year-old girl (ibid.,
127). In a frame semantic analysis, man, boy, woman and girl evoke frames that
include not just the biological sexual distinction but also differences in attitudes
and behavior towards the sexes that would explain the traditional asymmetry in
the use of boy/girl and the more recent change in the use of woman, including
its hypercorrective use. Likewise, the difference between our understanding of
bachelor and our understanding of spinster involves much more than a simple
feature [MALE/FEMALE] (ibid., 131).
Many lexical contrasts contain semantic asymmetries that cannot be captured
by features (except in an ad hoc fashion), but lend themselves easily to a frame
semantic account. For example, the opposing terms used for the vertical extent of
an erect human being are tall and short, for vertical distance from a bottom baseline
(e.g. a branch of a tree) they are high and low, but for the vertical dimension of
a building they are tall and low (Fillmore 1977a:71). It would be difficult if not
impossible to come up with a unitary feature definition of tall that captured its
different contexts of use from high, and did the same for short vs. low. Instead,
one can simply describe the frames for humans, buildings and other objects, and
specify which words are used for vertical extent or distance in that frame.
Similarly, no simple unitary definitions would capture the contrast between the
adjectives live and alive given in (2)–(4) (Fillmore 1977a:76–77):
including different plural forms; a unitary definition of brother would miss the
frame contrast (ibid., 76).
Fillmore notes that his frame semantic model shares significant properties with
lexical (semantic) field theory (Fillmore 1985:225–26; 1992:76–77). Lexical field
theory groups together words that are associated in experience, not unlike frame
semantics. However, lexical field theory differs from frame semantics in that words
are defined relative to other words in the same lexical field, whereas in frame
semantics, words are defined directly with respect to the frame. For example, in
lexical field theory, one would observe that large in the field of sizes of packages
of soapflakes is in contrast with jumbo, economy giant and family size and hence
describes the smallest size in the field, unlike uses of large in other lexical fields
(Fillmore 1985:227).
In frame semantics, the same observation can easily be captured: large labels
the smallest size in the SOAPFLAKES frame. But lexical field theory predicts that
the meaning of a word in a field can only be defined in contrast to neighboring
words in the field. Lexical field theory has difficulties if there are no neighboring
words, or a speaker does not know the neighboring words: it predicts that the term
has a different meaning. Fillmore notes that while German has a word for the
sides of a right angle triangle other than the Hypotenuse, namely Kathete, most
English speakers do not have such a word (ibid., 228–29). Yet the understanding
of English hypotenuse and German Hypotenuse is the same, provided the speaker
understands what a right angle triangle is. This is not a problem in frame semantics,
where the word concept is linked directly to the frame, in this case the RIGHT
ANGLE TRIANGLE frame.
Another argument in favor of a frame-based approach to lexical semantics are
words whose corresponding concepts inherently refer to other concepts extrinsic
to the concept denoted by the word. Some word concepts refer to a prior history of
the entity denoted. A scar is not just a feature of the surface of someone’s skin, but
the healing state of a wound; a widow is a woman who was once married but whose
husband has died (Fillmore 1977a:73). Other word concepts, especially for prop-
erties and actions, cannot be understood without understanding something about
the participant in the action or possessor of the properties: one cannot understand
gallop without knowing about the body of a horse, or hungry without understand-
ing the physiology of living things (ibid., 73–74). This is true of object concepts
as well: lap cannot be understood except in reference to a person’s posture and the
function of one’s lap in supporting another object (ibid.).
Another clear class of examples that requires reference to extrinsic entities are
deictic expressions that evoke the speech act situation (Fillmore 1982a:117). For
example, the past tense situates an event in a point or interval or time relative to
the speech act situation. The speech act situation, including its time of occurrence,
Frames, domains, spaces 11
functions as the frame against which past time reference is profiled. Likewise, all
other deictic words and inflections, such as person deixis (I, you, he/she/it, we, they
and person-based agreement inflections) and spatial deixis (this, that, here, there),
evoke the speech act situation. Other types of grammatical words and inflections
also have meanings evoking the speech act situation. For example, the definite
articles the and a define the identity of the noun referent relative to the mutual
knowledge of speaker and hearer (the basically indicates mutually known, a not
mutually known, in most contexts). The meanings of the and a evoke the speech
act situation because they make reference to the mental states of speaker and hearer
(see also §3.4).
Above all, many word concepts cannot be understood apart from the intentions
of the participants or the social and cultural institutions and behavior in which
the action, state or thing is situated. For example, the concept VEGETARIAN
only makes sense in the frame of a culture in which meat-eating is common; the
concepts STRIKE or BORROW can only be understood in the frame of a culture
in which such actions occur (Fillmore 1982a:120). Even something as simple as
an apple core evokes a frame describing a particular way of eating apples: ‘an
apple-core is that part of the apple that somebody who eats apples the way most
of us do has left uneaten’ (Fillmore 1977a:73).
Another respect in which a word meaning makes reference to extrinsic entities is
that a word allows the speaker and hearer to focus their attention on only part of an
entire frame; no one word gives the full structure of the frame. The classic example
is the commercial transaction frame (Fillmore 1977a:58–59; 1977b); but a much
clearer case is the RISK frame (Fillmore and Atkins 1992). Fillmore and Atkins
identify the following elements of the RISK frame: Chance (uncertainty about the
future), Harm, Victim (of the Harm), Valued Object (potentially endangered by the
risk), Situation (which gives rise to the risk), Deed (that brings about the Situation),
Actor (of the Deed), (Intended) Gain (by the Actor in taking a risk), Purpose (of
the Actor in the Deed), Beneficiary and Motivation (for the Actor). The verb risk
occurs in many syntactic constructions, some of which are exemplified in (5a–e),
but none of them include all or even most of the elements of the RISK frame
(Fillmore & Atkins 1992: 83, 87, 89, 94, 96; all but the first are corpus examples):
(5) a. You’ve (Actor/Victim) risked your health (Valued Object) for a few cheap thrills
(Gain).
b. Others (Actor/Victim) had risked all (Valued Object) in the war (Situation).
c. She (Actor/Victim) had risked so much (Valued Object) for the sake of vanity
(Motivation).
d. Men (Actor/Victim) were not inclined to risk scalping (Harm) for the sake of
settlers they had never seen (Beneficiary).
e. I (Actor/Victim) didn’t dare risk a pause (Deed) to let that sink in (Purpose).
12 A conceptual approach to linguistic analysis
In a frame semantic analysis, any of the uses of risk evokes the entire RISK frame,
even if only part of that frame is overtly focused on by the construction in which
risk is used.
The semantics of understanding also allows Fillmore to account for linguistic
facts that do not lend themselves to a truth-functional analysis. For example, the
collocations in (6) could be reversed as in (7) without producing semantic anomaly
(Fillmore 1977a:75–76):
In other words, the difference between (6) and (7) cannot be accounted for by
semantic constraints. But the examples in (6a–b) sound much more natural because
the noun and the verb in each sentence both evoke the same frame.
Likewise, a truth-conditional semantics cannot capture many aspects of our
understanding of (8) (Fillmore 1985:230–31):
Fillmore notes that choosing father or dad (without the possessive) would express
a different relationship between the speaker and the speaker’s father; the morning
is understood to be defined against the frame of the working day (i.e, around 8am
to noon) rather than the calendar day (midnight to noon); waste frames the use of
time very differently from spend; and on the bus frames the speaker’s location in
terms of the bus being in service, rather than simply a physical container (which
would be evoked by in the bus).
A truth-conditional model also cannot account for the anomaly of frames that
are appropriate at one time of utterance but not at another because the world has
changed in the meantime. Fillmore uses the contrived example in (9), noting that
it could be said in 1984 but not in, say, 1919 (Fillmore 1985:238–39):
(9) During World War I, Ronald Reagan’s birth mother dropped his analog watch
into the sound hole of the acoustic guitar.
Such a sentence could be uttered in 1984, because World War II had occurred,
allowing the 1914–18 war to be renamed World War I; medical technology had
allowed the dissociation of the birth mother from the genetic mother (who donates
the egg); and electric guitars and digital watches had been invented. None of these
framings of the objects, persons or events was available in 1919, and so (9) would
be an impossible utterance at that time, even if true retrospectively.
Finally, frame semantics offers a natural account of a number of problematic
phenomena that seem to be caught between semantics and pragmatics, including
Frames, domains, spaces 13
Example (10) denotes (or entails) a particular mental state of John, namely his
feeling of regret, but (in one analysis) is said to presuppose that John did sign
the letter. The problem for truth-conditional semantics is that if John did not sign
the letter, (9) has no truth conditions. In order to preserve truth conditions one
may shunt the presupposition problem off to pragmatics. This seems odd since
the presupposition is associated with a particular word and its meaning, which
is semantic. But presuppositions display a further type of peculiar behavior, in
negative sentences such as (11):
In one interpretation, the entailment is negated – John does not have any regrets –
but the presupposition is not – John signed the letter. However, there is another
interpretation of (11), namely that John did not regret signing the letter because
he did not sign it (Fillmore 1985:251).
Fillmore argues that the behavior of presuppositions can be easily accounted
for in a frame semantic analysis. The concept REGRET includes in its frame the
accomplishment of an action towards which the regretter has his/her regrets. If
the action is absent from the frame, understanding of the positive sentence fails.
Negation, on the other hand, can negate either the concept denoted or the frame
itself. Negating the state of affairs in the frame preserves the rest of the frame,
including the action that could have led to the regrets. This is the first interpretation
of (11) described above. Negating the entire frame, on the other hand, also negates
the action that could have led to the regrets. This is the second interpretation of
(11): the speaker denies the framing of the situation as including the action of John
having signed the letter.
We may compare (11) to other examples of frame negation, such as the one in
(12) (Fillmore 1985:245, from Wilson 1975:138):
(12) You didn’t spare me a day at the seaside: you deprived me of one.
In (12), the speaker denies the positively evaluated framing of the action as sparing
her, and replaces it with the negatively evaluated framing of depriving her.
Finally, the semantics of understanding plays a major role in text understand-
ing. For example, the well-known example of initial definite reference, e.g. the
carburetor in (13) (described as ‘bridging’ by Clark and Haviland [1977] and
‘evoked’ by Prince [1981a]), are in fact due to the frame evoked by the first
sentence (Fillmore 1977a:75):
14 A conceptual approach to linguistic analysis
(13) I had trouble with the car yesterday. The carburetor was dirty.
The car in the first sentence evokes a frame that allows the hearer to identify which
carburetor of the millions in the world the speaker was referring to. But frame
semantics contributes more than the resolution of definite reference to the analysis
of the coherence of texts. Fillmore contrasts (13) with (14) (ibid.):
(14) I had trouble with the car yesterday. The ashtray was dirty.
The second sentence in (14) is incoherent with the first, even though the definite
reference can be resolved (most cars have ashtrays). The reason for this is that there
is nothing evoked in the frame of having trouble with the car that has anything to
do with the ashtray – unlike (13), because dirty carburetors do cause problems for
cars.
Fillmore’s arguments present a wide range of data that justify the introduction
of frames to the analysis of linguistic semantics, and the replacement of a truth-
conditional semantics with a semantics of understanding. In the following sections,
we lay out more systematically the frame semantic model and follow its further
development in cognitive linguistics.
a line segment, but not any line segment: the line segment is defined relative to the
structure of the circle. In other words, one can understand RADIUS only against
a background understanding of the concept CIRCLE, which can be geometrically
illustrated as in Figure 2.1.
RADIUS
CIRCLE
Figure 2.1 RADIUS and CIRCLE
In other words, the concepts RADIUS and CIRCLE are intimately related, and
this relationship must be represented in conceptual structure. Langacker describes
the relationship between RADIUS and CIRCLE as one of a concept profile against
a base. The profile refers to the concept symbolized by the word in question. The
base is that knowledge or conceptual structure that is presupposed by the profiled
concept. Langacker also uses the term domain for the base (this term is also used
in Lakoff 1987). This is identical to Fillmore’s frame (§2.1): ‘by the term “frame”
I have in mind any system of concepts related in such a way that to understand any
one of them you have to understand the whole structure in which it fits’ (Fillmore
1982a:111). The term ‘profile’ has also come to be used as a verb to describe
the relationship between word form and word meaning (profile+base): e.g. radius
profiles a particular line segment in the CIRCLE base/domain/frame.
A concept profile is insufficient to define a word concept, because it presup-
poses other knowledge in its definition, namely its base. But a single base, such
as CIRCLE, is a complex conceptual structure that includes a wide range of con-
cept profiles, such as RADIUS, ARC, CENTER, DIAMETER, CHORD and so
on. Hence the base alone is insufficient to define a linguistic concept either. The
conclusion that follows from this is that the meaning of a linguistic
unit must specify both the profile and its base. This is identical
to Fillmore’s conclusion regarding concept frames.
The fact that a base supports multiple concept profiles is what makes the base a
domain, in the intuitive sense: several different concept profiles have it as a base.
We can now define a domain as a semantic structure that functions
as the base for at least one concept profile (typically, many
profiles). As Taylor (1989[1997]:84) notes, ‘In principle, any conceptualization
or knowledge configuration, no matter how simple or complex, can serve as the
16 A conceptual approach to linguistic analysis
cognitive domain for the characterization of meanings.’ We may now say that the
domain CIRCLE includes the concepts of an arc, a diameter, a radius, a chord and
so on.
The canonical example of a profile-base relation is the part-whole relation: all
agree that a concept such as ARM cannot be defined without reference to BODY. A
similar class of concepts are kin terms such as daughter. The concept DAUGHTER
presupposes the concept PARENT, and the particular type of kin relationship that
holds between them. The concept NIECE presupposes other kinship concepts, and
more complex kin relationships. In other words, the base against which a profile is
defined can be more complex than just the whole of which some entity is a part. In
some cases, one cannot always find a single simple word to describe the base: for
NIECE, perhaps the best description of the base is KINSHIP SYSTEM, or some
part of that system (see §2.4).
But it is not only relational nouns that represent a concept profile against a
base, as we saw in §2.1. Consider another example, the word weekend (Fillmore
1985:223–24). The concept WEEKEND can only be understood against a whole
background system of the calendrical cycle, defined partly by natural phenomena
(the sequence of day and night) and cultural conventions (the seven-day week
cycle, and its division into working days and nonworking days). Likewise, the
concept BUY can only be understood against a background knowledge of the
commercial transaction situation. Different aspects of the commercial transaction
are profiled by BUY, SELL, PAY, COST and so forth. Such domains/frames cannot
be readily represented in a geometric form in the way that RADIUS and CIRCLE
are represented in Figure 2.1, although schematic diagrams are often resorted to
in cognitive linguistics in order to represent the complex interconnectedness of
concepts in domains or frames.
In fact, no concept exists autonomously: all are understood to fit into our general
knowledge of the world in one way or another. What matters for semantic analysis is
the profile-base relation, and the relationships between bases and domains. Some of
the corollaries of this analysis of word meaning into profile and base/frame/domain
will be explored in the following section.
linguists. However, the terms frame and domain continue to compete for usage,
and base is also used among cognitive grammarians. We will use the terms frame
and domain interchangeably here. Nevertheless, there are still other terms that have
been proposed to describe types of semantic analyses that bear a strong affinity
to frame semantics. We mention three influential theories here, which originated
in artificial intelligence (scripts), cognitive psychology (the ‘theory theory’) and
sociology (communities).
The examples of frames given above appear to be largely static in character.
But this is not necessary: a frame is any coherent body of knowledge presupposed
by a word concept. In particular, frames can include dynamic concepts, that is,
extending through time. For example, PURIFIED presupposes in its frame a prior
impure state of the entity which is then changed by some process; in contrast,
PURE does not presuppose anything about prior states and processes. Of course,
process terms such as RUN or BUY presuppose a sequence of events and prior and
posterior states. The term script is often used for a frame/domain with a sequence
of events, following Schank and Abelson (1977). They use the term to describe a
canonical sequence of events presupposed by a social activity such as going to a
restaurant. We subsume scripts under frames/domains.
Another theoretical construct that can be understood as a type of frame or domain
is the so-called ‘theory theory’ of categorization found in cognitive psychology.
Advocates of the theory theory argue that our understanding of categories such
as HORSE or HAMMER is based not on perceptual features but on theories of
biological kinds and artifacts respectively (Murphy and Medin 1985). For instance,
we have at least a folk theory of biological kinds that indicates that individuals
of the same category (e.g. HORSE) are members of that category by virtue of
descent and reproduction, and perceptual similarity of horses (and the distinctness
of individuals of other species) are a result of those basic biological patterns.
Likewise, hammers are defined by the fact that they are manufactured by human
beings for a particular function, and perceptual similarity of hammers (and the
distinctness of other kinds of artifacts) are a result of their intended function. In
frame semantic terms, the base for HORSE includes the ‘theory’ of biological
kinds and the base for HAMMER includes the ‘theory’ of artifacts (see Fillmore
1986a:54).
Fillmore also uses the notion of framing to describe differences in the com-
munity or social domain of use of a word (Fillmore 1982a:127–29). For exam-
ple, he notes that in the legal domain, that is, the community that engages in
legal activity, the concepts of MURDER and INNOCENT differ from those con-
cepts used outside that domain/community. In the legal domain, MURDER is pro-
filed in a frame/domain where it contrasts with MANSLAUGHTER, but outside
that domain, MURDER is profiled in a domain lacking that contrast. In the legal
18 A conceptual approach to linguistic analysis
domain, INNOCENCE is profiled against a frame in which innocence and guilt are
the result of judgements in a trial (and in fact, guilt can be established only after
the completion of the trial). Outside that domain, INNOCENT is profiled against a
frame in which innocence and guilt are defined by whether the person in question
committed the crime or not. Other concepts such as FLIP STRENGTH exist only
in a specialized community, in this case publishers of pornography (the interested
reader may turn to Fillmore 1982a:12 for further details). Hence, frame semantics
is being extended to describe differences that appear to be defined on social rather
than conceptual grounds. But there is a link between the two. Communities are
defined by the social activities that bind the members together. Clark argues that
communities involve the possession of shared expertise among their members: the
specialized knowledge that is acquired by engaging in the activities that define the
community (Clark 1996:102–4). This shared expertise is the conceptual structure
that is found in the frame/domains of the concepts symbolized by the specialized
vocabulary used by members of the community.
The distinction between profile and frame/domain is a useful tool for analyzing a
number of interesting semantic questions. In particular, some distinctions in word
meaning apply not to the profiled concept – what is usually thought of as ‘the
definition’ of a word – but to its frame/domain.
For example, some concepts appear to denote the same thing in the world but
profile it against a different frame. For example, LAND and GROUND denote
(profile) what seems to be the ‘same thing,’ but against different frames: LAND
describes the dry surface of the earth in contrast with SEA, while GROUND de-
scribes the dry surface of the earth in contrast with AIR (Fillmore 1982a:121).
The frame chosen by one word or another allows one to make different inferences:
Fillmore notes that a bird that spends its life on land does not go in the water, but
a bird that spends its life on the ground does not fly (ibid.). Langacker offers the
example of ROE and CAVIAR, both being fish eggs: ROE is profiled against the
frame/domain of the reproductive cycle of fish, while CAVIAR is profiled against
the frame/domain of food preparation/consumption (Langacker 1987:164–65).
Another example is FLESH, profiled against the frame/domain of the body’s
anatomy, vs. MEAT, profiled against the frame/domain of food. The semantic
difference is reflected in the collocations flesh and bones, describing an emaciated
body, and meat and potatoes, describing a bland but filling type of meal (contrast
meat and bones and flesh and potatoes).
The alternative framing of the same profile is particularly common with terms
that are evaluative in character. For example, STINGY profiles one end of a scale,
the opposite of which is GENEROUS; while THRIFTY appears to profile the
same end of the same scale, and its opposite end is profiled by WASTEFUL
(Fillmore 1982a:125). The difference is the orientation of the associated evaluative
Frames, domains, spaces 19
In the examples of frames for mouth in (15), mouth can be thought of as denoting
the same type of profile, namely the opening to a container (however, a cave may
have several openings to the earth’s surface, and the container of a river is defined
by both the riverbed and gravity). The word mouth is generally considered to be
polysemous, that is, it has a sense for each of the profile-base pairings (senses that
may not share the same word in other languages). In other words, the profile alone
is insufficient in defining the senses of mouth.
The profile-frame/domain distinction is particularly useful in understanding
the nature of semantic differences between words and their apparent translation
equivalents in different languages. The profile-frame/domain distinction may shed
light on some aspects of why translation is difficult and often unsuccessful.
20 A conceptual approach to linguistic analysis
One can find frame-based semantic contrasts across languages that are similar to
those found within a single language such as LAND/GROUND. Fillmore contrasts
the English word concept LUKEWARM with the Japanese word concept NURUI.
Both concepts profile the state ‘at room temperature,’ but for English speakers
LUKEWARM is used for liquids that are ideally hot or ideally cold, whereas
for some Japanese speakers NURUI is used only for liquids that are ideally hot
(Fillmore 1982a:121).
Sometimes linguistic differences across languages represent differences in how
much information is specified in the frame, rather than something about the inherent
structure of the profiled concept. English river profiles a more or less permanent
flow of water in a natural channel, presupposing a frame specifying the relevant
topographical features. French contrasts rivière and fleuve. The concepts for both
French words is essentially the same as the concept for the English word – a natural
flowing waterway – but the frame is more specific than the English frame for each
word: fleuve specifies in addition that it is a major waterway that flows into the sea,
unlike rivière. Hence a translation of fleuve as river is partly accurate (the profile),
but not completely so (its frame). Another example is English eat, which profiles
the process of consuming food. German contrasts essen and fressen: both describe
the process of consuming food also, but the former specifies that the eater is human
and the latter that the eater is an animal (nonhuman). This is a framing effect and
is therefore subject to construal (§3.3). That is, one can use the term fressen to
describe the action of a human being, but it frames that action much differently
than essen, leading to a construal of the action as being animal-like (crude, sloppy,
etc.).
The nature of word meaning across languages is sometimes obscured by analysts
who do not distinguish between profile and frame in their word definitions. For
example, some languages are described as having words that correspond to whole
sentences in English. Two candidate examples are given in (16)–(17), the first from
a native American language and the second from a European language:
(16) Alabama ispaspaakáhmit ‘to be shaped into a patty, shaped like a biscuit (said
of the shape of the mixture of brains and moss used for curing hides)’ (Sylestine
et al. 1993:203)
(17) Swedish tura – ‘sitting on the boat going back and forth between Helsingborg
and Helsingør’ (Karina Vamling, pers. comm.)
Rasa has two primary meanings: ‘feeling’ and ‘meaning’. As ‘feeling’ it is one
of the traditional Javanese five senses – seeing, hearing, talking, smelling and
feeling, and it includes within itself three aspects of ‘feeling’ that our view of
the five senses separates: taste on the tongue, touch on the body, and emotional
‘feeling’ within the ‘heart’, like sadness and happiness. The taste of a banana is
its rasa; a hunch is a rasa; a pain is a rasa; and so is a passion. As ‘meaning’,
rasa is applied to words in a letter, in a poem, or even in common speech to
indicate the between-the-lines type of indirection and allusive suggestion that is
so important in Javanese communication and social intercourse. And it is given
the same application to behavioral acts generally: to indicate the implicit import,
the connotative ‘feeling’ of dance movements, polite gestures, and so forth. But
in this second, semantic sense, it also means ‘ultimate significance’ – the deep-
est meaning at which one arrives by dint of mystical effort and whose clarifi-
cation resolves all the ambiguities of mundane existence [etc.]. (Geertz 1973:
134–35)
3 Langacker argues that the locational-configurational distinction applies to domains, but Clausner and
Croft demonstrate that the same domain can support locational and configurational profiles.
Frames, domains, spaces 23
= EGO
NIECE
The relevant part of the kinship system for defining NIECE is called the scope of
predication for the concept (Langacker 1987:118–19; renamed immediate scope
[Langacker 1999:49]).
An example of different scopes of predication can be found in the behavior of
human body parts such as the following parts of the arm (Langacker 1987:119):
KNUCKLE ⊂ FINGER ⊂ HAND ⊂ ARM ⊂ BODY. Each one has its imme-
diate successor as its scope of predication. Possessive constructions referring to
wholes within the scope of predication are acceptable, but if the whole is beyond
the scope of predication, then the sentence is odd (Langacker 1987:119; but see
§6.2.1.7):
Langacker argues that some domains involve more than one dimension
(Langacker 1987:150–51). An obvious case is space, which involves three di-
mensions (some concepts such as CIRCLE need only two dimensions for their
definition; others such as LINE need only one). Many physical qualities that are
grounded in the experience of sensory perception, such as TEMPERATURE and
PITCH, are one-dimensional. Others, such as COLOR, can be divided into HUE,
BRIGHTNESS and SATURATION. Generally, dimensions of a domain are all si-
multaneously presupposed by concepts profiled in that domain. This is the critical
point: a concept may presuppose several different dimensions at once.
In fact, a concept may presuppose (be profiled in) several different domains.
For example, a human being must be defined relative to the domains of physical
objects, living things and volitional agents (and several other domains, e.g. emo-
tion). The combination of domains simultaneously presupposed by a concept such
as HUMAN BEING is called a domain matrix. Langacker makes the important
point that there is in principle only a difference of degree between dimensions of a
domain and domains in a matrix (Langacker 1987:152). In practice, we are more
likely to call a semantic structure a domain if there are a substantial number of
concepts profiled relative to that structure. If there are few if any concepts profiled
relative to that structure alone, but instead there are concepts profiled relative to
that structure and another one, then those structures are likely to be called two
dimensions of a single domain. The term ‘domain’ implies a degree of cognitive
independence not found in a dimension (see also §5.3.1).
The domain structure presupposed by a concept can be extremely complex. Let
us now consider how one would define what seems to be a kind of physical object,
the letter T. It is directly defined as a letter of the alphabet; its base (domain) is
hence the alphabet. The alphabet is itself an abstract domain presupposing the
notion of a writing system – it is not just an instance of a writing system, since
the latter involves not just a set of symbols such as an alphabet but also the means
of putting them together, including the direction of letters on a page, spaces for
words and so on. The domain of writing systems in turn presupposes the activity of
writing. The activity of writing must be defined in terms of human communication,
which presupposes the notion of meaning – perhaps a basic domain, since the
symbolic relation appears not to be reducible to some other relation – and visual
sensations, since writing is communication via visually perceived inscriptions,
rather than auditorily or through gestures. And since writing is an activity, the
domains of time and force or causation (both basic domains; see §3.5) are also
involved in the domain matrix of writing, since the letter T is the product of an
activity. Since writing is a human activity, it presupposes the involvement of human
beings. Human beings are living things with mental abilities, such as volition,
intention, cognition and emotion (themselves dimensions of the mental domain or
26 A conceptual approach to linguistic analysis
better domains in the matrix of the domain of the mind). Living things in turn are
physical objects endowed with life. Physical objects possess material existence and
are spatial entities (although material objects always have spatial extent, spatial
objects like geometric figures can exist without material embodiment).
A diagram exhibiting all of the basic-abstract domain relations presupposed
in defining the concept of the letter T is given in Figure 2.3 (based on Croft
1993[2002]:170, Fig. 2.1; the profiled concept is given in boldface, and the basic
domains are given in capitals).
alphabet
writing system
writing
communication VISION
MATTER SPACE
Figure 2.3 Domain structure underlying the concept of the letter T
From this, it can be seen that it is incorrect to describe the concept of the letter
T simply as belonging to the domain of writing, as a typical informal theory of
domains would most likely have it. The vast majority of concepts belong to abstract
domains, which are themselves profiled in complex domain matrices, often also
abstract and so ultimately presuppose a large array of basic domains that can be
called a domain structure (Croft 1993[2002]:169; this corresponds to Langacker’s
maximal scope [Langacker 1999:49]).
It is not easy to distinguish profile-base relations from taxonomic/schematic
relations (that is, type vs. instance). For example, is writing an instance of human
communication, or is writing an instance of an activity that can only be understood
Frames, domains, spaces 27
in terms of the goals of human communication? Figure 2.3 assumes that the
latter is a more accurate description. Likewise, since writing is an instance of
human activity, human activity does not appear as its domain, but the various
domains that it presupposes – time, change, force, volition – do appear, because
anything presupposed by a human activity will be presupposed by any instance
of it.
It is also difficult to determine direct vs. indirect reference to a domain. The
definition of an arc does not directly presuppose two-dimensional space, but rather
it presupposes a circle which in turn presupposes two-dimensional space. Thus,
an arc is not directly a two-dimensional object per se, but only such by virtue of
being a part of a circle. Likewise, the letter T is not directly a shape, but only such
by virtue of being a letter of the alphabet. But in fact, is the letter T a shape by
virtue of being a letter of the alphabet, or by virtue of being the physical product of
the activity of writing? Figure 2.3 assumes that it is best described as the former,
since the set of symbols is a set of shapes.
Another similar problem in this example is the location of the domain of mental
ability. The activity of writing is a volitional, intentional activity, so it presupposes
the domain of mental ability. But mental ability is presupposed by writing because
writing presupposes human involvement, and the human involvement involves
volition and intention. Determining the exact structure of the array of domains
upon which a profiled concept is based requires a careful working out of the
definitions of concepts.
A further complication in the relation between profiles and domain matrices is
that a word sometimes profiles a concept in only one of the domains in the domain
matrix, or even just a domain deeply nested in the domain structure. The contrast
can be illustrated by the concepts PERSON and BODY. PERSON is profiled against
the abstract domain of HUMAN BEING (along with MAN, WOMAN etc.). The
concept of HUMAN BEING is in turn profiled against the domain matrix of
LIVING THING + MIND: human beings are living things with certain mental
states and abilities (recall the classical definition of man as a rational animal).
LIVING THING is in turn profiled against the domains of PHYSICAL OBJECT
and LIFE: living things are physical objects endowed with life. The concept BODY
represents a person’s physical reality (alive or dead). Its base is nevertheless still
the abstract domain of HUMAN BEING (or more precisely ANIMAL), but it
profiles just the PHYSICAL OBJECT domain in the domain structure underlying
HUMAN BEINGS. Contrast BODY with SOUL, which profiles a nonphysical
domain of a human being, what we have called MIND for convenience; or with
CORPSE, which profiles the PHYSICAL OBJECT domain but also profiles a
particular region in the LIFE domain, namely DEAD.
28 A conceptual approach to linguistic analysis
[19e]). In other words, the ICMs for BACHELOR are going to be as detailed and
as hedged as reality in order to describe the ‘ideal’ life history and lifestyle that is
implied by BACHELOR.
Searle (1979) argues that in fact the frame for any word concept is going to
be infinitely complex. Searle is interested in what he calls the background as-
sumptions for defining the literal meaning of words; in frame semantic terms, the
background assumptions are the frame(s) for understanding the literal meaning of
a linguistic expression. Searle argues that the sort of background knowledge that
is relevant to a linguistic expression’s meaning cannot be enumerated in such a
way that all contexts of use can be predicted. That is, a basically infinite set of
background assumptions are required to characterize the literal meaning of an ut-
terance, and hence its appropriate use in context. Consider the following example
(Searle 1979:127):
(20) Give me a hamburger, medium rare, with ketchup and mustard, but easy on the
relish.
We assume we understand what the meaning of this request is; we invoke a back-
ground frame of fast food restaurants, the ordering and serving of food, how a
hamburger is cooked and garnished, and so on. But there is more to it than that:
These situations are admittedly unlikely to be encountered in real life, in the way
that unmarried men living with their girlfriends or homosexual men commonly are
encountered. Nevertheless, in the frame for ordering a hamburger we would want
to represent the assumptions that it is not supposed to be too large, nor encased in
solid lucite plastic, nor any of an indefinitely large number of other things that one
could do to a hamburger.
Langacker makes a similar observation with a similar type of example, given in
(21) (Langacker 1988:16):
imagine a race over the ocean by helicopter, where the contestants must transport
a severed head, suspended by a rope from the helicopter, from the starting line to
the finish; a contestant is disqualified if the head he is carrying ever dips below
the water’s surface. (Langacker 1988:16–17)
30 A conceptual approach to linguistic analysis
In other words, we have to bring to bear our full knowledge of the way the world
is or, more accurately, the way we expect the world to be, in order to describe the
precise meaning of an utterance.
Another way of saying this – the more common way of saying it in cognitive
linguistics – is that we have to call on our encyclopedic knowledge in order to
properly understand a concept. Some semanticists have argued that only a small
subset of our knowledge of a concept needs to be represented as the linguistic
meaning of a word; this is known as the dictionary view of linguistic meaning.
But the frame semantic model of linguistic meaning highlights the failings of
the dictionary view (Fillmore 1982a:134; 1985:233). The dictionary view fails
because it generally describes only the concept profile, or at best a very simplified
version of the concept frame implicit in a concept profile (see Haiman 1980 for
further arguments; see also Quine 1951[1961]). Once one begins to specify the
conceptual structure of the frame that supports the concept profile for a word or
linguistic expression, the semantic structure quickly expands to encompass the
total (encyclopedic) knowledge that speakers have about the concept symbolized
by the word or construction.
Of course, encyclopedic knowledge is all interconnected in our minds. If the
meaning of a word includes the frame as well as the profile, then one must abandon
the concept of word meanings as small discrete chunks of conceptual structure.
Langacker proposes an alternative model of the meaning of a word as an access
node into the knowledge network (Langacker 1987:161–64):
from the ICM. For example, Lakoff describes the ICM for mother as involving a
cluster of several different ICMs (Lakoff 1987:74–76):
The cluster ICM (as Lakoff names it) is essentially a domain matrix. Thanks
both to modern medicine and to traditional social arrangements, the real world
has many cases where only parts of the cluster model for MOTHER applies to
particular individuals. These deviations from the cluster ICM are indicated by
conventional compounds and adjective + noun expressions:
(23) a. stepmother: fits the NURTURANCE and MARITAL models but none of the
others
b. foster mother: fits the NURTURANCE model but none of the others
c. birth mother: fits the BIRTH model but none, or not all, of the others
d. genetic mother: fits the GENETIC model but not all of the others
e. unwed mother: fits (probably) all but the MARITAL model [etc.]
Nevertheless, one might still obtain varying results if asking of individuals falling
under any of the categories in (23) whether she is the ‘real mother’ of the child
(see chapter 5).
In other cases, there is clearly an ICM but linguistic convention has allowed the
word, unmodified, to describe situations that lack some of the properties of the
ICM. Fillmore gives the example of the ICM for breakfast, which has as its frame
a cycle of meals, and profiles ‘the one which is eaten early in the day, after a period
of sleep, and for it to consist of a somewhat unique menu’ (Fillmore 1982a:118).
But you can work through the night and have eggs, toast and so on at sunup and
call it breakfast; you can sleep till 3pm, get up and have eggs, toast and so on
and call it breakfast; and you can sleep through the night and in the morning have
chocolate cream pie and a martini and call it breakfast (ibid., 118–19). Each of
these cases lacks one feature of the ICM for BREAKFAST. One can also call a
meal breakfast that lacks both ‘early in the day’ and ‘after a period of sleep’ too:
restaurants exist that serve breakfast all day (ibid.; the menu feature appears to be
more important than the other two).
Another example similar to BREAKFAST is the ICM for lie (Coleman and
Kay 1981). The ICM for LIE, such that a speaker S telling an addressee A the
proposition P is a lie, is:
32 A conceptual approach to linguistic analysis
(24) a. P is false.
b. S believes P to be false.
c. In uttering P, S intends to deceive A.
Coleman and Kay performed an experiment with stories designed to test every
combination of the features listed in (24), and found that, in general, the situations
with more of the three properties (24a/b/c) tended to be described by experimental
subjects more often as lies than situations with fewer of the properties. Two situa-
tions (at least) have conventional expressions that indicate their deviation from the
ICM. Polite social lies such as saying What a lovely party! or How nice to see you!
can be said in circumstances in which (24a/b) hold but (24c) does not. The other
situation can be illustrated with the exchange in (25) (Coleman and Kay 1981:29):
In the situation in (25), (24a/b) do not hold but (24c) does (just the opposite of social
lies); in this situation an English speaker could say that Mary is being economical
with the truth.
In the case of breakfast and lie, the word profile extends to a range of situations
whose features vary. Nevertheless, there appears to be agreement as to the situation
that counts as the ICM for these words. ICMs thus give rise to judgements of graded
centrality to members of a category, a phenomenon that is usually described as
prototype effects (see Lakoff 1987 and chapter 4).
is also represented, but it has a different status: the event has not taken place, it is
only something in Gina’s mind. In fact, even the sports car may exist only in Gina’s
mind. Example (26e) is more similar to (26d) than (26a), even if it is a prediction
about the real world: the event has not taken place. Finally, in example (26f) the
event is again hypothetical, and so is the event described in the consequent clause.
In a truth-conditional semantics, (26a) is unproblematic, but (26b–f) are. The
situation ‘Gina has bought a sports car’ is false in (26d–e), but not necessarily false
in (26b–d) or even (26g). One must be able to distinguish between the status of
situations depending on whether they are true in the real world, or whether they
are only true in someone’s beliefs or desires, or true at another time in the real
world.
In a truth-conditional semantics, the standard way of representing the status of
situations is as possible worlds: there is the real world, and then there are worlds
with situations that are possible but not (necessarily) actual. Possible worlds are
then identified with a person’s beliefs or wishes or some other mental attitude. Pos-
sible worlds pose metaphysical problems for many people, however. Do possible
worlds exist? If so – or especially if not – where are they?
Fauconnier (1985, 1997; see also Fauconnier and Sweetser 1996) proposes an
alternative model of representing the status of knowledge that is metaphysically
more attractive and allows for elegant solutions to a number of problems in semantic
and pragmatic analysis. Fauconnier replaces the notion of a possible world with
that of a mental space, and argues that the mental space is a cognitive structure.
That is, the allocation of a situation to ‘Gina’s desire,’ ‘Paolo’s belief’ or ‘The
hypothetical situation’ is done in the mind of the speaker (and hearer), not in
some as yet unclear metaphysical location. Fauconnier then proposes a set of
principles for the interpretation of utterances and the assignment of situations to
the appropriate mental space. We briefly present Fauconnier’s model and a number
of examples here; the reader should consult his work for detailed arguments in favor
of his model over truth-conditional approaches to the same phenomena.
Utterances such as (26a) are normally construed as situating events or states in
a base space (Fauconnier 1997:38–39), normally the present reality (more pre-
cisely, the mutually known world of the interlocutors; in Fauconnier 1985 this is
called the reality space). Utterances such as (27b–f) have elements that Fauconnier
describes as space builders: included in their meaning is the setting up of a new
space different from the base space and linked to it. Space builders include a
wide range of semantic phenomena corresponding not only to possible worlds in
logical semantics but also a variety of other operators, including temporal expres-
sions ([27a]; see Fauconnier 1985:29–30, 33–34; Fauconnier 1997, chapter 3),
image or ‘picture noun’ contexts ([27b]; Fauconnier 1985:10–12), fictional situ-
ations ([27c]; ibid., 73–81), games and other systems ([27d]; ibid, 31), negation
34 A conceptual approach to linguistic analysis
and disjunction ([27e–f]; ibid., 92, 96–98) and the separate cases in quantification
([27g]; see Fauconnier 1986):
All of these examples have in common the building of a mental space in which
a situation is held to be ‘true’ in that space only. More generally, we can say
that just as words and constructions evoke semantic frames/domains, words and
constructions also build spaces; at the very least, they ‘build’ or evoke the base
space. The relevant word or construction then conventionally specifies that the
asserted situation holds in the appropriate space.
Between the base space and any built space, there must be a mapping of the
elements found in each space. Many interesting and puzzling semantic and prag-
matic phenomena are a product of the possible mappings between spaces. We
may divide the phenomena found in the mapping into two parts. First, what do
the named elements of the built space (e.g. Gina and sports car) correspond to,
if anything, in the base space? Second, what conceptual structures from the base
space also occur in the built space(s), and vice versa? We begin with the first
question.
In (26d), it seems straightforward to say that the person named Gina in the desire
space built by Gina wants . . . is mapped onto Gina in the base space. But the object
described as a sports car may or may not correspond to anything in the base space:
Gina may have seen a particular car on the lot, or she may not have any specific
car in mind. This is the distinction between a specific and nonspecific reading,
respectively, of a sports car in (26d). The specific and nonspecific readings are
represented in Figure 2.4.
Fauconnier crucially distinguishes between roles and values in mappings be-
tween spaces. A role is a linguistic description describing a category; a value is an
individual that can be described by that category. Roles can be a category or type
with various instances or tokens; sports car is such a role, since there are many
instances (values) of sports cars. A role can also be a category that is filled by a
single individual at one time but by different individuals over time; the President
of the United States is an example of such a role. Roles and values are specific to
a single mental space, and all counterpart relations between roles and values in
different spaces must be established cognitively by the interlocutors.
Frames, domains, spaces 35
sc sc sc sc
x x' x'
G R G
R
Figure 2.4 easily represents the difference between the specific and nonspecific
readings. In the specific reading of (26d), the value x of a sports car in Gina’s
want space G has a counterpart value x in reality (the car she saw on the lot). In the
nonspecific reading, there is no counterpart value in reality: she imagines a sports
car she wants, but has not identified it with any existing car.
One of Fauconnier’s central insights is that many puzzling semantic phenomena
are the result of the fact that a value in one space can be described by the role its
counterpart in another space has, even if that role is invalid for the value in the
first space. This is the Access Principle (Fauconnier 1997:41; it is called the ID
Principle in Fauconnier 1985:3). For example, (28) is not contradictory (Fauconnier
1985:29):
The value in the 1929 temporal space – the blonde girl – is being described with
a role, lady with white hair, from the base space (the current reality).
Armed with the distinctions between mental spaces, between roles and values
within spaces and across spaces, and the Access Principle, Fauconnier goes on to
explain a wide range of semantic and pragmatic phenomena using these distinc-
tions. Only a selection of these can be described here.
The phenomenon described as referential opacity is illustrated in 29:
In the Greek myth, (29) is true under one reading (his mother = ‘the person you
and I know is Oedipus’ mother’) but false under another (his mother = ‘the person
Oedipus believes is his mother’). This distinction is due to the fact that Oedipus
does not know that Jocasta is his mother. In mental space terms, the individual value
named Jocasta does not fill the role his mother in Oedipus’ belief space, although
she does in reality space; see Figure 2.5 (adapted from Fauconnier 1985:49).
36 A conceptual approach to linguistic analysis
om om
j j′
R O
,
om: Oedipus mother (role)
j: Jocasta (value in R)
j′: Jocasta (value in O)
R: reality space
,
O: Oedipus belief space
In the true reading of (29), the description his mother for j in R is used for the value
j in O by the Access Principle. The false reading of (29) uses the description his
mother in O, but it does not apply to j in O. (A similar analysis can be applied to
referential/attributive ambiguities; Fauconnier 1985:159–60.)
Only a few further examples can be given of how Fauconnier’s model handles
a variety of complex reference and identity phenomena (Fauconnier 1985:45, 36,
32, 39, 31, 155):
asserted in the hypothetical space. On the other hand, at least other information
about Gina, and about sports cars, not to mention much other knowledge about the
world, may be attributed to the hypothetical space.
Fauconnier first addresses the question of presuppositions (Fauconnier 1985,
chapter 3). As noted in §2.1, presuppositions are situations that are part of the
frame of a concept, but are not asserted. The question is, what is the relationship of
presuppositions in a built space to those in the base space? For example, consider
the sentences in (31) (Fauconnier 1985:89–90):
(31) a. If Max has gone to the meeting, then Max’s children are alone.
b. If Max has children, Max’s children are American.
The phrase Max’s children presupposes that Max has children; that is, a referring
expression presupposes the existence of its referent(s). The traditional pragmatic
analysis is that one must determine the presupposition of the whole sentence from
the presuppositions of its parts (presupposition projection; see, e.g., Levinson
1985:191–225). In (31a), the presupposition that Max has children ‘projects’ to the
base space. But in (31b), it does not project because it is asserted in the antecedent
clause: Max may or may not have children in the base space (reality).
Fauconnier instead introduces the principle of presupposition float: ‘informally:
a presupposition floats up [from a built space to its base space] until it meets itself or
its opposite’ (Fauconnier 1997:61). In example (31a), the built space presupposes
that Max has children but does not assert it. Hence the presupposition can float to
the base space. In example (31b), however, the built space asserts that Max has
children, and hence the presupposition cannot float beyond it to the base space.
Two more complicated examples are given in (32)–(33) (Fauconnier 1985:95,
93):
(32) It is possible that John has children, and it is possible that John’s children are
away.
(33) Luke believes it’s raining and hopes that it will stop raining.
Examples (32)–(33) demonstrate that space builders may build the same space
or related nonreal mental spaces. In (32), it is possible in the second conjunct
can be construed as evoking the same possibility space that was built in the first
conjunct. In this case, the presupposition that John has children is asserted in the
hypothetical space in the first conjunct and therefore it does not float to the base
space. Example (33) demonstrates that certain built spaces are related in privileged
ways that allow presuppositions to float (see also McCawley 1981[1993]:415–30).
A hope for some situation can be built on one’s beliefs. Hence the presupposition
that it is raining in the second clause of (33) is built on the assertion in the first clause
(and therefore does not float to the base space). Reversing the relation between the
38 A conceptual approach to linguistic analysis
(34) ??Luke hopes that it is raining and believes that it will stop raining.
(35) If Boris had not come, Olga would have come anyway.
(36) a. If Napoleon had been the son of Alexander, he would have been Macedonian.
b. If Napoleon had been the son of Alexander, he would have won the battle of
Waterloo.
c. If Napoleon had been the son of Alexander, he would not have been Napoleon.
d. If Napoleon had been the son of Alexander, Alexander would have been
Corsican.
Fauconnier writes:
It would not make sense to evaluate the ‘absolute’ truth of any of these statements,
but they can all be used to make some point, which requires only very partial
structuring of H [the counterfactual space] . . . such examples suggest that there is
no general linguistic algorithm for going from R [the base space] to H. (Fauconnier
1985:118)
(37) At this point, Great America II is barely maintaining a 4.5 day lead over Northern
Light.
Obviously, Northern Light is nowhere to be seen in 1993, but the blend of the
1853 temporal space and the 1993 temporal space in (37) ‘makes a point’ about
the progress of Great America II. In Fauconnier and Turner’s blending theory,
example (37) evokes four mental spaces: two input spaces (in [37], the 1853 and
1993 temporal spaces); a generic space, which abstracts the commonalities from
the two spaces (the route of travel, distance traversed, time taken etc.) and thereby
defines the cross-space mapping between the elements in the two input spaces; and
a blended space, which creates a novel expressive effect, in this case an image of
a race between the current boat and a boat from the nineteenth century. Fauconnier
and Turner (2002) argue that blending is a process of space mapping that pervades
human reasoning, and explore the phenomenon of blending in a wide range of
phenomena, most notably metaphor.
At this point, blending theory has moved quite a distance from mental space
theory. Mental space theory illustrates how utterances evoke not just semantic
frames but also spaces representing the status of our knowledge (beliefs, desires,
hypotheticals, counterfactuals) relative to reality, how language uses links between
different spaces in referring to individuals, and how knowledge can float between
spaces. Blending theory has shifted the focus to how information from two spaces,
construed broadly to include domains, is combined to produce novel conceptual
structures. This aspect of blending theory is discussed with respect to metaphor in
§8.3.3. In this chapter we have focused on the fact that the original mental space
theory describes a significant dimension for the structuring of our conceptual
knowledge orthogonal to semantic frames/domains, and offers solutions to many
semantic and pragmatic problems in addition to those illustrated in this section.
3
Conceptualization and
construal operations
3.1 Introduction
40
Conceptualization and construal operations 41
Examples (1a–b) and (2a–b) differ in the choice of nominal and verbal inflection,
(1a–b) by plural count noun and mass noun, and (2a–b) by the choice of a simple vs.
a progressive form. Examples (3a–b) differ in the choice of a function word, in this
case a preposition, and a reversal of subject and prepositional complement choice.
Examples (4a–b) differ derivationally in part of speech, between a verb and its
derived noun. Examples (5a–b) and (6a–b) differ in the grammatical construction
used to describe the scene, active vs. passive voice in (5) and presentational vs.
ordinary declarative in (6). All of these sentences seem to be truth-functionally
equivalent. But English is not being unnecessarily profligate here: the a and b
members offer a different conceptualization of the experience in every case. These
and other examples will be explicated in the following sections.
Similar evidence is found in cases when the same word can be used with two
different inflections, derivations or constructions, and there are subtle but definite
conventional truth-functional differences in the two uses:
In (7a), the mass noun refers to a substance, but in (7b), the same noun as a count
noun refers to an object which is covered with the substance but may or may not
be filled with the same substance. In (8a), the simple present describes a behav-
ioral trait of Ira’s, while (8b) describes a particular activity of his. Examples (9a)
and (9b) differ in Timmy’s position relative to the speaker and the tree; to make
(9a) and (9b) truth-functionally equivalent, the speaker would have to move to the
other side of the tree. Example (10a) describes a behavioral trait of Jill, but (10b)
describes the same trait as a constant (and annoying) aspect of Jill’s personality.
In example (11a), the object role of the bone indicates that the bone itself is being
affected by the dog’s action, but in (11b), the oblique role indicates that it is only
meat and gristle on the bone that is being affected.
The examples in (7)–(11) are equally indicative of the role of construal in lan-
guage as the examples in (1)–(6). The truth-functional differences in (7)–(11) are
indicative of favored conceptualizations that have led to the extension of a partic-
ular construal to a situation that does not (easily) allow for alternative construals.
For example, the fact that certain sweets come in individuated units allows for a
construal of the substance noun chocolate as a count noun (compare I’d like an
orange juice, please), and chocolate as a count noun is extended to other such
sweets where only the (perceptually salient) outer surface is made of chocolate.
But both a chocolate and an orange juice share the construal as an individuated
unit, even if the link to the substance is more tenuous in the former case. In fact, in
cognitive linguistics conceptualization is the fundamental semantic phenomenon;
whether alternative construals give rise to differences in truth conditions or not is
a derivative semantic fact.
In some cases where construal is accompanied by a truth-functional seman-
tic shift in meaning, English allows speakers to express the construal-plus-shift
overtly, as in:
Example (12) uses the process verb act and a manner construction to denote
the activity and (13) uses a container word glass and the partitive construction to
denote the individuated amount of juice. This is not always possible in English; for
example, a chocolate does not have an obvious overt expression of its individuation,
and there is no simple overt expression of the different types of chewing actions
in (11a–b).
Other languages require overt expression of construal-plus-shift, in contrast to
English. For example, Vietnamese cannot simply use a mass noun in a count noun
construction or vice versa, as English does in (7); and Russian requires overt
marking in the following examples were English does not:
Conceptualization and construal operations 43
is given in (17):
(17) I. Selection
II. Perspective
A. Figure/Ground
B. Viewpoint
C. Deixis
D. Subjectivity/Objectivity
III. Abstraction
An inventory of image schemas drawn from Johnson 1987 and Lakoff and
Turner 1989 is given in (18) (based on Clausner and Croft 1999:15; the headings
and items in italics were added by Clausner and Croft).
(18) SPACE UP-DOWN, FRONT-BACK, LEFT-RIGHT,
NEAR-FAR, CENTER-PERIPHERY, CONTACT
SCALE PATH
CONTAINER CONTAINMENT, IN-OUT, SURFACE,
FULL-EMPTY, CONTENT
FORCE BALANCE, COUNTERFORCE, COMPULSION,
RESTRAINT, ENABLEMENT, BLOCKAGE,
DIVERSION, ATTRACTION
UNITY/MULTIPLICITY MERGING, COLLECTION, SPLITTING,
ITERATION, PART-WHOLE, MASS-COUNT,
LINK
IDENTITY MATCHING, SUPERIMPOSITION
EXISTENCE REMOVAL, BOUNDED SPACE, CYCLE,
OBJECT, PROCESS
Image schemas are also construals of experience, though they exhibit some of
the characteristics of domains as well (see §3.5).
In this chapter, we present all of the construal operations and image schemas
discussed by cognitive linguists under a new classification.3 This classification is
given in Table 3.1 on page 46.
A chief aim of this classification is to demonstrate the close relationship be-
tween construal operations proposed by linguists and psychological processes
proposed by cognitive psychologists and phenomenologists. If linguistic construal
operations are truly cognitive, then they should be related to, or identical with,
general cognitive processes that are postulated by psychologists. In fact, most if
not all of these construal operations are special cases of general cognitive processes
described in psychology and phenomenology. This view follows from the basic
hypothesis of cognitive linguistics that language is an instance of general cognitive
abilities.
The classification of construal operations in Table 3.1 is not intended to be
a reduction of construal operations to just four processes. The various construal
operations listed under the four headings are all distinct cognitive processes. The
analysis we propose is that the various construal operations are manifestations of
the four basic cognitive abilities in different aspects of experience. The remainder
of this chapter describes and illustrates the construal operations under these four
headings.
I. Attention/salience
A. Selection
1. Profiling
2. Metonymy
B. Scope (dominion)
1. Scope of predication
2. Search domains
3. Accessibility
C. Scalar adjustment
1. Quantitative (abstraction)
2. Qualitative (schematization)
D. Dynamic
1. Fictive motion
2. Summary/sequential scanning
II. Judgement/comparison (including identity image schemas)
A. Categorization (framing)
B. Metaphor
C. Figure/ground
III. Perspective/situatedness
A. Viewpoint
1. Vantage point
2. Orientation
B. Deixis
1. Spatiotemporal (including spatial image schemas)
2. Epistemic (common ground)
3. Empathy
C. Subjectivity/objectivity
IV. Constitution/Gestalt (including most other image schemas)
A. Structural schematization
1. Individuation (boundedness, unity/multiplicity, etc.)
2. Topological/geometric schematization (container, etc.)
3. Scale
B. Force dynamics
C. Relationality (entity/interconnection)
3.2 Attention/salience
model of the mind. The phenomenon of attention focuses on the human cogni-
tive ability involved, but there are also natural properties of phenomena in the
perceived world that lend themselves to being attended to by human beings, and
these properties are said to enhance those phenomena’s salience to human beings’
attention.
Attention is a complex psychological ability whose different aspects can be most
easily illustrated by visual ability: one can select one object or another to focus
one’s attention on; focus of attention is surrounded by a scope of attention; one
can take a more coarse-grained or more fine-grained view of a scene; and one can
fix one’s gaze on a scene or move one’s eye over it. These four aspects of attention
are found across all domains of thought.
3.2.1 Selection
The focal adjustment of selection is our ability to attend to parts of our
experience that are relevant to the purpose at hand and ignore aspects of our
experience that are irrelevant. The phenomenon of profiling a concept in a semantic
frame, described in detail in chapter 2, is an example of selection. In most cases,
different words in a semantic frame or domain focus our attention on the different
elements in the frame, for example radius, arc, circumference in the CIRCLE
frame. In other cases, derivational morphology shifts the profile, as in writer,
whose -er suffix shifts the profile of write from process to agent. The participant
that the -er suffix selects is not fixed to a single participant role but depends on
salience, manifest both in conventionalized forms such as stapler (the instrument)
or in novel forms such as clapper (Jane T., describing a lamp that turns on when
you clap your hands).
Selection of the profile by a single underived word stem is also flexible and
subject to construal. For example, many English nouns are also used as verbs
(Clark and Clark 1979): pan can be construed as profiling either a metal object or
a process in the GOLDSEEKING frame. Both the process and the metal object
are salient in this frame, hence the choice of one word for both. Likewise, British
English speakers can construe bin as profiling either a wastebasket or the action
of tossing something into the wastebasket.
Such examples are not usually analyzed as examples of construal since the
profile is of course central to a word’s meaning and any shift in profile has truth-
functional consequences. However, two semantic processes that involve subtler
and/or more systematic shifts in profile lend themselves to a construal analysis.
The first example is the highlighting of different facets (see chapter 5) or domains
in a domain matrix, as in (19)–(22):
48 A conceptual approach to linguistic analysis
active zone. Thus, in (25) the meaning of heard is ‘S b j heard the sound of Obj’ –
compare She heard the sound of the piano – and the active zone in the verb meaning
is the sound emitted by the object referent. In (26) the meaning of be in (the phone
book) is ‘S b j’s name is printed in Loc’, and the active zone is a name and its
relation to the subject referent.
Langacker argues that the difference between (27a–c) is due to alternative con-
struals of the adjective, not a syntactic alternation:
notes that in (25)–(26), and (27b–c) as well, the effect of the semantic shift is
to allow a more salient entity to be the semantic as well as syntactic argument
of the verb (I instead of my name, the piano instead of the sound of the piano,
Monopoly instead of to play Monopoly; Langacker 1991b:193). Nunberg argues
that the primary ‘pragmatic’ constraint on predicate transfer is the noteworthiness
of the relationship of the predicate to its argument in the context – in cognitive
linguistic terms, its salience in the semantic frame. For example, I’m in the Whitney
Museum (said by an artist about her painting) confers a noteworthy property to the
artist of being represented in a major museum; the same artist saying ??I’m in the
second crate on the right does not (Nunberg 1995:113). Nunberg also notes that
noteworthiness is subject to construal: a jealous painter might say Those daubers
get one-person shows while I’m relegated to a crate in the basement (ibid., 129,
n. 7).
(29) A: We’ve found every bit of the body, sir – even the knuckles.
B: How many did you find?
A: Twenty-seven, sir.
B: Come on, now! How many knuckles does a body have?
A: Oh, you’re right, sir. Twenty-eight.
(30) The money is in the kitchen, under the counter, in the lefthand cabinet, on the top
shelf, behind the meat grinder.
Each locative expression profiles an entity in the scope defined by the preced-
ing locative expression (i.e., the locative expression defines successively nar-
rower search domains; Langacker 1987:286). Scrambling the order of locative
Conceptualization and construal operations 51
A third person pronoun such as them construes the referent as being in the focus
of attention of the hearer, which is appropriate in (32) because the details have
just been mentioned. However, when details is first uttered, the details are not in
the focus of attention – they have not been mentioned as such – but they are in
the scope of attention – the description in the preceding intonation units has been
about the details of the film. The choice of the proximal demonstrative adjective
these details construes the details as being in the hearer’s scope of attention but
not in focus (Gundel et al. 1993:275).
Langacker proposes a highly generalized concept of scope in terms of the do-
minion made accessible by a reference point which functions as the (initial)
focus of attention (Langacker 1999, chapter 6). Langacker argues that reference
point and dominion constitute the construal underlying the possessive construc-
tion: the ‘possessor’ in examples such as my watch, your anxiety and Lincoln’s
assassination functions as a reference point to establish a dominion in which the
appropriate referent of the head noun can be selected. Langacker also extends his
analysis to metonymy. For example, in (20c) above, the speaker focuses the hearer’s
attention on the city of Paris by using Paris. Paris then functions as a reference
point whose scope or dominion includes its inhabitants; the inhabitants of Paris
are therefore accessible as the subject referent for the predicate. Choosing The
people of Paris instead of Paris in (20c) would construe the situation differently,
putting the people in the focus of attention instead of in the dominion of another
focus.
Examples (33a–b) could describe the same scene, but (33b) invites the hearer
to attend to the thickness of the vegetation in the field by using a preposition
requiring a three-dimensional volume; (33a) instead construes the field as a two-
dimensional surface without thickness. To describe the conceptualization involved
here, a metaphor of magnification or granularity is often used. Example (33a)
offers a coarse-grained view of the field, seen as if from a distance so that the
thickness of whatever covers the field is invisible to us. Example (33b) offers a
fine-grained view of the field, as if our view was magnified to reveal its thickness.
Examples (34a–c) provide a more elaborate example of granularity:
(34) a. We drove along the road.
b. A squirrel ran across the road.
c. The construction workers dug through the road.
4 Langacker calls this ‘abstraction,’ but that term is used for such a wide range of theoretical concepts,
even in cognitive linguistics, that we choose a more precise term here.
Conceptualization and construal operations 53
The road is not actually going anywhere, but it is conceptualized as if it is: the
mind’s eye, so to speak, represents one as going along the road. Talmy describes
this as fictive motion, because it is a construal of a static scene in dynamic terms.
Of course, most of the time speakers construe static scenes statically and dynamic
scenes dynamically; this underlies the distinction between state and process in
predicates (see §3.6). But example (35) and many similar examples demonstrate
that this semantic property is subject to construal.
Langacker also makes use of the static/dynamic attentional contrast in construal,
but for a different phenomenon. One of the basic conceptual distinctions between
predicates (prototypically verbs) and arguments or modifiers (nouns and adjec-
tives), according to Langacker, is the mode of scanning of the scene. Langacker
distinguishes summary scanning, a holistic conceptualization of a scene in its
entirety, and sequential scanning, a scanning of a scene in conceived time, which
is not the same as objective time (Langacker 1987:144–45, 248–49). For exam-
ple, when a verb predicates an action as in Boston Bridge collapsed, the event is
54 A conceptual approach to linguistic analysis
3.3 Judgement/comparison
3.3.1 Categorization
Perhaps the most fundamental judgement of comparison is categorization,
which was described in terms of framing in chapter 2. The act of categorization –
applying a word, morpheme or construction to a particular experience to be com-
municated – involves comparison of the experience in question to prior experiences
and judging it to belong to the class of prior experiences to which the linguistic
expression has been applied. There are many ways in which a situation can be com-
pared and judged to be like a prior experience. As we saw in §2.3, the choice of a
linguistic category based on comparison to a prior situation frames – construes – the
Conceptualization and construal operations 55
current situation in different ways, as in fetus vs. unborn baby or thrifty vs. stingy.
In addition to the flexibility of framing a situation by comparing it to one or an-
other prior situation, speakers also have the flexibility of comparing the current
situation to a prior one and in effect redefining the frame. For example, upon en-
tering a holding pattern over Milan airport, a pilot said We’ll be on the path they
call a racetrack; that’s essentially a circle with two straight sides – a significant
reconceptualization of the category CIRCLE.
Langacker describes the comparison process between the current situation and
the category to which it is assigned as sanction (Langacker 1987:66–71). He
recognizes a gradient between full sanction – unproblematic subsumption of the
new situation – and partial sanction – a more creative extension of the category
to the current situation. Categorization involves schematization (§3.2) as well as
judgement: in comparing the new experience to prior ones and categorizing it
in one way over another, we attend to some characteristics and ignore others.
The pervasiveness of construal in the process of categorization has already been
discussed in chapter 2 and will be described in detail in chapter 4.
3.3.2 Metaphor
Another construal operation widely discussed in cognitive linguistics,
metaphor, also involves judgement or comparison. Metaphor involves a relation-
ship between a source domain, the source of the literal meaning of the metaphor-
ical expression, and a target domain, the domain of the experience actually being
described by the metaphor. For example, to waste time involves comparing TIME
(the target domain) to MONEY (the source domain) in the metaphor represented
by the Lakoffian formula t i m e i s m o n e y (Lakoff and Johnson 1980). Time is
construed as a valuable asset that is possessed by human beings and can be ‘used’
in the same way that money is.
The choice of metaphor to describe a situation in a particular domain construes
the structure of that domain in a particular way that differs depending on the
metaphor chosen. For example, the metaphor in stockmarket crash construes the
low level of the market as abnormal, the result of defective operation, whereas
a high (or rising) market is normal. On the other hand, stockmarket correction
construes the low level of the market as normal, its correct level, whereas the high
level is abnormal.
The exact relationship between the source and target domains in a metaphorical
expression is a matter of debate within cognitive linguistics. Metaphor, like cate-
gorization, is sufficiently important for conceptualization to merit its own chapter
in this book (chapter 8), and is discussed in greater detail there.
56 A conceptual approach to linguistic analysis
The figure and ground are asymmetrical. Although near is a spatially symmet-
rical preposition, (40b) sounds odd compared to (40a) (Talmy 2000:314):
Talmy identifies the following properties of objects that favor figure or ground
construal, in the narrower domain of spatial relations (based on Talmy 1983:230–
31; see Talmy 2000:315–16):
(42) Figure Ground
location less known location more known
smaller larger
more mobile more stationary
structurally simpler structurally more complex
more salient more backgrounded
more recently in awareness earlier on scene/in memory
favoring contexts can also be overridden for the opposite figure-ground construal,
with appropriate contextualization, as in (44):
The two events could be coextensive, but since dreaming is contingent on sleeping,
sleeping must function as the ground and therefore (46a) is acceptable while (46b)
is not.
For most figure-ground subordinators, there is no natural inverse for the figure-
ground relation specified by the subordinator (Talmy 2000:326):
In (49a), Tom’s resignation is presumed to let loose the forces of chaos; whereas in
(49b), Tom succeeded in cutting out when he saw what was happening (or perhaps
before the consequences of his actions became apparent to everyone).
58 A conceptual approach to linguistic analysis
3.4 Perspective/situatedness
Perspective, especially deixis, is perhaps the most obvious and most com-
mented upon of the construal operations. Particularly for spatial descriptions, per-
spective is essential, and its dependence on the relative position and viewpoint of
the speaker is well known. But perspective is also found in nonspatial domains:
we have a perspective based on our knowledge, belief and attitudes as well as
our spatiotemporal location. The closest cognitive property to perspective taken
broadly is probably the philosophical notion of our situatedness in the world in a
particular location – where location must be construed broadly to include temporal,
epistemic and cultural context as well as spatial location. This broad interpreta-
tion of location is related to what the phenomenological philosopher Heidegger
calls Being-in-the-world. Heidegger argues that Being-in-the-world is more than
simple spatial inclusion; rather, it is the fundamental situatedness of existence in
5 Despite this, Langacker subsumes figure-ground under perspective among his focal adjustments. In
fact, Langacker himself gives an argument to show that figure-ground alignment is conceptually
independent of foreground-background perspective (see §3.4 below). Langacker also argues that
figure-ground is conceptually distinct from focus of attention, which suggests that the figure-ground
distinction does not belong under the general category of attention either (pace Talmy 1988a:195;
Talmy does not include figure-ground in the schematic systems in Talmy 2000).
Langacker makes greater use of the concepts trajector and landmark. A trajector is defined as
the figure in a relational profile (Langacker 1987:217; see §3.5 for the definition of relationality);
landmarks function as grounds to the trajector.
Conceptualization and construal operations 59
3.4.1 Viewpoint
It is easiest to begin illustrating perspectival construals with spatial ex-
amples. Langacker proposes viewpoint as a focal adjustment with two subtypes:
vantage point and orientation (Langacker 1987:122–26). Vantage point was il-
lustrated in (9a–b) in §3.1: the description of Timmy’s position as being in front
of the tree or behind the tree depends on the vantage point of the speaker.6 A
particular vantage point imposes a foreground-background alignment on a scene
(ibid., 124–25).7 Alternative construals of Timmy’s position are achieved simply
by the speaker moving to another position – that is, the linguistically expressed
spatial relation is dependent on the speaker’s situatedness. Vantage point is sensi-
tive to construal; in (51), it is the vantage point of the addressee (at the relevant
future time) that is used to interpret behind:
(51) Follow my instructions carefully. Enter the woods by the south gate. Follow the
path until you come to the big oak tree. You will find the box behind it.
3.4.2 Deixis
Deixis is the phenomenon of using elements of the subject’s situated-
ness – more specifically, the subject qua speaker in a speech event – to designate
something in the scene. Deixis has been widely studied (see, e.g., Levinson 1983,
chapter 2), and we focus on deixis as construal here. Person deixis – the pronouns
6 This analysis applies only to the situational use of in front of/behind. In sentences such as (9a–b), the
choice of preposition is determined purely situationally, by the relative positions of speaker, Timmy
and tree. In a sentence such as The cat is in front of the house, there is available another interpretation
in which the house has an inherent orientation such that the side with the main entrance is the front
side, regardless of the speaker’s relative position. The remarks in this paragraph refer only to the
purely situational interpretation.
7 A number of linguists have argued that clausal subordination represents a foreground-background
distinction, but it appears to be better analyzed as a figure-ground distinction (Talmy 1978, 2000;
Reinhart 1984; Croft 2001).
60 A conceptual approach to linguistic analysis
I, you, he/she/it, we and they – are only defined relative to who is speaking, and this
variation is an example of alternative construals defined by the speech act situa-
tion (§2.1). Likewise, deictic demonstratives such as this and that, and deictic time
reference such as present and past tense, are only defined relative to the location
and time of the speech event.
In addition to the relativity of spatiotemporal reference to the situation or per-
spective of the speech event, it is possible to construe another time and place as
the deictic center. Examples (52a) and (52b) are construed with a deictic point of
view at a place and time in the narrative; (52c) is bizarre because the point of view
clashes with the reported information (Fillmore 1982b:262–63):
(52) a. He was coming up the steps. There was a broad smile on his face.
b. He was going up the steps. There was a wad of bubblegum on the seat of his
pants.
c. He was coming up the steps. There was a wad of bubblegum on the seat of his
pants.
Likewise, the use of the so-called narrative present in (53b), or the ‘sportcaster’s
present’ in (54b), presents a construal of the time of the linguistic event, which
has the effect of bringing the reported event conceptually closer to the listener
(this construal also involves a ‘moving’ deictic center, moving with each reported
event):
(53) a. He came up behind me, I stopped suddenly, and he rammed into me.
b. He comes up behind me, I stop suddenly, and he rams into me.
(54) a. He hit the ball and the first baseman missed it.
b. He hits the ball – the first baseman misses it . . .
hearer knows:
Example (55a) construes the hedgehog as unknown to the hearer, while (55b)
construes it as part of their common ground. The construal in (55b) could also be
used in a context when the hearer in fact does not know about the hedgehog, as
a way of surprising the hearer with the discovery by manipulating the epistemic
deictic construal.
But in fact the wholesale structuring of clauses is determined by the epistemic
perspective of common ground. This structuring often goes under the name of in-
formation structure or information packaging (Lambrecht 1994). Examples (6a–b)
in §3.1 illustrate alternative construals of a scene based on a difference in informa-
tion structure. The presentational sentence in (6a) (a subtype of what are also called
thetic or sentence-focus structures; Lambrecht 1994:177) presents all of the infor-
mation as part of the assertion directed to the hearer. The ordinary topic-comment
(categorical or predicate-focus) sentence in (6b) presents the subject referent as part
of a presupposition of current interest (and therefore part of the interlocutors’ com-
mon ground) and only the predicate as part of the assertion (Lambrecht 1994:121).
In fact, information structure is one of the clearest syntactic examples of alterna-
tive construals of what is the same scene from a truth-conditional point of view.
Another example of perspectival construal that has been discussed more by prag-
matically oriented linguists than by cognitive linguists is the notion of empathy
(Kuno and Kaburaki 1977). By empathy, Kuno and Kaburaki mean the partici-
pant in the reported event whose perspective is taken by the speaker. Empathy is
perspectival and as such is subject to alternative construals. Kuno and Kaburaki
argue that empathy is involved in the semantics of a number of grammatical con-
structions. For example, the choice of describing Bill as John’s brother, that is, as
a person anchored by his relation to John, instead of as Bill independent of John,
implies that the speaker empathizes more with John than Bill. Likewise, choos-
ing the passive Bill was hit by John, instead of the active John hit Bill, implies
speaker empathy with the subject referent, namely Bill (compare [5a–b] in §3.1).
Example (56c) is therefore odd compared to (56a–b), because of a clash in con-
strued empathy: the speaker uses a description of Bill that implies empathy with
John, combined with a grammatical voice that implies empathy with Bill (Kuno
1987:203–6):
Some linguists have argued that the distinction between the grammatical re-
lations of subject and object is one of deixis or empathy. DeLancey (1981), in
his analysis of split ergativity and subjecthood, argues that the unmarked subject
category involves construing the orientation of the speaker towards the temporal
and causal beginning of the event. On the other hand, Langacker (1991a:305–
17) defines subjecthood in terms of the most prominent figure, which combines
the construal operations of attention (most prominent) and judgement (figure).
Whichever is the correct analysis – and it is possible that different languages use
different construals for subjecthood – the point is that fundamental grammatical
categories such as subject, which are treated as ‘meaningless’ in some syntactic
theories, represent a construal of the referent in the situation described by the ut-
terance, and the construal analysis can predict patterns of (un)acceptability such
as that found in (56a–c).
3.4.3 Subjectivity
The last construal operation under perspective in Table 3.1 is Langacker’s
notion of subjectivity/objectivity. This refers to how one conceptualizes a scene
that includes the speaker herself/himself. Two simple if restricted examples illus-
trate the alternative construals (Langacker 1987:131):
Example (57a) represents the more common subjective construal of the speaker
using a deictic personal pronoun, defining her identity relative to the speech act
situation. Example (57b) involves objectification: the speaker describes herself in
terms independent of the speech act situation.
It is also possible to subjectify reference to an entity (Langacker 1987:132):
In (58), an entity that is not the speaker, namely the physical image in the pho-
tograph, is described using a deictic expression (me), as is in fact common in
so-called picture noun contexts (see also §2.6). Another, more common example
of subjectification is the construal implied when using certain spatial expressions
that can leave a ground object unexpressed (Langacker 1991b:326, 328):
3.5 Constitution/Gestalt
but there are also many examples of alternative construals of objects (Herskovits
1986:76):
If there is a lot of dust or a few drops of milk, the actual spatial configuration
of figure and ground in (60a) and (60b) is not that much different. But since the
function of bowls is to contain potable liquids, the bowl is construed as a container
with in in (60a), and since dust is thought of as an extraneous substance, the bowl
is construed as a surface with on in (60b).
The geometric construal of an object often requires selective attention (Her-
skovits 1986:65, 67):
In (61a), she is unlikely to be under the ground or inside the trunk (though such
a construal is possible if, say, the suspect is leading the detective to the location
of the body of the murder victim). The usual construal selects only the lower
surface of the foliage as the underside of the object. In (61b), selection ignores
the table legs, and scalar adjustment reduces the tabletop to a two-dimensional
surface. In (61c), selection (driven by encyclopedic knowledge) profiles only the
top surface of the water to specify the figure-ground relationship (and all of these
examples are construed relative to the speaker’s canonical upright orientation, of
course).
Another image schema that imposes a structure, this time more typically as-
sociated with properties, is the scale image schema, which provides a gradable
dimension to a domain, which may or may not be measurable. The ways in which
an entity is construed as possessing a scale or multiple scales are described in
detail in §7.4. Here we simply note that the same domain may be construed with
a scale (in contrast to a polar construal, as in [62a–b] and [63a–b]), or construed
as calibratable, as in (64), a domain not usually considered measurable:
Example (65a) represents the prototypical causative type: an antagonist (the causer)
forces an agonist (the causee – the ball) that tends towards rest to move. Example
(65b) extends the notion of causation to maintaining a rest state: the antagonist
resists the agonist’s tendency to move. Example (65c) further extends to notion
of causation to enablement: the antagonist acts in a way that allows the agonist to
exert its tendency towards motion.
Croft (1991, 1998b, in prep.) argues that the force-dynamic structure of events
largely determines the encoding of subject, object and oblique arguments of pred-
icates. For instance, the choice of for in I baked brownies for Mary vs. with in I
beat the eggs with a fork is determined by the fact that Mary is the endpoint, the
beneficiary of the baking event, while the fork acts upon the eggs and is therefore
an intermediate participant in the force-dynamic chain. The difference in degree
of affectedness of chew the bone and chew on the bone in (11a–b) in §3.1 is a
(conventionalized) consequence of the alternative construals of the degree of af-
fectedness of the bone evoked by the object-oblique contrast.
Different choices of verbs, or different voice forms, or different argument-
linking constructions, express different conceptualizations of the force-dynamic
structure of the event. For example, (66a) construes the situation as force-
dynamically neutral (being a static situation), but (66b) construes the situation
as having a force-dynamic value of resisting the effects of some (unspecified)
force-applying process.
Force and resistance play a role in the construal of semantic domains other than
causation. For example, Talmy (1988b) and Sweetser (1990, chapter 3) argue that
deontic modals such as may and must in (68a–b) construe the deontic modality
as letting causation or the absence of resistance ([68a]; compare [65c]) vs. the
application of force ([68b]; compare [65a]):
May in (69a) indicates the absence of resistance from concluding that the propo-
sition She is ill is true, while must in (69b) forces one to the conclusion that She is
ill is true.
understanding. The schema for these gestalts have parts and dimensions that stand
in various relationships that allow us to make sense of our experience. (Johnson
1987:61)
Johnson notes that many image schemas are experienced together and describes
this as a superposition of schemas, using the example of things we co-experience
as both near us and central to our vantage point vs. things far away and peripheral:
‘The center-periphery schema is almost never experienced in an isolated or
self-contained fashion . . . Given a center and a periphery, we will also experience
the near-far schema as stretching along our perceptual or conceptual perspec-
tive’ (Johnson 1987:125). The superimposition of image schemas is identical to
the combination of domains in a domain matrix. In fact, image-schematic domains
are usually combined in a matrix with ordinary domains. For example, our expe-
rience of degrees of weight combines the SCALE image schema(tic domain) with
another basic domain, WEIGHT. It is very difficult to separate WEIGHT and
SCALE, but in this respect, WEIGHT and SCALE represent the tightest relation-
ship between domains in a domain matrix, that is, what Langacker describes as
dimensions of a domain (see §2.4).
The analysis of (constitutive) image schemas as image-schematic domains is
not incompatible with their function as construal operations, because domains
themselves are construals, framing the experience to be communicated in a certain
way. What makes image schemas worthy of separate treatment here is their per-
vasiveness in experience: to be communicated, our experience must be construed
in terms of basic structure, scales and force dynamics.
3.6 Conclusion
(Talmy 2000:84):
The formula in (71b) describes the most natural construal of (71a): a scene in
which (at least) part of the spatial region occupied by the bird is included – the
meaning of in – in the interior space defined by the outline of the visible part of
the spatial region occupied by the bush.
The second general observation is that the layers of construal operations must
yield a conceptually unified construal of the meaning of the utterance (Croft
1993[2002]:163, 194–99). Croft argues that all of the concepts in a single clause
must be construed as part of a single unified domain.8 Examples (72)–(73), for
example, must be construed wholly in the domain of emotion and semantic content
respectively; this requires a metaphorical construal of the spatial preposition in in
(72) and a metonymic construal of the human proper name Proust in (73) (ibid.,
195):
In (72)–(73), the alternative construal of one word in the clause (in, Proust) is
driven by the normal or ‘literal’ construal of another word or phrase in the clause
(mood, read). However, all that matters is that the entire clause is construed in a
single domain. It is possible that alternative construals are available for the entire
8 In the original 1993 paper, Croft argues that only immediate clause dependents must be conceptu-
ally unified; the internal structure of argument phrases may be unified around a different domain.
Nunberg’s analysis of metonymy suggests that even argument phrases must conform to the unity of
domain (Nunberg 1995; see also §3.2.1).
Conceptualization and construal operations 71
The conceptual unity of domain is only one of three conceptual unities that a
clause must obey. The referents in a clause must be construed as belonging to a
single place and time in a mental space (Fauconnier 1985; Croft 1993[2002]:200).
Thus in (76), the hearer must construe the referent of her sister to exist in Margaret’s
belief space (whether or not Margaret has a sister in reality, or whether or not the
description her sister applies to the referent in Margaret’s belief space), and must
construe the referent of a car to exist in the same space (again, whether or not a
specific car fitting the description exists in reality):
The referents in a clause must also conform to a single instantiation of the event
plus participants (Croft 1993[2002]:201). This unity of selection accounts for the
interaction between verbal aspect, noun countability and adverbial construals, as
in (77)–(78):
In (77a), the situation is construed as generic (hence unbounded and not referring
to a specific event); so wine must be construed as referring to the type. In (77b),
the situation is construed as specific and bounded, so wine must be construed
as referring to a specific bounded amount of the liquid. In (78a), the situation is
construed as bounded by the definite noun phrase and the compatible adverbial
phrase in two hours; so write is construed as bounded (telic). In (78b), the situation
is construed as unbounded and specific by the adverbial phrase for two hours, so
the unbounded bare plural letters is construed as specific (not generic as in I hate
to write letters), and write is construed as specific and unbounded.
The more philosophical question that construal raises is, what is the relation-
ship between language, thought and experience? Are there any constraints on the
relationship, and in which direction do the constraints operate: from language to
thought to experience, or the reverse direction, or both?
In a number of places in this chapter, we have referred to the typical construal
of a particular experience: for example, pregnancy is typically construed as not
gradable, a smaller, movable object is typically construed as figure, an action is
72 A conceptual approach to linguistic analysis
typically scanned sequentially, and so on. The nature of our experience in many
instances favors certain construals over others. It is these widespread typical or
default construals that have led semanticists to posit a more rigid model of the
mapping from linguistic meaning to experience. Cognitive linguists emphasize that
flexibility is necessary for understanding conceptualization, and this flexibility is
due to the nature of the human mind as it engages with the world. But experience
does constrain human conceptualization to some degree, rendering some construals
difficult and others almost impossible.
It appears that the pervasive role of conceptualization in linguistic expression
entails a relativistic approach to the relation between language and thought: the
way we conceive our experience is determined by the grammatical structure of our
language. This strongly relativistic formulation is not generally found in cognitive
linguistics, however.
Langacker argues that language-specific semantic structure must be distin-
guished from a universal conceptual structure (Langacker 1976). Langacker rejects
the claim that ‘semantic structure can, in some unclear but hopefully straightfor-
ward way, be related directly to thought and cognition, i.e. the structures manip-
ulated in cognition are essentially the same as the semantic structures underlying
sentences’ (ibid.). In discussing an example from Whorf, who compares the English
He invites people for a feast to its nearest Nootka equivalent, which literally trans-
lates into something like ‘He goes for eaters of cooked (food)’ (ibid., 342–44),
Langacker suggests that both could be expressing the same cognitive experience,
but employ different semantic structures to express the experience. As Lakoff puts
it, ‘experience does not determine conceptual systems, [it] only motivates them’
(Lakoff 1987:310).
Langacker calls language-specific semantic structures conventional imagery.
The semantic representations of Cognitive Grammar, many examples of which
have been used in this chapter, are intended to describe this conventional imagery,
not the presumably universal cognitive structures that these conventional images
construe. Langacker uses the example of the expression of bodily states (Langacker
1976:345) to illustrate his approach: English speakers say ‘I am cold’, whereas
French speakers say literally ‘I have cold’ and Modern Hebrew speakers say ‘It
is cold to me.’ In Langacker’s view, ‘these expressions differ semantically even
though they refer to the same experience, for they employ different images to
structure the same basic conceptual content’ (Langacker 1987:47).
The question remains as to the status of these conventional images in thought.
Langacker suggests that it is relatively ephemeral:
But bodily states can be construed as gradable, which is incompatible with pos-
session (Croft 2001:115):
(80) a. J’ai très froid. ‘I am very cold’ [lit. ‘I have very cold’]
b. *J’ai très une voiture. [lit. ‘I have very a car’]
4.1 Introduction
The act of categorization is one of the most basic human cognitive ac-
tivities. Categorization involves the apprehension of some individual entity, some
particular of experience, as an instance of something conceived more abstractly
that also encompasses other actual and potential instantiations. For instance, a
specific animal can be construed as an instantiation of the species DOG, a specific
patch of color as a manifestation of the property RED, and so on. We shall call
this abstract mental construct a conceptual category. Conceptual categories can
be regarded as cognitive tools, and are usually credited with a number of general
functions:
(a) Learning. Experiences never recur exactly: our ability to learn from past
experience would be severely impaired if we could not relate the present to sim-
ilar aspects of past experience, that is, by putting them into the same conceptual
categories.
(b) Planning. The formulation of goals and plans to achieve them also requires
knowledge to be disassociated from individuals and packaged into concepts char-
acterizing categories of entities.
(c) Communication. Language works in terms of generalities, that is, in terms
of categories. Any linguistic expression, however detailed, in the end represents
only a category of referents.
(d) Economy. Knowledge does not (all) need to be related to individual mem-
bers: a significant amount can be stored in relation to groups of individuals. New
knowledge gained on the basis of interaction with one or more individuals can be
easily generalized to other members of category. Conversely, knowing, on the basis
of a limited number of criteria, that an individual belongs to a particular category,
can give access to a much wider range of information about that individual.
There is an important distinction to be made between generic concepts like
CAT and TERRORIST, and individual concepts like TONY BLAIR and CLEOPA-
TRA. The process of categorization presupposes a more basic one, namely, that of
74
Categories, concepts and meanings 75
what is here called ‘graded centrality’ constitutes a problem; that is, the fact that
some members of a category are judged ‘better’, or ‘more representative’ of the
category than others: in a classical category, all members are equal. Thirdly, the
classical model can offer no account of why category boundaries, in practice, seem
to be vague and variable (they are frequently described as ‘fuzzy’, but our account
will be somewhat different). A model of category structure is supposed to provide
a basis for an account of how we use categories in remembering, planning, reason-
ing and so on. A classical definition is not a very efficient vehicle for this purpose,
because the information it contains is too sparse.
Several theories of the nature of natural categories have been proposed, mostly
in the psychological literature, but the theory that has had the most influence on
the development of cognitive linguistics is undoubtedly prototype theory, to which
we now turn.
instance, if the category was VEGETABLE, the ratings of various items (by British
subjects) might be as follows (these scores represent the ratings of one of the
authors):
GOE rating
LEEK, CARROT 1
BROCCOLI, PARSNIP 2
CELERY, BEETROOT 3
AUBERGINE, COURGETTE 4
PARSLEY, BASIL 5
RHUBARB 6
LEMON 7
whether or not the string forms a word. Presenting a semantically related word, or
the same word, before a test item has the effect of speeding up subjects’ responses:
this phenomenon is known as priming. The relevant case here is when the prime
is a category name, like FRUIT. The degree of speeding up is the priming effect.
The priming effect correlates with the GOE score of the category member, that is,
for Britons, FRUIT will speed up the response to APPLE to a greater degree than
the response to, for instance, DATE.
Psycholinguistic variables such as verification speed and priming are regarded
as particularly significant correlates of GOE because they are not under conscious
control and therefore can be claimed to reveal underlying properties of categories.
There has been some dispute in the literature regarding the relationship between
the GOE of an item and its degree of membership (henceforward DOM) in
the category. Some say that, in giving GOE scores, subjects are in fact giving
DOM scores. However, this is misleading. What they were asked to do was to
rate items as to how good they were as members of particular categories. Saying
that they were giving DOM ratings is a subsequent interpretation. Those who
object to the equation of GOE and DOM (for instance, Lakoff [1987:45], Pulman
[1983], Cruse [1992b]), point to examples like OSTRICH in the category BIRD.
There is no doubt, they say, that an ostrich is a fully paid-up member of the
BIRD category, but also undeniably has a low GOE, hence the two parameters
must be independent. Ungerer and Schmid (1996) claim not to see a problem, but
they do not throw any light on the matter. Taylor (1989[1997]) claims that both
assessments of OSTRICH are DOM judgements, but they are made with respect
to differently construed categories. An ostrich is judged a full member relative to
an ‘expert’ category, which has clear membership criteria; the graded membership
judgement is made relative to the everyday category BIRD, which does not have
clear membership criteria. This is ingenious, and we are sympathetic to the appeal
to different construals of categories denoted by the same lexical item, but Taylor’s
account does not stand up to close scrutiny.
The first point to make is that yes/no judgements and graded judgements co-
exist as alternative construals in many semantic domains. Take, for example, the
case of dead and alive. The domain of what might be called ‘vital status’ is often
construed dichotomously: saying John is dead normally commits one to the truth
of John is not alive. But it also possible to say John is more alive than Mary. This
does not change the domain, but reconstrues it as a gradable scale. The same is
true of category membership. In the case of BIRD (whether construed as an expert
category or an everyday one), anything on the right side of the boundary is in the
category, but at the same time, variable centrality allows a gradable construal of
some things as more in the category than others, hence there is some legitimacy
in interpreting GOE as DOM.
80 A conceptual approach to linguistic analysis
At the same time, there is something counterintuitive about saying that an ostrich
is, say, only 30% a bird, and perhaps the term DOM should be reserved for a
distinctive property. We may think of a category as a container (i.e., a result of
the imposition of the Lakoffian CONTAINER image-schema on a domain). What
would we normally understand by a description of an object as 30% in a container?
Surely something like a teddy bear that is partly in and partly out of the toy box,
rather than one that is nearer the side of the box than the middle? There is a
category equivalent of this picture. When we say that, for instance, a priest is to
some extent a social worker, we are effectively placing him part-in, part-out of
the category (cf. Lakoff 1973). That is to say, we are construing the categories
PRIEST and SOCIAL WORKER as partially overlapping. This is surely a more
useful conception of DOM. (NB: the view expressed here regarding DOM is
significantly different from that in Cruse 1992.)
Two problems may be signaled here in connection with GOE experiments and
results. One concerns the meaning attributed, in the context of the experiments, to
expressions such as How good is X as an example of category Y? How, exactly, is
How good? interpreted? Used in actual contexts, good and better do not normally
give rise to communicative problems. For instance, the goodness of a dog, if
thought of as a pet for a young child, is different from what it would be if it was
thought of as providing security for a house, or as contributing to the life of a
farm. This does not destroy the notion of GOE, but suggests that truly significant
results would require specific construals of both the categories being judged and
the meaning of good. Various at least partially distinct notions of goodness can
be teased out (see, for instance, Lakoff 1987:84–90). The following are the main
types:
(i) Typicality/representativeness. This indicates how accurate/useful an idea of a
category one would get from familiarity with only one subcategory. This dimension
has a clear relation to frequency. Lakoff points out that we are much more likely
to generalize properties from typical to non-typical members than vice versa. In
certain cases, a known individual member may be assumed to be typical of a class
(for instance, if a person has limited experience of the class).
(ii) Closeness to an ideal. This is related to what in Cruse 1990 was called
‘quality.’ The example given there was that of emeralds. The most highly valued
emeralds have a deep, intense color, and are without flaws; but these are also the
most rare (and the bigger, the rarer), so they are in no way typical. As Lakoff
points out, ideals operate in many domains. They may be represented by a set of
abstract properties, as in the case of emeralds, or they may be centered around an
individual (called by Lakoff a ‘paragon’).
(iii) Stereotypicality. This is interestingly different from typicality, but a fully
convincing explanation of the difference is not yet available. Lakoff’s account
Categories, concepts and meanings 81
(1987:85–6) is suggestive, but not fully explanatory. Lakoff says that the use of
typicality is usually unconscious and automatic, whereas the use of stereotypes is
conscious: this is plausible. He also says that stereotypes change, but typicality is
constant over a person’s lifetime. However, typicality changes as reality changes
(think of a typical car or computer or camera), whereas a stereotype can persist
in the face of change. Stereotypes are also typically associated with evaluative
features.
There is another problem. A lot of the classical experiments deal with sub-
categories as category members, rather than individuals: for example, subjects
are given a category such as FRUIT, and a range of fruit types such as APPLE,
STRAWBERRY, MANGO, PASSION FRUIT, DATE, OLIVE and so on for GOE
scoring. Other experiments involve individuals: for instance, the work on proto-
typical colors by Rosch (Heider 1971, 1972, and Berlin and Kay 1969) and others;
also experiments with young children typically use individual items, not category
names. It does make a difference. Using categories as examples of other categories
suppresses properties that can enter into the notion of goodness. This is true of
quality as described above. Another example is the property of well-formedness:
it is all very well saying that an apple is the best example of a fruit, but what if it is
a rotten apple? As far as individuals are concerned, well-formedness is yet another
variety of goodness.
Normally, one level of specificity in each set, called the basic (Rosch et al.
1976) or generic (Berlin et al. 1973) level of specificity, has a special status,
and importance. (The basic level items in [1] are printed in bold italic.) Apart
from the basic level, two further levels of specificity with different characteristics
are usually identified: superordinate level and subordinate level. These are not
defined simply by their position in the chain – there are substantive characteristics
that distinguish one level from another. (For an extended discussion of hierarchical
structure in concepts, see Murphy and Lassaline 1997.)
Categories, concepts and meanings 83
(i) It is the most inclusive level at which there are characteristic patterns of behavioral
interaction.
To appreciate this point, imagine one is asked to mime how one behaves with,
say, a dog: this is not too difficult, most people would mime, for instance, patting
and stroking the dog. But suppose one were asked to mime how one behaves with
an animal: this is very difficult unless one knows what kind of animal it is. The
same is true of furniture relative to chair, and spoon relative to cutlery.
(ii) The most inclusive level for which a clear visual image can be formed.
This includes relations between parts. For most superordinate artifactual cate-
gories, such as TOOL, CUTLERY, CLOTHES or FURNITURE, there is no com-
mon part-whole structure for members. Biological superordinate categories show
more regularity in part-whole structure, but there is much less commonality in the
relations between the parts.
Basic level terms (i.e., terms whose default construals are basic level categories)
are often felt by speakers to be the ‘real’ name of the referent. Cross-linguistic
studies have shown that they tend to be shorter than terms at other levels, nor-
mally monomorphemic, and are original in the sense of not being borrowed by
metaphorical extension from other domains (Berlin et al. 1973). They are also
84 A conceptual approach to linguistic analysis
more frequently used by parents in talk to children, and thus, not surprisingly, they
are the first words children learn.
(v) Individual items are more rapidly categorized as members of basic level categories
than as members of superordinate or subordinate categories.
(i) Superordinate categories are less good categories than basic level categories,
because although members are relatively distinct from members of neighboring
categories, within-category resemblance is relatively low.
(ii) Superordinate categories have fewer defining attributes than basic level categories.
In experiments by Rosch et al. (1976), where subjects were asked to list the
attributes of basic level items in a superordinate category, few attributes were
generated that could serve as defining attributes of the superordinate category.
However, as suggested in Cruse 1992b, this is perhaps not the best way to elicit
attributes of superordinate categories. The most salient attributes of a category are
Categories, concepts and meanings 85
those that differentiate it from other members of its default contrast set. The salient
features of a horse are those that distinguish a horse from other animals. Subjects
asked to list the attributes of a horse are unlikely to produce ‘has bones,’ ‘breathes’
and so on, because these are shared by other members of the contrast set; they are
more likely to mention ‘can be ridden,’ ‘has a mane,’ ‘has a long tail,’ ‘neighs,’ and
so on. The only way to get significant attributes of ANIMAL is to set up contrasts
with categories such as FISH, PLANT, INSECT. The same is true of FURNITURE:
it is no use looking at informants’ responses to chair, table and the like. More
revealing would be to ask what features distinguish items of furniture from, say,
curtains, carpets, appliances, fireplaces and windows. Thinking of furniture in
this way suggests that items of furniture are prototypically hard (unlike carpets),
mobile (unlike fireplaces) and are places where things happen (unlike appliances,
which are for doing things with). (For a similar, but independent, analysis of
the category FURNITURE, see Bolinger 1992). However, it remains true that
characteristic features of superordinate categories are fewer, and, as a consequence,
family resemblance relations are less marked.
Examples of this are crockery (cups and plates), cutlery (spoons and forks),
furniture (tables and chairs), footwear (boots and shoes), (computer) hardware
(hard disks and modems). Less frequently mentioned are cases where the con-
verse is true: the superordinate is a count noun and the basic level term a mass
noun: metals (iron and copper), beverages (beer and wine), spices (pepper and
coriander). There is never a discrepancy in this respect between basic-level and
subordinate-level terms. Superordinate terms are also frequently morphologically
complex and/or polysyllabic.
(i) They are less good categories than basic level, because although members have
high mutual resemblance, they have low distinctiveness from members of neigh-
boring categories.
(ii) They are much less informative relative to their immediate hyperonymic category,
hence, when subjects are asked to list distinctive attributes, the lists differ very
little from the lists given for the hyperonymic basic level items.
86 A conceptual approach to linguistic analysis
(iii) They are frequently polymorphemic, the most common pattern being modifier-
head (e.g. teaspoon, rocking-chair).
This is taken by, for example, Ungerer and Schmid (1996) to indicate that they
are distinguished from basic level by a single property, rather than encyclope-
dically (e.g. teaspoon, rocking-chair). However, a distinction needs to be made
between naming practices and conceptual content: the ‘single property’ is a matter
of naming, while there are virtually always unencoded encyclopedic distinctive
characteristics. For instance, although spaniel is a single-morpheme word, and
long-tailed tit is a complex expression incorporating a single property (the posses-
sion of a long tail), the extra specificity in each case over the basic level category
is encyclopedic.
The above account is close to the account given by cognitive psychologists
such as Murphy and Lassaline (1997). Anthropological linguists have also made
extensive studies of the hierarchical organization of categories (see, for instance,
Brown 2002). Their approach differs in many ways from that of the psycholo-
gists. Firstly, they have a strong cross-linguistic orientation. Secondly, the most
extensive studies have been of biological kinds (a distinction is usually made be-
tween ‘folk-classifications’ and ‘expert systems’: most studies are of the former):
some (e.g. Atran 1990) claim that only biological kinds are truly hierarchized,
and among biological kind concepts only ‘general purpose’ categories such as
animal, dog, spaniel, beech, copper beech, bush and so on are hierarchized, but
not utilitarian categories such as vegetable, weed or pet. Thirdly, they recognize
a greater ‘depth’ of hierarchization, and use different terms for the levels. The
following is an example (the equivalent psychological categories are given in
brackets):
The properties attributed to the generic level do not significantly add to what was
said above concerning the basic level. The remarkable constancy of hierarchical
structuring across a wide variety and degree of complexity of cultures suggests
that it is a cognitive universal and probably innate. There is a dispute among
anthropologists as to the underlying motive force for the evolution of classificatory
systems. One school holds that it is driven mainly by utilitarian considerations:
the categories evolved because they were an aid to survival. The other school
holds that the evolution of the systems for classifying biological kinds is driven
Categories, concepts and meanings 87
by intellectual curiosity. There are a number of reasons for the latter claim. The
systems are remarkably similar across the world, even though the cultures and
their environments differ markedly; many cases are found of distinctions that have
no utilitarian value in the culture that uses them; they tend to coincide to a high
degree with scientific classifications. Brown (2002), while acknowledging the force
of these arguments, points out nonetheless that hunter-gatherer societies typically
have far fewer categories than settled agrarian societies, and suggests that there is
a functional reason for this. Small agrarian communities are typically larger than
hunter-gatherer communities, and when harvests fail, it is very important for them
to have access to alternative food sources, hence a detailed knowledge has survival
value. Hunter-gatherer societies, on the other hand, are typically much smaller,
and their essentially mobile lifestyle makes them much less dependent on food
available in a particular locality; hence detailed knowledge of local flora and fauna
is of less value.
hopping,’ ‘larger than average human,’ ‘has scales’ would all have negative weight-
ing for dog). It does not appear that this strategy has ever been followed.
1 It should be emphasized that the existence of a boundary does not entail the existence of a ‘core
definition’ with necessary and sufficient criteria, as proposed in what Hampton (1997:93) calls the
‘binary view’ put forward by Smith et al. (1974).
90 A conceptual approach to linguistic analysis
airplane
glider
hang
glider
Here, the prototype of GLIDER falls within the category boundary of AIR-
PLANE, but HANG GLIDER falls outside it.
There is a psycholinguistic correlate of the position of a boundary, involving
speed of response to a categorization task. If subjects are asked whether or not an
item belongs to a given category, the speed of response depends on how near the
item is to the category boundary: the closer it is, the slower will be the response.
Hence, in (5), the bold items will be responded to slowest:
Another indication of marginal status is the following. Take the case of SHOE as
a member of the category CLOTHES. One would not hesitate to say of a suitcase
that contained underwear, shirts, socks, jackets and trousers, but nothing else, that
it ‘contained only clothes.’ However, one would hesitate to say of a suitcase full
of shoes, that it contained only clothes; on the other hand, there would be no
such hesitation if the contents were all shirts. This is arguably a function of the
marginal status of shoes in the category of clothes compared with shirts. It appears
that we construe the category CLOTHES differently, that is, more generously,
when confronted with a variety of types of clothes, including marginal ones; when
confronted with a collection of shoes, we construe the categories CLOTHES and
SHOES as mutually excluding.
The location of the boundary of a category is independent of its prototype, that
is to say, two categories may have the same prototype but different boundaries;
likewise, two categories may have the same boundaries but different prototypes.
Take the French word corde and its default English translation rope. A questioning
Categories, concepts and meanings 91
of native speakers of the two languages suggests that the prototypes of the two
categories are very close: both put forward the same sorts of thing as best examples.
However, their boundaries differ. Le Petit Larousse defines ficelle (‘string’) as ‘une
corde mince’; a parallel definition of string as ‘a thin rope,’ would seem very odd.
That is to say, ficelle falls within the (default) boundary of the category CORDE,
but string falls outside the boundary of the category ROPE. The converse case,
of identical boundary but different core, is perhaps exemplified by courage and
bravery in English. It would be hard to think of an act that was a manifestation
of courage but not of bravery, or vice versa. But their core regions are arguably
distinguishable. Student informants were asked to give a relative rating of (6a) and
(6b) as (i) an example of a brave act and (ii) an example of a courageous act.
(6) a. A person jumps into a fast-flowing river in an attempt to save someone who
has fallen in.
b. A person risks his/her career and livelihood by exposing malpractice and in-
justice at the heart of government.
There was substantial agreement that (6a) was the better example of bravery
and (6b) the better example of courage.
A fundamental problem with boundaries is that they do not arise naturally from a
prototype representation. Even in Hampton’s version of the model, the boundaries
are simply stipulated in an arbitrary fashion. Prototype theorists typically say that
natural conceptual categories have fuzzy boundaries. Indeed, this is one of the
main arguments against the classical model. Claimed pointers to fuzziness are, for
instance, the fact that different subjects make different judgements as to the location
of boundaries, and the same subject will make different judgements under different
contextual conditions. Even the psycholinguistic experiment quoted above yields
a borderline region rather than a sharp line. However, it should be pointed out that
even a fuzzy boundary has a location. The notion of a fuzzy boundary will be
critically examined below.
Three ways in which this occurs can be identified. First, there is the question of
the convergence between the individual and the profiled region of the frame. Take
the case of car and tractor within the category of VEHICLE. Most informants
award car a somewhat higher GOE score than tractor. The reason appears to be
that the ideal vehicle is designed for travel along roads, rather than across fields,
hence there is a better fit between CAR and VEHICLE than between TRACTOR
and VEHICLE. The second type of graded centrality involves items that do have
a traditional definition, like bachelor. In this case, graded centrality can arise
from similarity between the ideal background domain and the actual background
of the individual. As we have seen, the definition operates against a set of cul-
tural background assumptions concerning marriageability: the reason we regard a
(Roman Catholic) priest as not a very good example of a bachelor, even though he
satisfies the basic definition, is that our background assumptions about priests do
not fit our assumptions about an ‘ideal’ bachelor. A third case is when a concept
is characterized by a cluster of ICMs, as in the case of MOTHER (§2.5). Here,
the ICMs behave like features, in that the more of the members in the ideal cluster
are present in a particular instance, the more central the instance is within the
category.
knowledge, and that the alleged properties of fixed concepts can be given other
explanations.
According to Smith and Samuelson, the elements out of which a concept is
created are past history, recent history, current input. On the topic of past history,
that is to say, accumulated memories of previous experiences, they point out that
each experience has a permanent effect on our ‘ways of knowing,’ and further have
this to say:
What is recorded on each past experience will include such things as salient
contextual factors, perceived and inferred relations (causal and other) with other
things, accompanying language and so on. The second element is immediately
preceding mental activity. They adduce the ubiquitous effect of priming as an
example of this. More particularly for concept formation they claim:
[T]here is a pull for coherence from one thought to the next one, for the meaning
of an event to depend on its place in a stream of events. If we think first about
eating and then about frogs we will think differently than if we think first about
ponds and then about frogs. (ibid.)
When this question is put to a typical class of undergraduates, a typical result is that
a minority, but a significant minority, answer Yes, while the majority say No. This
is, of course, a typical ‘fuzzy’ result. Now suppose the question in (8) is asked:
(8) Is a cyberpet a real pet?
The response this time is overwhelmingly No, because the word real encourages
a particular construal of the position of the category boundary. On the other hand,
suppose a scene is set such as the following: an educational psychologist, say, is
advising the parents of a child with behavioral problems, and says (9):
(9) I advise you to get her some kind of pet – even an electronic one might be
beneficial.
In this case, no one in a typical class finds anything anomalous in the psychologist’s
utterance, even though pet is used to include the electronic variety. The expressions
some kind of and even in the context encourage us to construe a broader category
of pets.
Another example is dog in (10)–(12):
(10) A dog has four legs.
At first sight this seems an obvious truth. But what about dogs that have lost one
or more legs in an accident? It seems that when we interpret (10), we construe the
category of dogs to include only well-formed dogs. Yet another construal of the
boundaries of the category of dogs is illustrated in (11):
(11) Dogs are mammals.
For (13), we must exclude flightless birds and injured birds incapable of flight
from our construal of bird. In (14) we are constrained to interpret bird (if uttered
by an inhabitant of a typical Manchester suburb) as ‘most familiar type of small
garden bird,’ on the assumption that no one would expect to see ostriches or eagles
in their garden.
Categories, concepts and meanings 95
4.4.2 Frames
Frames/ICMs (in some cases cluster ICMs) are presented by Fillmore
and Lakoff as more-or-less invariant structures having a stable association with
lexical items, which allow for variable boundary construal, presumably in terms
of the goodness-of-fit required between perceived reality and aspects of the frame.
However, although the frame may be relatively more stable than the boundaries, the
96 A conceptual approach to linguistic analysis
dynamic construal approach allows also for variable construal of the frame itself.
The experiments of Barsalou, reported above, where ad hoc categories are formed
that have all the characteristics of established categories, suggest that frames may
be construed on-line. Also, the type of variation shown by the category DOG in
dogs and other pets and Dogs are mammals seems more convincingly explained
by a modulation of the frame, rather than an adjustment in the degree of fit with a
constant underlying the DOG frame. In any case, whether or not frames are subject
to construal, the mechanism of boundary placement is still in need of elucidation.
They may be able to form an image of a generic (garden) bird, but not have enough
experience or knowledge to be able to visualize individual species; their patterns
of behavior will be very much the same towards all garden birds. Insofar as they
have both a superordinate level and a basic level construal of bird, these will not
be different level construals of the same category, but two different categories that
fit at two levels.
What happens in the case of a speaker who can operate with two different
systems? Consider the case of a dog-breeder who in his work environment exhibits
basic level characteristic behavior in respect of categories such as SPANIEL,
COLLIE, ALSATIAN, TERRIER. What happens when he converses with a non-
specialist? Presumably many, at least, will adjust themselves cognitively to the
new situation, and revert to the societal default construal of the terms. But do
they actually change their categories? Or do they simply construe a new level for
them? If they actually restructure the categories, for instance by backgrounding
aspects of knowledge that are highly relevant in a professional setting, then they
are effectively creating new conceptual categories. It would seem reasonable to
assume that level cannot be construed independently of content, that is to say, any
observed movement up or down a taxonomic hierarchy will be a consequence of
different construals of the category denoted by a lexical item.
a vital component of the raw material required for the construal of meaning. But,
as we shall see, they represent only one component among several. On this view,
words do not really have meanings, nor do sentences have meanings: meanings are
something that we construe, using the properties of linguistic elements as partial
clues, alongside non-linguistic knowledge, information available from context,
knowledge and conjectures regarding the state of mind of hearers and so on.
Our account of word meaning will incorporate, albeit in an adapted form, the
basic insights of the dynamic construal picture of conceptual categories. It should
be borne in mind, however, that concepts are not necessarily equatable with con-
textually construed meanings, or, as we shall call them, interpretations. Consider
the following sentences:
(15) Dogs are not allowed in this building.
(16) I like cats, but I can’t stand dogs.
In any situated use of these sentences, the interpretation of dog will be the
same as the concept DOG construed in the same context. However, cases like this,
although widespread, are probably in the minority. Take sentence (17), said in
reference to the family dog, whom someone forgot to feed at mid-day:
(17) Oh, look: that poor animal hasn’t had anything to eat since this morning!
Here we can say that the word animal causes the construal of an appropriate
conceptual category ANIMAL. However, the fully construed meaning involves
an individual concept, namely the family dog, which is itself further construed in
response to contextual factors (which include, among many others, the fact that
the word animal was used in the referring expression, rather than the word dog).
Even without statistical evidence, it seems a safe guess that the bulk of everyday
communication ultimately concerns individual things or people, rather than classes
of individuals.
There are four basic notions in the present account of meaning, namely, contex-
tualized interpretation, purport, constraints and construal and they will be discussed
in that order.
(18) Bertie was lying on the doorstep in the sunshine as Ingram drew the jeep to a halt
beside his gate . . . The dog raised his shaggy head and thumped his tail on the
mat before rising leisurely to his feet and yawning.
(19) Bibi . . . sat cross-legged on the floor at Tony’s feet . . . nervously [she] raised her
head.
It is not necessary to dwell on these passages in detail. They are cited simply to
draw attention to the ‘deadness’ of the individual signs, in contrast to the vividness
of the interpretations we construct. Think of what a dictionary entry will tell us
about the meanings of, for instance, thump, raise, rise, and compare this with the
detailed picture the words evoke in the passages quoted. And contrast raise in the
dog raised his shaggy head and nervously she raised her head. These are different
actions, not simply because one is performed by a dog and the other by a girl:
the dog’s head actually moves to a higher position relative to the ground; the girl
merely tilts her head so as to look upwards.
Of course, when we construe these scenes, we draw on our stored knowledge
of the behavior of girls and dogs in different circumstances: if Bibi had been lying
face-down on the floor with her chin on a book, say, and someone had asked to see
the book, then the head-raising referred to in Bibi raised her head would have been
much closer to that referred to in The dog raised his head. Clearly these subtleties
are not inherent in the word raise, but they are part of construed meaning in context,
and are construed as a direct result of the occurrence of the word raise (cf. the
discussion of Fillmore’s semantics of understanding in §2.1).
When we encounter a piece of language in the course of normal communication,
there is an instant of comprehension, a kind of crystallization of the perception of
meaning – we know what somebody has said (or written etc.). This is similar to
our recognition of a familiar face, or when we realize that what we are seeing is a
dog and so on. In the case of the face, we do not merely recognize whose face it
is, but at the same instant we see perhaps that the person is tired, or worried, and
the hair is windblown and so on. On further reflection, we might infer what the
person has been doing, or what the cause of worry is. The processing can continue
indefinitely, but there is nonetheless a prior moment of recognition.
Something similar happens when we encounter a piece of language. We recog-
nize in an instant what has been said, but we can go on working out consequences
100 A conceptual approach to linguistic analysis
and further inferences indefinitely. It is what constitutes the focus of our attention
at the moment of understanding that is referred to here as the interpretation of
an expression. Phenomenologically, it is a fairly clear-cut event. It will be use-
ful to distinguish pre-crystallization processes, processes preceding and leading
up to crystallization, and post-crystallization processes. In many approaches to
meaning, there is a determinate starting point for the process of constructing an
interpretation, but an indeterminate end point. For example, in Relevance Theory
(Sperber and Wilson 1986), the starting point is an explicature, and the end point
(insofar as there is one) is an indeterminate series of implicatures of diminishing
strengths. The present model of comprehension has an indeterminate starting point
(a purport) and a determinate end point.
An interpretation resembles a picture in that it is not susceptible of finite charac-
terization in terms of semantic features, or whatever. Any features are themselves
construals. Of course, a meaning must in some sense have a finite neural represen-
tation, but the elements out of which the representation is composed are more like
the pixels underlying a picture on a computer screen: the resulting experienced
picture is a Gestalt and so is an interpretation. The nature of this experience is still
mysterious.
Notice that the above characterization focuses on the hearer. Presumably there
is something in the mind of the speaker which precedes the utterance, but there is
a sense in which a speaker does not know what they have said until they process
their own utterance. A major task for the speaker is to devise an utterance that will
lead to the desired interpretation forming in the hearer’s mind.
4.5.2 Purport
Each lexical item (word form) is associated with a body of conceptual
content that is here given the name purport. Purport is part of the raw material
contributed by the word to processes of construal of an interpretation (the other
part being a set of conventional constraints). A purport does not correspond to
any specific interpretation, even an abstract one, nor does a word, in general, have
a stable association with specific conceptual categories. At the same time, there
is an intuitive sense of coherence among most of the uses of a word. This can
undoubtedly be partly explained by the constant association between word form
and purport.
Purport may consist of a relatively coherent body of content, or it may display
relatively disjunct parts (as in traditional ‘homonymy’); or, indeed any intermediate
degree of coherence or lack of it. (We do not say of a word such as bank, which
has a disjunct purport, that it has two [or more] purports. There is not a great
deal of harm in saying this for extreme cases, except that it obscures the fact
Categories, concepts and meanings 101
4.5.3 Constraints
Of course, the construal of interpretations is not unconstrained: the con-
straints are many and varied. They vary in strength and may reinforce one another
or cancel one another out (for examples of conflicting constraints, see the dis-
cussion of facets and microsenses in chapter 5). They can also be overcome by
cognitive effort, but the stronger the constraint, the greater the cognitive effort
required to impose a construal that defies the constraint. They also vary in their
stability under change of context. I have classified constraints informally under a
number of headings.
is difficult not to construe a sense boundary between the two meanings of bank,
and this has something to do with the raw/brute reality concerning river banks and
money banks.
4.5.3.3 Convention
Another very important constraint is convention: how the society in which
we live habitually construes situations and uses words, and so on (see Lewis 1969,
Clark 1996 and Croft 2000 for the theory of convention and its limits). There are
two aspects of convention. One is the mapping between word forms and regions
of conceptual content. This in itself constitutes a constraint: the principal source
of difference in semantic potential between one word and another arises from
difference in associated purport. The other aspect of convention is a limitation of
the possibilities of construal of a particular purport. Certain construals, because
of the strength of the constraints, will acquire a special default status, and extra
cognitive effort will be required to impose a different construal. Conventional
constraints are frequently context sensitive, that is to say, given a particular purport
in a particular context (or, more likely, a context-type) conventional constraints
may favor certain construals over others.
4.5.3.4 Context
Last, but not least, there are contextual constraints. These, in general,
correspond to what Clark (1996) calls common ground.
(iii) Type of discourse: under this heading are included such matters as genre
(whether we are dealing with a poem, novel, textbook, newspaper report, personal
letter, friendly conversation, police interrogation etc.), register (whether formal
or informal, if formal, whether technical or non-technical, if informal, whether
jocular etc.) and field of discourse (legal, ecclesiastical, sporting, political etc.).
Categories, concepts and meanings 103
4.5.4 Construal
Construal, in the sense introduced by Langacker, is the central notion in
this account of lexical semantics. It is by means of a series of processes of construal
that an essentially non-semantic purport is transformed into fully contextualized
meanings. For a detailed discussion of construal processes see chapter 3.
represented in the minds of individual speakers, but their origin lies outside the
individual, in the speech community. Different conventional constraints vary in
strength. A weak constraint will yield no more than a favorite, or most likely con-
strual, which can be easily overridden by contextual constraints; a strong constraint
will require heavy countervailing pressure to be overcome. Different aspects of
construal may be independently subject to conventional constraints: the result of
a constraint will rarely be a fully construed meaning, and is much more likely to
be a pre-meaning involving, say, only a boundary placement and leaving scope
for further enrichment by construal. Conventional constraints are not necessarily
independent of context: some may operate only in certain context types. They can
also operate at different levels of specificity, with narrow scope constraints refining
the work of wider scope constraints. Some constraints will be defeasible, but will
govern some aspect of construal if there is insufficient or no indication from the
context as to required construal. The outcomes of these constraints will function as
default construals. Default construals can also be context dependent. It is probably
default construals that give the illusion of fixity of meaning.
One of the main aims of the dynamic construal approach to word meaning
is to achieve a unified account of both hard and soft aspects of word meaning, both
flexibility and rigidity, and to locate the origins of these at first sight contradic-
tory properties. The ‘hard’ properties include sense relations such as hyponymy,
incompatibility, meronymy and antonymy, and the existence of structured lexical
sets (word fields), as well as logical properties such as entailment. For instance,
it is generally accepted that It’s a dog entails It’s an animal. But what does this
mean? And how can it be the case if dog and animal do not have fixed mean-
ings? It will be argued here that such properties properly belong to pre-meanings
resulting from the construal of boundaries, principally, and scales and reference
points. Confident assertions regarding entailment in decontextualized sentences
can be attributed to default construals of boundaries. Hence, flexibility comes
from the nature of purport and the sensitivity of construal processes to contex-
tual factors; conventional constraints ensure that contextual variability remains
within certain limits. Rigidity comes from the operation of image schemas such as
the container schema and the scale schema. Boundary construal can also account
for the appearance of componentiality in word meaning, without the necessity
of assuming that semantic features are permanent elements of the meaning of
a word.
Categories, concepts and meanings 105
This, too, is untenable, however, as it does not allow an adequate role for context
in the construal of the meaning of the complex. The principle could be further
modified to take account of this:
(23) The meaning of a complex expression is the result of a construal process one of
the inputs to which are the construals of its constituent parts.
This allows context to act at two levels: the initial construal of the word meanings,
and at the level of the whole expression. But this is a long way from the original
principle of compositionality. It is more reminiscent of cookery. Is cookery a
compositional art? Certainly, the final result is determined by (a) the ingredients
and (b) the processes applied, so there is an element of compositionality. But it is
not what the proponents of the principle usually have in mind. If we think of global
construals, then they are almost certainly compositional only in the cookery sense.
But there may be aspects of meaning that do obey the classical principle, at least up
to a point. Logical properties are determined by boundary placements, so perhaps
the pre-meanings created by boundary construals behave in the classical way?
This seems plausible: when we construe red hats, we construe a category of red
things and a category of hats, and it seems inescapable that the resulting category
will be the intersection of these two categories. This is only valid, however, for
class membership: we have no guarantee, for instance, that the result will be either
prototypical hats nor that their color will be a prototypical red. In other words, for
certain aspects of meaning, at certain levels of construal, classical compositionality
holds, but not for all aspects or levels.
With this chapter we conclude the exposition of the basic principles and
key concepts underlying the enterprise of cognitive linguistics as we see it. In
106 A conceptual approach to linguistic analysis
the chapters that follow, these principles and concepts will be refined, expanded
and further illustrated through their application, first, to aspects of word meaning
in Part II, and second, to grammar in Part III. We aim to show that the cognitive
approach to language not only opens up new aspects of language, but also addresses
the traditional concerns of grammarians and semanticists in a more satisfying
fashion.
PA RT I I
Cognitive approaches to
lexical semantics
5
Polysemy: the construal of
sense boundaries
5.1 Introduction
The operative factor in this case is of course the immediate linguistic context.
Polysemy is interpreted in this chapter somewhat more broadly than the traditional
lexicographic acceptation involving distinct, established senses, but it includes the
traditional view as a special, perhaps in some ways prototypical, case.
On the present account, bounded sense units are not a property of lexical items
as such; rather, they are construed at the moment of use. (The notion of sense
boundaries that are sharp, but subject to construal, distinguishes the present account
from that of Cruse 2000, and also from accounts such as Deane 1988 and Geeraerts
1993, which question the existence of boundaries.) When we retrieve a word from
the mental lexicon, it does not come with a full set of ready-made sense divisions.
What we get is a purport, together with a set of conventional constraints. However,
in particular cases there may be powerful stable constraints favoring the construal
of certain sense units. If the permanent constraints are pushing very strongly in
one direction, a correspondingly strong countervailing pressure will be necessary
to go against them; if the permanent constraints are weak, whether a boundary is
construed or not will depend on other, mainly contextual, factors. We can portray
the total meaning potential of a word as a region in conceptual space, and each
individual interpretation as a point therein. Understood in this way, the meaning
potential of a word is typically not a uniform continuum: the interpretations tend
109
110 Cognitive approaches to lexical semantics
Full sense boundaries delimit the sort of sense units that include those
that are the stock-in-trade of traditional dictionaries. For us, the class is wider:
traditional dictionaries include only those sense units that are well entrenched in
Polysemy 111
the language and are strongly supported by conventional constraints; however, the
same characteristics can be found in nonce construals.
5.2.2 Entrenchment
The interests of lexicographers are necessarily focused on aspects of word
meaning that are strongly supported by stable, mainly conventional constraints,
and that have attained some sort of default status. They are more likely to recognize
gross distinctions than subtle ones. This is not a criticism. The possible readings
112 Cognitive approaches to lexical semantics
of any word are nondenumerable. A dictionary can only offer a finite list, so a
high degree of selectivity is inevitable. In the absence of clear criteria, it is also
inevitable that different dictionaries will make different distinctions, especially
where more subtle distinctions, or ones less clearly licensed by convention are
concerned. As far as the present discussion is concerned, there is no difference
between entrenched readings and nonce readings in respect of boundary properties
or the nature of delimited units.
There are two main consequences of antagonism. The first is the well-known
identity constraint:
There are strong constraints favoring the construal of two autonomous pre-
meanings, namely, ‘light in color’ and ‘light in weight.’ Because the two readings
of light are antagonistic, we are strongly constrained to construe the same reading
of light in both conjuncts. Two points may be noted here. The first is that the identity
constraint is not absolute: it is possible to construe (6) with different readings for
light in the two conjuncts, but it requires more cognitive work, and the penalty is
a sense of wordplay. The second point is that if we construct the sentence without
verb-phrase anaphora, while there is still arguably some pressure to use a single
reading of light, it is much weaker:
(7) Mary was wearing a light coat; Jane was wearing a light coat, too.
Polysemy 113
The so-called ‘zeugma test’ has been criticized on the grounds that it does not
give consistent results, in that cases occur where an identical semantic contrast
will in some contexts produce zeugma and in other contexts will not, and, most
damagingly, contexts can be found where even the prototypical examples of alleged
ambiguity will not yield zeugma. One type of example is illustrated in (10):
The first occurrence of bank in A’s response to B is not felt by most people to
be zeugmatic, although it ought to be, if bank is truly ambiguous and the zeugma
test is reliable (so the argument goes). It ought to be clear that this is no problem
for the approach adopted here. Whether or not a sense boundary appears in par-
ticular cases is a matter of construal: ‘consistency’ would only be expected if the
boundary was an inherent property of the word, which our approach denies. Now,
in the case of bank, there are certainly strong conventional constraints favoring the
construal of a boundary, which means that in the majority of contexts a boundary
will be construed. However, in (10), no boundary is construed (the two readings
are ‘unified’ – see discussion below), and the above occurrence of bank is not
ambiguous. This occurs only in a very small range of context types. (Incidentally,
any two meanings can be unified in this way.) However, the fact that it occurs does
not constitute a valid reason for questioning the validity of the test.
construals; for this reason, the more stable the relations are, and the more sup-
ported by conventional constraints, the stronger the evidence they provide for the
autonomy of readings, and hence for the presence of a boundary. An example is
light in (11):
Here light has two distinct antonyms, dark and heavy. Cases such as light are
to be distinguished from cases such as old. Old might be held to have two distinct
antonyms in an old car (new) and an old man (young). However, there are good
reasons for viewing new and young as jointly constituting the antonym of old. New
and old have a common meaning component ‘has been in existence a relatively
short time’ which forms a satisfactory antonymic partner to an interpretation that
covers the range of possibilities for old, namely, ‘has been in existence a relatively
long time.’ The unity of old is supported by the lack of zeugma in (12):
The first mention of bank here represents a unification of the two units. The
context in this case enjoins us to find a distinctive unifying property for the two
kinds of bank. As it happens, this is not too difficult, although it requires a con-
ceptual shift to a metalinguistic level. The unifying factor is that both concepts are
designated/mapped onto by the same word form.
116 Cognitive approaches to lexical semantics
5.3.1 Facets
Facets are distinguishable components of a global whole, but they are not
capable of being subsumed under a hyperonym.
5.3.1.1 Introduction
Although facets display a significant degree of autonomy, they are not
generally considered to represent polysemy in the traditional sense; for instance,
they are rarely given separate definitions in dictionaries. Prototypical examples
of words with facets are the readings of bank (in the ‘financial’ sense) and
book:
(14) bank = [PREMISES] The bank was blown up.
[PERSONNEL] It’s a friendly bank.
[INSTITUTION] The bank was founded in 1597.
(15) book = [TOME] a red book
[TEXT] an interesting book
The items in (16) illustrate cases of words possessing facets analogous to
[TOME] and [TEXT]:
(18) Some of the books we read were novels and the others were biographies.
(20) *Some of the books [GLOBAL] were hardbacks, the rest were novels.
As (20) illustrates, novel and hardback do not coordinate in a context that expects
items to be construed as incompatibles within a single domain. The normality
of generic decontextualized statements such as (21a) and (21b) shows that the
potentiality for the construal of facets for book is well supported by conventional
constraints:
In these phrases, the adjective modifies only the [TOME] facet of book. In (23),
on the other hand, the adjectives modify only the [TEXT] facet:
Example (24a) can designate either two copies of the same text (i.e. two ‘tomes’)
or two texts; (24b) is interesting because the numerals two and one, respectively,
modify different facets (yet there is no zeugma); a new book may be a new copy
of a very ancient text, or a copy (whether in pristine condition or not) of a recently
composed text; a long book may be one with lots of words in it, or it may have a
non-canonical physical shape (notice that two long books or two new books have
to be long or new in the same way).
In an appropriate context, a question containing book can be truthfully answered
both in the affirmative and the negative, that is to say, responses can be relative to
one facet to the exclusion of the other. Consider (25):
referring equally to the [TOME] facet. In such a case, the potential exists for both
Yes and No to be true answers.
Both (26) and (27) are normally interpretable and involve different readings of
the book itself, namely, [TOME] in (26) and [TEXT] in (27). This will be taken to
indicate that book has two ‘cores’. Compare, however, the same procedure applied
to novel:
(28) ?I’m not interested in the plot, etc., I’m interested in the novel itself.
(29) I’m not interested in the binding, etc., I’m interested in the novel itself.
It seems that it is easy to isolate the [TEXT] facet of novel, but more difficult
to construe an autonomous [TOME] facet. This is also suggested by the relative
oddness (at least in the absence of more specific context) of a red novel, a dusty
novel, a shiny novel, compared with a red book, a dusty book and a shiny book
(we shall return to the ‘novel problem’ below).
Finally, each facet can have an independent proper name. Middlemarch is the
name of a text, not of a physical object, nor yet of a text-physical object complex.
(33) It’s a very helpful book, but rather heavy to carry around.
What is written is the text, but what is red is the physical object, and combining
the two seems to be cognitively difficult. It is not clear what the conditions are for
the appearance of zeugma here, since it does not appear to be present in (35):
(35) John wrote that red book on the top shelf over there.
It is not possible to publish something that does not comprise both a text and some
physical manifestation.
Polysemy 121
(37) This is a very interesting book, but it is awfully heavy to carry around.
In (37), interesting modifies the [TEXT] facet, and heavy to carry around the
[TOME] facet.
To interpret read in (38), we must access our knowledge of how texts are processed;
to interpret open, we need to access knowledge of books as physical objects.
In the case of book, it appears that the stable constraints do not push us strongly
in either direction, so we are (relatively) free to construe the facets either as au-
tonomous or as unified in response to other constraints, such as contextual or com-
municative ones. This is not so much because conceptual constraints are absent, or
inoperative, but because they act in contrary directions, effectively canceling one
another out.
One obvious factor pointing toward an autonomous manifestation of facets is
‘conceptual/semantic distance.’ As we have already noted, the more similar two
readings are, other things being equal, the easier it will be to unify them, and the
more difficult to construe a boundary between them; conversely, the more different
they are, the more difficult it will be (generally speaking) to unify them and the
easier it will be to construe them as autonomous. Clearly, a difference of basic
ontological type, for instance, ‘concrete’ vs. ‘abstract,’ represents a substantial
semantic distance, and will be expected to constitute a constraint opposing unifi-
cation. The facets of book, and those of every multifaceted word that have been
identified, are of distinct ontological types, and therefore it is no surprise that they
show autonomous tendencies.
The problem with facets is rather one of explaining why they are so easily
unified, given their conceptual distinctness. One possibility is that there is a coun-
terconstraint operative here. First of all, the facets typically co-occur in a range
of significant contexts – in fact, they prototypically co-occur; secondly, they do
not simply co-occur, rather, they operate in a kind of functional symbiosis. The
only reason for having the [TOME] is to concretely manifest the [TEXT]; the
text is useless without some physical manifestation. In Langackerian terms, we
can say that they are jointly profiled against a single domain matrix. This acts as
a strong constraint favoring unity. So, in the case of facets, we have significant
constraints favoring a unified construal and constraints favoring the construal of a
boundary. When this state of affairs obtains, whether a boundary is construed or
not in particular circumstances will depend on other factors. For instance, in (28)
and (29) above, the linguistic context motivates the hearer strongly to look for a
reading excluded by the first conjunct, i.e. to construe a boundary.
Cases like (43) are presumably similar (cf. example 1 in Fauconnier 1994:143):
(43) A: Pass me the Keats and the Wordsworth: Keats is red and Wordsworth is
green.
B: I can see a blue Keats, but not a red one.
These examples are convincing enough, but are hard to invent and clearly require
special contextual justification. But there are still facts about novel that need to
be explained. For instance, as mentioned earlier, why are the following odd out of
context, while the equivalent expressions with book are not?:
(44) a. ?a red novel
b. ?a dirty novel (in the physical sense)/?This novel is dirty
c. ?a dusty novel
Furthermore, certain contexts that are clearly ambiguous with book require a
considerable cognitive effort to see as ambiguous with novel:
(45) a new novel
two novels
Perhaps most strikingly, as has been already noted, the X itself construction has
two readings with book, but only one with novel:
(46) I’m not interested in the typography or the cover design, I’m interested in the
novel itself.
(47) ?I’m not interested in the plot or the characters, I’m interested in the novel itself.
At the very least, we can say that the [TOME] reading of novel, in spite of the
fact that it informs a good deal of our everyday interaction with novels, for one
reason or another, is much more difficult to construe than that of book.
Why, then, does novel behave in a different way from book? One approach to
answering this question is to look at features of meaning that distinguish items
124 Cognitive approaches to lexical semantics
that behave like book from those that behave like novel. It might be argued, for
instance, that what is distinctive about a novel is its text, not its physical format;
one might even say that the immediate default contrasts of novel are other literary
genres, like poem or short story. In the case of book, on the other hand, the facets
are perhaps more equally balanced. There are two sorts of contrast: first, with other
everyday physical objects such as clock, vase, ornament or in-tray, and second,
with other text-types, such as newspaper, magazine, brochure, directory.
There is a certain intuitive plausibility about this explanation, but on close exam-
ination, the matter is much more complex. Take the case of dictionary. The most
distinctive property of a dictionary is the text it contains. Yet dictionary behaves
not like novel but like book in that, for instance, two dictionaries is ambiguous
between ‘two different dictionary-texts’ and ‘two copies of the same text.’ At this
point one might point out that while a novel does not have a distinctive format,
a dictionary does, at least prototypically, and this might account for the higher
salience of the [TOME] facet. There are two objections to this. One is that there
are other book-like entities with distinctive formats that behave like novel: one of
these is thesis, which at least in a British university has a highly distinctive for-
mat. Yet two theses can only mean two different texts, and not two copies of the
same thesis. The other objection is that novels do prototypically have a distinctive
format, or at least there are possible book formats that are unlikely to be novels:
think of the characteristic large-format art books, or atlases.
A different line of argument is to say that there is no novel problem: novel behaves
as one might expect. The problem lies with dictionary in that it unexpectedly
behaves like book in spite of being defined by its contents. At this point, the
example of bible is perhaps relevant: a bible is defined by its contents, but two
bibles normally refers to two copies of the same text. This is presumably because
the text of a bible is (in everyday experience) unique – there are no others of the
same type. There is some plausibility in the suggestion that, to an ordinary person,
the same is true of ‘The Dictionary.’ That is to say, there is a naive assumption
that there is only one text, and this forces a [TOME] interpretation of plurality
(something similar would occur with, for instance, two David Copperfields). (The
fact that one says Look it up in the dictionary at least as readily as Look it up in a
dictionary is confirmatory evidence.)
There is still a problem, though. Consider the following cases:
Just as novel is defined by its text, secretary, barman and professor of linguistics
are defined by their jobs, yet there appears to be no prohibition on the use of
Polysemy 125
‘irrelevant’ adjectives, as there is with novel and thesis. (It may be that this is an
aspect of the wider problem of restrictions on active zones: why, for instance, can
we say Mary is fair, meaning that she has fair hair, but not ?Mary is blue [without
elaborate contextualization], meaning that she has blue eyes?)
However, out of a typical class of student informants, only about half will accept
(50c) as normal. In contrast, there is unanimity concerning the normality of (51c):
(51) a. A red book
b. A funny book
c. You’ll find that red book on the top shelf very funny.
This suggests that the components of factory are less integrated than those of
book. However, there seem to be cases of dual-nature concepts that are even more
integrated than book, and that show significantly less facet-like behavior. One ex-
ample is woman (many words referring to human beings are similar). Although
informants are intuitively not happy to put woman into the same semantic category
as book, it does shows some signs of facethood (the facets will be referred to as
[BODY] and [MIND], for convenience, but this carries no philosophical implica-
tions). For instance, there are adjectives that attach themselves to one facet or the
other:
(52) a. A tall woman [BODY]
b. An intelligent woman [MIND]
126 Cognitive approaches to lexical semantics
(53) I’m not interested in the woman’s body, I’m interested in the woman herself.
(54) ?I’m not interested in the woman’s mind or personality or feelings, I’m interested
in the woman herself.
Another example concerns the verb weigh. This involves a physical action allied
to a mental action:
(55) a. John weighed the potatoes with trembling hands. [PHYSICAL ACTION]
b. John weighed the potatoes accurately. [MENTAL ACTION]
5.3.2 Microsenses
Microsenses are distinct sense units of a word that occur in different
contexts and whose default construals stand in a relation of mutual incompatibility
Polysemy 127
at the same hierarchical level, rather like, for instance, the names of animals such
as cat, dog, sheep, cow, pig, horse. Typical examples of words with microsenses
are knife and card. Consider how these words are understood in the following
contexts:
(56) John called the waiter over to his table and complained that he had not been given
a knife and fork.
(57) The attacker threatened the couple with a knife.
(58) I got a card the other day from Ralph, who’s on holiday in Tenerife.
(59) Let me give you my card; let me know as soon as you have any news.
These are to be contrasted with the readings of knife and card that appear in (60)
and (61), respectively, which are hyperonymic to the readings in (56)–(59) above:
Words like knife and card have a hyperonymic reading and a cluster of hypony-
mous readings, whose default construals are sister incompatibles. All of these
units exhibit a significant degree of autonomy. It is the specific units, which are
here termed microsenses, that do the bulk of the ‘semantic work’ of the lexeme.
The hyperonymic construal has a secondary role. It requires positive contextual
pressure for activation: it is never the default selection (for instance, in [60] and
[61] any kind of and of various sorts are contextual triggers for the hyperonymic
reading). Furthermore, it is relatively ill-defined, and does not have an established
place in any lexical field. The words knife and card show what may be termed
default specificity, that is, when we encounter them, our first assumption is that
one of the specific construals is intended, and we look for evidence as to which
one.
A natural reaction at this point is to say that this is a purely pragmatic matter – the
different construals of knife and card in the above examples are no different from
those of friend in (62) and (63) (i.e. ‘female friend’ and ‘male friend,’ respectively):
However, there are significant differences. Firstly, in (62) and (63) we see merely
different contextual construals of a purport that is essentially neutral with respect
to sex. The readings ‘female friend’ and ‘male friend’ do not exhibit autonomy.
The neutral purport is inferentially enriched in different ways in (62) and (63) in
response to the contextual elements married my brother and married my sister.
In the absence of contextual pressure, the neutral reading of friend will be con-
strued. This type of construal, which does not involve the creation of autonomous
128 Cognitive approaches to lexical semantics
(65) Mother: (at table; Johnny is playing with his meat with his fingers): Use your
knife to cut your meat, Johnny.
Johnny: (who has a pen-knife in his pocket, but no knife of the proper sort) I
haven’t got one.
(66) Tom: (who has a football under his arm): Let’s play tennis.
Billy: Have you got a ball?
Tom: No, I thought YOU had one.
Of course it is the case here that context makes clear what sort of ball is relevant,
and this means that if there were no appropriate microsense available for selection,
contextual modulation would lead us to much the same meaning. It is important to
Polysemy 129
The constraint appears to be stronger in some cases than others: this may be
a reflection of degree of autonomy. But the contrast with (69a) and (69b), where
there is no constraint, is clear:
In (71a-b) there is no pressure for specificity, and they are not (automatically)
anomalous in contexts that do not sanction a more specific interpretation. A corol-
lary of the property of default specificity is that the hyperonymic readings of words
such as knife require overt contextual pressure for their activation: the mere ab-
sence of specifying pressure is not enough. Thus, although the questions in (70)
are to some degree odd, those in (72) and (73) are not:1
(72) Do you have a knife/card/ball of any kind?
(73) a knife wound; a knife-sharpener
1 The account of default specificity given in Cruse 2000a is not quite correct. There it was suggested
that questions like those in (70) are odd if the context does not sanction the selection of one of the
microsenses. However, it was also suggested that there are situational contexts, for instance camping,
for which there is no corresponding microsense. It would seem to follow from this that questions
like those in (70) are invariably odd in such contexts. However, this is not true: (i) seems normal:
(i) (in a camping context) I need to cut this rope. Do you have a knife?
Relevance considerations here narrow down the range of suitable knives: they have to be capable
of cutting the rope in question. But this looks like contextual modulation rather than the selection
of a microsense. It seems therefore that the default specificity constraint is satisfied by any sort of
specification, whether or not it involves a microsense, and it is the unrestricted reading of the hyper-
onym that requires overt contextual justification.
Polysemy 131
However, plastic and paper cards can be united in special contexts such as (76):
(76) The box contained a variety of plastic and other sorts of cards.
This will explain the seeming anomaly of sets of incompatibles sharing the same
word form. These readings are not really in competition with one another – the
sister units are only half-sisters. It also plausibly motivates their truth-conditional
autonomy. If the different microsenses were not domiciled in different domains, it
would be difficult to explain why, say, child does not develop microsenses corre-
sponding to ‘boy’ and ‘girl,’ since it must frequently be used with such reference.
It is possible, then, that mere frequency of occurrence is not enough to cause auto-
nomy to develop, but that some other differentiating factor must be present. In the
case of microsenses, this extra factor could plausibly be occurrence in different
domains; in the case of facets, although they are not differentiated by domain
(sister facets tend to occur together), they are differentiated by ontological type.
What does this account fail to explain? Two apparently contradictory things.
Firstly, it does not explain why immediate intuition goes for the hyperonymic
reading. In other words, people need convincing that the word shows default
specificity: unlettered intuition is aware more of the unity of ‘knife’ than of its
plurality. Secondly, given that this is so, what would we expect in a context that
does not sanction the selection of one of the specifics? Would we not expect the
hyperonymic reading to emerge? But it does not – it only appears when it is overtly
sanctioned, that is, one component of default specificity is not accounted for.
There is another way of looking at microsense complexes within an overall
Langackerian approach. The above approach implicitly pictures microsenses as
developing from an initial unified reading by ‘entrenchment + differentiating
factor.’ Perhaps this might explain the intuition of unity of words like knife, but it
is not a wholly plausible scenario. A more plausible scenario is that we start with
distinct readings that then become united. Langacker’s network model works like
this (cf. Langacker 1991: 266–71). We start with knife being used to designate a
particular kind of knife; then this is metaphorically extended to refer to a distinct
type of implement, with sufficient resemblance to the original to justify the use of
the same word; then a hyperonymic reading develops, which subsumes these two.
In this case, without taking differential entrenchment into account, the difference
in ‘richness’ of the resulting concepts may be governed by how much common
content can be abstracted from the specific readings. Obviously, if child developed
this way, then the common abstracted content would be rich, but in the case of
knife, much less so.
However, the ‘bottom-up’ scenario fails to explain the intuitive primacy of the
hyperonymic reading of knife. One possible explanation for this is that metalin-
guistic functions of language obey different rules to ‘actual situated use.’ Perhaps
for metalinguistic purposes, semantic similarity is the stronger constraint, whereas
in everyday use, domain allegiance is more important. Perhaps there is enough se-
mantic commonality in the various microsenses of a word like knife to fuel a unified
response in reflective metalinguistic use, but not enough to counteract the domain
134 Cognitive approaches to lexical semantics
Here a specific direct object has to be recovered from context by the hearer. Notice
that the identity constraint is operative here:
This is to be contrasted with, for example (79), where there is no identity con-
straint:
In answering (81), one is only obliged to answer with respect to the latent direct
object. Suppose the latent element is us; then we can answer ‘No’ to the question, as
long as John is not watching us, and it does not matter if he is watching something
else. Notice, too, that the autonomy can be overridden in contexts that make the
objects explicit; (81) is not zeugmatic:
This example of latency is of course syntactic in nature, and the latent element is
a separate word. Perhaps a little closer to present concerns is the example of cub:
this has a kind of default specificity, not unlike that of knife, which requires us to
identify which kind of cub is being talked about, if it is not made explicit. Thus,
for example, if an identification cannot be made, the result is odd:
(82) I was walking through the woods this morning and I spotted a cub.
Similarly (83) is odd if what I saw were a lion cub, a fox cub and a bear cub.
Polysemy 135
And different readings have independent truth-conditions. The question Did you
see any cubs? can be truthfully answered No if the subject of conversation/common
interest is foxes, and the speaker did not see any fox cubs, irrespective of whether
he/she saw any other kind of cub. Specifying the type of cub explicitly, removes
the restrictions and allows a hyperonymic reading:
(85) The lioness and the vixen were playing with their cubs.
This is to all intents and purposes ambiguous: French informants say they have to
intend one meaning or the other, one cannot be non-committal. Furthermore, there
is a strong identity constraint in (87):
However, the two readings can coordinate without zeugma, provided the sense
modalities are made explicit:
So perhaps the best solution is to propose a continuum ranging from the pure
latency of watch, where the choice of ‘filler’ is purely pragmatic, through cases
like cub where the choice is open along one dimension, but constrained in others,
to cases like knife, where the choice is restricted to a range of entrenched units.
Polysemy 137
But we are still lacking a satisfactory explanation for the difference between the
types.
5.3.3 Ways-of-seeing
What are called here ways-of-seeing are derived from what Pustejovsky
calls qualia roles (Pustejovsky 1995). These are in some ways analogous to the-
matic roles, but instead of detailing ways arguments may attach to a verb, they
govern ways in which predicates can attach themselves to nouns. Pustejovsky
proposes four qualia roles:
The constitutive role: this describes the internal constitution of the object, and
refers to such matters as material, weight, parts and components.
The formal role: the features under this heading serve to distinguish the ob-
ject from other objects within a larger domain, and refer to such matters as
orientation, magnitude, shape, color and position.
The telic role: this describes the purpose and/or function of the object.
The agentive role: the matters referred to here are, for instance, how the object
comes into being, whether an artifact, a natural kind, created by whom or
what, what causal chain leads up to it, and so on.
Qualia roles are reconceptualized here as ways-of-seeing (henceforth WOS) as
follows:
The part-whole WOS: views an entity as a whole with parts (e.g. a horse, as
viewed by a vet).
The kind WOS: views an entity as a kind among other kinds (e.g. a horse as
viewed by a zoologist).
The functional WOS: views an entity in terms of its interactions with other
entities (e.g. a horse as viewed by a jockey).
The life-history WOS: views an entity in terms of its life-history, especially its
coming into being (e.g. a book as viewed by an author or publisher).
The degree of autonomy of ways-of-seeing is less than that of facets or mi-
crosenses. Ways-of-seeing do not correspond to distinct concepts, and they are
not referentially distinct: they represent different ways of looking at the same
thing. There is some limited evidence of relational autonomy, which is one fea-
ture that differentiates them from the active zones considered below. Take the
example of hotel. This can be viewed as a piece of real estate, in which case it
will contrast with houses, offices, factories and so on. Or it can be viewed as a
type of accommodation, in which case it will contrast with B & B, Youth Hostel
and so on. Ways-of-seeing provide a possible explanation for certain cases of
138 Cognitive approaches to lexical semantics
ambiguous phrases with no ambiguous words and univocal syntax, that is to say,
they apparently show compositional autonomy:
There are problems in connection with ways-of-seeing. They are less well sup-
ported than facets or microsenses. It is not clear, for instance, that autonomy is
greater between perspectives than within ways-of-seeing. A number of cases of
WOS-like differential construals are difficult to assign unambiguously to one of
the proposed possibilities. Also, the number proposed seems to be largely arbi-
trary. At the same time, something of the sort is clearly needed: there is a need for
a construed pre-meaning unit intervening between facets and the purely composi-
tionally relevant features and active zones. It would be entirely consistent with the
approach adopted here to refrain from enumerating possible ways-of-seeing, and
to treat the above examples merely as particularly salient possibilities, where it is
relatively easy to construe a (minor) sense boundary.
(92) ([EQUINE][MALE])(overworked)
[EQUINE]([MALE](overworked))
That is, the animal is either overworked as a male or as a horse, generally. In other
words, a semantic unit [MALE] may be construed as separable, and can function
as scope for the predicate [OVERWORKED]. (Notice that the same construal is
not so easy for an overworked husband.) The two readings can coordinate without
zeugma:
that ‘mare’ is the default construal of ‘not a stallion.’ But, for instance, it is not
difficult to devise a scenario where not a horse will not be construed as ‘some non-
horse animal,’ but as ‘some non-horse means of transport’: the particular situation
will determine what is salient, and the possibilities would seem in principle to be
open.
Some active zones have the same sort of degree of autonomy as semantic com-
ponents, that is, they manifest themselves only in compositional terms. Like com-
ponents, they can give rise to ambiguity, as in red eyes (‘bloodshot eyes’; ‘pupil
is red, as in a photograph’ [potentially] ‘the iris is red’) and a red pencil (‘writes
red’; ‘red on the outside’). Notice that although the parts of a pencil selected as
the domain of relevance of the adjective red are readily distinguishable in reality,
and can be pointed to (unlike [MALE] and [HORSE] for stallion), a definite noun-
phrase like that red pencil, whichever interpretation is given, does not refer to the
part in question, but to the whole pencil.
6.1 Hyponymy
6.1.1 Introductory
The following are examples of linguistic expressions whose semantic
well-formedness depends on hyponymy (X is hyponymous to Y):
(1) Xs are Ys (Koalas are marsupials)
(2) Xs and other Ys (Koalas and other marsupials)
(3) Of all Ys, I prefer Xs. (Of all fruit I prefer mangoes.)
(4) Is it a Y?
Yes, it’s an X. (Is it a tit? Yes, it’s a coal-tit.)
(5) There was a marvelous show of Ys: the Xs were particularly good. (There was a
marvellous show of flowers: the roses were particularly good.)
All the above examples involve nouns. A similar relation can be found between
items belonging to other parts of speech:
141
142 Cognitive approaches to lexical semantics
However, pairs of lexical items related by hyponymy are far more frequently found
among nouns than among adjectives or verbs.
In principle the relation of hyponymy is a simple one (contrary to the view
expressed in Cruse 2002b), and can be seen as an instance of the operation of
the Lakoffian container image schema (see, for example, Lakoff 1987:271–73).
Hyponymy can be regarded as simple class inclusion:
The notion of inclusion can also be pursued intensionally. This yields some
insights and is more directly related to semantic concerns, since we are dealing
with meanings. It therefore promises greater explanatory power, but at the same
time raises some problems. To characterize hyponymy intensionally, we need to
say something like:
In one sense this is obvious enough. For instance, if we assume that the relation
of hyponymy applies to semantically composite expressions as well as to single
lexical items (this convention is not universally followed, but it is harmless), then
a red hat is hyponymous to a hat in that the class of red hats is a proper subset of
the class of hats. Intensionally, we can see that the meaning ‘a red hat’ contains
as a proper part, the meaning ‘hat’ (taking a naive view of compositionality). This
picture seems readily transferrable to single lexical items such as stallion and
horse: the meaning ‘horse’ is a proper part of the meaning ‘stallion,’ which can be
analysed as ‘horse’ + ‘male.’ (This does not conflict with what was said in chapter
5, as long as ‘male’ and ‘horse’ are viewed as potentially autonomous components
of a particular construal of ‘stallion,’ and not as inherent properties of the words.)
One way of testing for meaning inclusion is in terms of entailment relations be-
tween sentences (more properly, propositions) containing the relevant lexical items
in corresponding structural positions. One formulation, which is not restricted to
nouns, goes as follows:
This sort of definition, but not in this exact form, was first put forward by Lyons
(1963). Here, F(–) is a sentential function satisfied by X and Y.
Thus, the fact that It’s a dog unilaterally entails It’s an animal indicates that
dog is a hyponym of animal. The reasoning behind this definition is obvious. The
entailment definition has the advantage that it does not require us to specify what
the components A, B and C actually are. Unfortunately, it also has a couple of
disadvantages. First, a hyponymous relation between X and Y does not always
lead to entailment between F(X) and F(Y) (Basil became a Catholic does not
entail Basil became a Christian – he may have started out as a Baptist). Secondly,
entailment between F(X) and F(Y) does not guarantee hyponymy (The wasp stung
John on the knee entails The wasp stung John on the leg – knee is not a hyponym
of leg.) However, there is no reason in principle why these disadvantages cannot
be overcome by a more careful formulation of the test.
speakers find perfectly normal. The claim here is that the construals of dog and pet
in this context are such that dog is a hyponym of pet, in that the class of dogs is a
proper subpart of the class of pets. This is in spite of the fact that most speakers
judge that the truth of This is a dog does not guarantee the truth of This is a pet.
How can these two apparently contrary positions be reconciled? They are recon-
cilable because two different construals of dog are involved. The construal in This
is a dog, therefore it is an animal involves some sort of default construal of dog
(and animal), either one that emerges in minimal contexts or one that is triggered
by the ‘logical’ domain evoked by the fairly unfamiliar sentence type; the construal
in dogs and other animals, on the other hand, is strongly constrained by the X and
other Ys format, which requires the construals of X and Y to be adjusted so that
hyponymy holds. The adjustment can affect either X or Y or both: what emerges
if there is no other context is the result of the easiest adjustment, that is, involving
the most easily accessible construals (there may also be an additional factor of the
contextualizability of the result). For instance, the construal of dog in dogs and
other pets is probably more specific than the default zero context construal, but
is easily accessible; in handbags and other weapons, however, the construal of
weapons is less specific than the default zero context construal.
Construability is not infinitely flexible. Sufficiently strong conventional con-
straints can prevent hyponymous construals from emerging. For instance, dogs
and other cats is virtually unconstruable in any imaginable context (at least it is
not literally construable). There are other constraints, too. For instance, dogs and
other dogs is not acceptable, even though there are readings of dog that would
satisfy the requirement of hyponymy. This is perhaps because of a constraint that
discourages the repetition of a form if a different construal is required.
As defined above, hyponymy is a transitive relation, based on containment,
which is also transitive. There are claimed examples of nontransitivity in the liter-
ature, which our conception of hyponymy does not permit. The apparent cases of
transitivity failure provide interesting examples of different construals of the same
word in different contexts. Two such examples are as follows:
the category boundaries of Y. This assumes that X and Y are lexical items, and that
they denote fixed conceptual categories. Our account is different, as it assumes
that informants’ judgements are based on construals in context. We suggest that
hearers are predisposed to look for nonanomalous interpretations of input utter-
ances, and hence an utterance of X is a type of Y will be judged acceptable if there
are easily accessible construals of X and Y such that X is hyponymous to Y. For
instance, A car seat is a type of seat is judged normal because there is an accessible
construal of seat that includes car seats. The sentence A seat is a type of furniture is
judged normal because there is a construal of seat (probably the everyday default
construal) that excludes car seats, and is hyponymous to furniture. Hence, from
our point of view, these examples do not illustrate a breakdown in transitivity.
A similar account goes for (9). We are not aware of any convincing examples of a
genuine breakdown in the transitivity of hyponymy.
Sense relations have traditionally been viewed as relations between items po-
tentially occurring in a fairly strictly defined ‘paradigmatic slot.’ This viewpoint
has cognitive and communicative validity. The choice of any term out of a paradig-
matic set carries implicit information about aspects of its meaning that are shared
with other possible choices, information about meanings that are excluded, and it
opens up a particular range of more specific meanings, any or all of which may
be important to the message being transmitted. However, the traditional assump-
tion that words have inherent meanings has the consequence that paradigmatic
sets and the interrelations among members have been viewed as relatively stable
structures. Here, while the paradigmatic viewpoint is accepted as valid, the items
in the paradigm are not lexical items but contextual construals of lexical items, and
the relationships are relations between a particular construal of the item actually
chosen and potential construals of other items that might have been chosen in that
context.
For instance, consider the following exchange (the reader is invited to imagina-
tively construct a fuller context):
A full understanding of significance of dog in B’s reply needs at least (a) a construal
of a domain of potential noise-producing agents, given the context, (b) a construal
of what is excluded (the use of just implies that there are potential alternatives to
dog that would have had more serious consequences) and (c) a construal of the
range of possibilities opened up by dog in this context. Relations of hyponymy
and incompatibility hold between these various construals.
The paradigmatic viewpoint underlies the entailment approach to sense rela-
tions, under which relations are defined in terms of truth-conditional relations
146 Cognitive approaches to lexical semantics
What result would this approach lead to in the case of dog and pet? At this
point we encounter a certain indeterminacy in the notion of a default construal.
It is arguable that the unmarked notion of ‘dog’ for the average western city-
dweller is the domestic variety, in which case dog will have to be considered a
bona fide hyponym of pet. But is this a truly context-free construal? And if not, is
there such a thing? There is another possible approach. The construal of dog that
includes wild dogs and is thus hyponymous to animal, but not to pet, has some
sort of scientific basis. The arguments for regarding this construal as the default
version do not seem particularly strong as it is probably of relatively infrequent
occurrence. But perhaps there is a need for what one might term a ‘core’ construal,
which emerges naturally in some way from purport. If we apply this notion to pet,
it seems clear that the core construal will exclude electronic pets. The fact that
the question Is an electronic pet a real pet? is interpreted restrictively is evidence
for this: real plausibly functions to direct our attention to the core. This, however,
leaves unanswered the question of the nature of the core.
6.1.4 Taxonymy
Lyons states that taxonomic lexical hierarchies are structured by the rela-
tions of hyponymy (class inclusion) and incompatibility (class exclusion). This is
true as far as it goes, but it is necessary to make a distinction between two relations
of inclusion. The first is the relation that is exemplified in An X is a Y (which
corresponds to ‘simple’ hyponymy); the second is the relation for which An X is
a kind/type of Y is diagnostic, which is more discriminating than hyponymy, and
which functions as the ‘vertical’ relation in a taxonomy. In Cruse 1986, the sec-
ond relation is called taxonymy. It is perhaps noteworthy that Wierzbicka (1996)
includes ‘a kind of’ as one of her semantic primitives, and does not subject it to
further analysis; here, an attempt is made to elucidate its nature. (The following
discussion draws heavily on Cruse 2002b.)
Taxonomizing is clearly not simply a matter of dividing a larger class into smaller
classes. Some logically impeccable subdivisions do not yield good taxonomies:
A significant number of good taxonyms, on the other hand, are not easily defin-
able in terms of their hyperonyms, and require encyclopedic characterization (for
instance, the difference between animal and horse cannot be captured in a single
feature). In Cruse 1986 it was suggested that single-feature definitions, in general,
are not good taxonyms because they do not create optimal categories. However,
there are problems with this analysis. It is not difficult to find cases where a sat-
isfactory taxonomy seems to be founded on a single-property division. Take the
case of spoons: these are taxonomized on the basis of what they are used in con-
nection with (teaspoon, coffee spoon, soup spoon etc.). It is significant that neither
large spoon, metal spoon, round spoon nor deep spoon is a satisfactory taxonym
of spoon:
Perhaps the explanation lies rather in the nature of the features? Some features
are conceptually ‘simple’ (LARGE, ROUND etc.), others are more ‘complex’
(FUNCTION): perhaps complex features are better able to support taxonomy?
What, then, about the subdivision of blondes, which relies on shade of hair color
(An ash-blonde/strawberry blonde is a kind of blonde)? The fact is, there are
reasons to believe that the problem of taxonomy cannot be solved merely by
looking at the nature of the subcategories, for the following reasons.
Hyponymy and meronymy 149
Firstly, take the domain of BLONDES. Ash-blonde and strawberry blonde are
satisfactory taxonyms of blonde, but they are by no means the optimal subcate-
gories of BLONDE. We can get much ‘better’ subcategories than ASH BLONDE
and STRAWBERRY BLONDE by dividing the domain into BLONDE HOUSE-
WIVES, BLONDE DOCTORS, BLONDE TEACHERS, BLONDE LAWYERS
and so on, but these are not good taxonyms of blonde.
Secondly, suppose that a particular species of bird has a number of varieties
and a very marked difference between males and females. In such a case, it is
not inconceivable that a male/female division would yield the best categories, in
that the males of different varieties resembled one another more than the male and
female of a particular variety. However, even if that were the case, a sex-based
division would still be taxonomically ‘wrong.’
Thirdly, a given category may be a satisfactory subdivision of one superordinate,
but not of another. For instance, ?A lumberjack is a kind of man is not good, but
A lumberjack is a kind of manual worker is fine. If the crucial factor was the nature
of the resultant category, this would be hard to explain.
It is possible that the good category principle has a role to play in the charac-
terization of taxonymy, but it clearly cannot provide a full explanation.
It seems, then, that a satisfactory taxonym must engage in a particular way with a
particular aspect of the meaning of the hyperonym. In Cruse 1994, it was suggested
that taxonym and hyperonym must share the same perspective. The reason stallion
is not a good taxonym of horse, it was argued, is that it has a ‘sexual’ perspective,
while horse does not; the reason blonde is not a good taxonym of woman is that
it adopts a ‘hair-color’ perspective, while woman does not. Ash-blonde, on the
other hand, has the same ‘hair-color’ perspective as blonde, and that is why it is a
satisfactory taxonym. This notion of perspective needs to be made more precise;
it is not identical to the perspective as defined in §3.4, so to avoid confusion it will
be referred to henceforth as focal orientation.
Two possible lines of approach may be singled out. The first recruits Langacker’s
concept of profiling (see §2.2). The proposal is that what is profiled in a taxonym
must further specify what is profiled in the hyperonym. From this it will follow
that the reason stallion and foal are not good taxonyms of horse is that what they
profile, namely, ‘male’ and ‘young’, are not specifications of what is profiled in
horse; a similar explanation holds for blonde and woman. Likewise, lumberjack and
navvy are not good taxonyms of man because their profiles are not specifications
of what is profiled in man, but they are good taxonyms of manual worker, because
they specify further and profile what is profiled in manual worker, namely, type
of work. (It seems likely that there is also a constraint involving the relationship
between the domains evoked by taxonym and hyperonym, but it is not at present
clear what this is.)
150 Cognitive approaches to lexical semantics
described as the ‘part-whole’ relation. However, this is not strictly correct, and a
failure to separate the part-whole relation from meronymy leads to a great deal
of confusion. Meronymy is a relation between meanings, whereas the part-whole
relation links two individual entities. Of course there is an important connection
between the two, which will be examined in due course, but the distinction between
them must not be lost sight of. Meronymy is also bedeviled by the range of possible
construals of the notion of ‘part,’ some of which are relevant to meronymy and
some of which are not. We shall begin by examining the part-whole relation.
The basic notion here is the containment of one region or regions (interpreted in
the broadest possible manner) within another region. The boundaries of a contained
region must be neither identical with, nor must they transgress, the boundary of
the containing region. The notion of containment is, of course, a very basic form of
construal, which we can assume corresponds to Lakoff’s container image schema.
For reasons of space, we shall concentrate on central instances and assume that
both contained and containing regions are bounded and continuous.
(28) All the parts of the aeroplane were carefully packed into crates, ready for shipping.
(29) After the explosion, pieces of the aeroplane were scattered over a wide area.
152 Cognitive approaches to lexical semantics
The most accessible construals of parts and pieces in these two sentences are
distinct. Each unit referred to in (28) falls into one of a limited number of nonar-
bitrary categories which group together similar items from different wholes, and
which are defined by a characteristic set of properties such as shape, size and rela-
tionship to the whole aeroplane. We shall designate such items by the term part:
‘part’ is thus a hyponym of ‘portion.’
In contrast, the pieces referred to in (29) belong to a single category with only
one defining property, namely, that they once formed part of a whole aeroplane.
The label aeroplane pieces can be construed as uniting portions from different
aeroplanes into one category, but these pieces do not fall into distinct groups
with stable properties that unite similar pieces from different aeroplanes, and
they do not have definable relations, other than origin, with the whole entity.
(Pieces from a particular type of entity rarely have distinct lexical labels: the
example of shard comes to mind, but such cases are few and far between.) We
shall use piece as a technical term alongside part to designate portions that do
not qualify as parts: ‘piece,’ too, is thus a hyponym of ‘portion.’ Although most
pieces are not parts, and vice versa, they may be, accidentally, so piece and
part are not strictly incompatibles. (The technical term ‘piece’ covers a wider
category than the everyday lexical item piece: the waterlogged portions of the
garden and the illegible portions of the letter mentioned above are technically
pieces.)
The notion of recurrence (identifiability across different wholes) may need to
be further restricted. Suppose plates of a certain type always fracture in the same
way, yielding recurrent portions. Would they then qualify as parts? Arguably not.
Another characteristic of pieces that distinguishes them from parts is that they
are not contemporaneous with their canonically constituted wholes: there are no
pieces until the whole is destroyed, whereas true parts can be construed even in a
well-formed whole. The pieces of the plates that always fracture in the same way,
although they can be predicted, have no existence in the unbroken plate. (Even if
they were to be admitted into the PART category they would not be prototypical
parts, because they have no function relative to the undamaged whole.)
A further feature that distinguishes parts from pieces is what in Cruse 1986 was
called ‘autonomy.’ This is related to the notion of ‘spare part.’ Parts of artifacts
such as machines are often replaceable by functionally equivalent items: in such
cases, the replacement counts as, for example, a machine part, even though it has
never entered the constitution of a complete machine. (The same is perhaps soon
to be true of human body parts.) On the other hand, an exact replica of a piece of
a broken plate cannot properly be described as a piece of the plate. This property
of parts is presumably related to the fact that they prototypically have a species
of self-sufficiency, that is, cohesion plus distinctness, as ‘objects’ in their own
Hyponymy and meronymy 153
right. They are, in other words, subordinate wholes; pieces prototypically lack this
self-sufficiency.
The discontinuity and cohesion may involve any or all of: shape, texture, color,
internal structure, makeup. Detachability and/or independent movability are strong
indicators of discontinuity.
What is meant here is function, such as wings for flying or legs for perching
in a bird. There is another notion of function that is relative to human users. For
instance, the parts of an animal carved out by a butcher may not have distinctive
functions for the animal.
The relevant notion of ‘type’ here is difficult to pin down. One aspect is what
is usually called ‘ontological type.’ There is no agreement on a basic ontology,
but the sort of thing referred to by Jackendoff (1983), namely, THING, STATE,
PROCESS, EVENT, PROPERTY, TIME, PLACE and so on seem relevant to parts.
That is, the parts of a period of time, for instance, should themselves be periods of
time; the parts of a thing should be things (rather than, for instance, substances);
the parts of an event or a process should be subevents or processes; the parts of
154 Cognitive approaches to lexical semantics
an abstract entity should be abstract entities; the parts of a place should be places,
and so on.
There is another notion of ‘type’ that is relevant. There are two broad types
of part, called segmental and systemic parts in Cruse 1986. Segmental parts
are clearly spatially delimited and are typically encountered sequentially as the
whole is traversed; they may well have heterogeneous internal consistency. Typical
examples are the externally visible parts of the body such as arms, legs, head,
trunk and so on. Systemic parts are typically spatially interpenetrating, but they
are functionally distinct and typically have a greater internal consistency. They are
less likely to be perceptually salient. Typical examples in the human body are the
nervous system, the vascular system, lymphatic system and the skeleton.
The two types can be observed in other entities besides the human body.
A house, for instance, can be divided into segmental parts (rooms) and systemic
parts (brickwork, plumbing system, wiring, etc.); a string quartet (to take a non-
spatial example) has movements as segmental parts, and first violin, second violin,
viola and cello as systemic parallels (notice that we speak of the second violin part
quite naturally); a similar case could could be made for the scenes and acts of a
play (segmental), and the various actors’ roles (systemic).
the fingertips are composed of skin, nerve fibers, capillary blood vessels and so on,
and these are composed of various chemical substances that can be further broken
down into molecules, atoms, electrons and so on; at the other end of the chain, a
body can be part of a family or team, which is part of a population, which is part of
the terrestrial biomass and so on. Yet, intuitively, there is a self-contained system
beginning with fingertips and ending with the body.
Looking first at ultimate parts, it might well be the case in some instances that
no subpart is motivated – there is simply no subregion bounded by a sufficiently
salient discontinuity to justify the construal of a boundary. In other cases we
can appeal to the notion of type-consistency. One reason why we do not divide
fingertips into nerve fibers, capillaries and so on is that they are of the wrong type.
Ultimate wholes pose more puzzling problems. A crucial factor is undoubtedly
the existence of a major discontinuity with surroundings, coupled with internal
cohesion. However, there are a number of apparent inconsistencies. The following
examples arose in the course of class discussions as part of an undergraduate course
in lexical semantics. When student informants were questioned about pan and lid,
the majority said that the lid was not part of the pan, but some said they might
reconsider their judgement if one or more of the following applied:
(34) a. the lid was sold as a unit together with the pan
b. the lid was essential to the functioning of the pan
c. the lid was attached to the pan by means of a hinge.
When asked about teapot and lid, however, the majority said that the lid was
part of the teapot. If the lid was attached by a hinge, the judgment was unanimous.
When asked why the teapot lid was different from the pan lid, the most frequent
comment was that it was essential to the functioning of the teapot, and was sold
together with it. The screw caps of soft drink bottles and the cork and capsule
of a wine bottle were unanimously rejected as parts of the bottle, in spite of the
fact that these are arguably essential to function, and bottles are normally sold
complete with caps. A possible confounding factor is that they are normally made
of different materials from the bottle, which gives rise to a salient discontinuity.
One final type of example will be mentioned. Most informants reject battery
as a part of flashlight, whereas bulb is unanimously accepted as a part (there are
many similar cases). This is reinforced by the fact that one normally has to buy the
batteries separately when one buys a flashlight, but one does not expect to have to
buy a bulb separately. Furthermore, the description of a flashlight in a shop will
say battery included if it is included in the price, but will not say bulb included.
This is strange, because the battery is contained within the body of the flashlight,
and is essential to its functioning.
156 Cognitive approaches to lexical semantics
(35) There were scratches on the hand, but not on the arm itself.
(The hand is part of the arm.)
(36) The monitor is faulty, but the computer itself is OK.
(In the case of an iMac, the monitor is part of the computer.)
The expression the X itself selects some sort of core. In the case of the part-whole
relation, what seems to be selected is the smallest portion that can be construed as
‘a whole X’ – any smaller unit Y can only be construed as a part of X (although it
can of course be construed as a whole Y): certain (genuine) parts may be stripped
off without completely destroying wholeness. We shall call the smallest possible
portion of an X that can be construed as a whole X the core part. (The term
‘core’ is deliberately used to suggest a parallel between core parts and the core of
a category.) The factors that determine the existence and boundaries of a core part
in particular cases are not clear. In the case of the iMac, the fact that most PCs have
separate monitors may influence our construal of the iMac. In any case, the core
part may be presumed to need adequate motivation as a whole in its own right.
The notion of core part overlaps with, but is not identical with, the ‘main func-
tional part,’ which can function as active zone in a use of the name of the whole
(see Cruse 1986). For instance, a stainless steel X is often ambiguous between an
X that consists entirely of stainless steel, and one whose main functional part only
consists of stainless steel: consider a stainless steel screwdriver/hammer/knife, any
of which may have a handle made of some other material. This only applies to
certain privileged parts, however: a screwdriver with a tungsten head and a stain-
less steel handle would not qualify as a stainless steel screwdriver. My intuitions
are that the description ‘a totally new type of computer’ would be valid for an
iMac whose central processor had been redesigned, but not for one whose only
new component was the monitor portion: in this case, the main functional part and
the core part are the same. However, it is not clear that the part of a screwdriver
that engages with the screw can be called the screwdriver itself, so in this case the
main functional part and the core part are distinct.
This distinction was then recruited as part of an explanation for one type of
so-called ‘transitivity failures’ in the part-whole relation, as in:
From the point of view of the current approach, both the distinction and the
transitivity failures are arguably illusory. Consider, first, the proposed distinction
between integral parts and attachments. Notice, first, that there is nothing odd in
principle about a part being attached to a sister part: we can say, for instance,
without oddness, that the arm is attached to the trunk at the shoulder. Second,
there is an easily accessible construal of arm that excludes the hand:
(43) There were burns on the victim’s hand, but none on the arm.
(44) He had a tattoo on his arm (contrasts with a tattoo on the hand).
(45) A broken arm.
Hence, the most obvious explanation of the normality of (38) is that arm receives
the construal that excludes hand. In (37), on the other hand, arm receives the
inclusive construal. Hence, (37) and (38) contain different construals of arm: the
hand is attached to arm1 , which is a sister part, but not to arm2 , the whole that
includes it. If one makes the effort to construe arm in (38) as arm2 , it becomes
as anomalous as (40). Turning now to (39) and (40), a slight twist is needed in
the explanation. Of course, the handle of the spoon is attached to the ‘bowl’ of
the spoon; the reason (40) is odd, however, is that there is no construal of spoon
that designates only that part. (This account of the oddness of [41c] and [42c]
differs from Langacker’s account as presented in §2.4. The two accounts are not
necessarily mutually exclusive.)
158 Cognitive approaches to lexical semantics
What about the apparent transitivity failures? It should perhaps first be acknowl-
edged that these are weird sentences: how many of us have encountered them in
naturalistic settings? They are also virtually uncontextualized, so one should hes-
itate before drawing general conclusions from them. The previous explanation
was that the relation cannot cross an attachment boundary, that is to say, the parts
of an attachment do not count as parts of the whole of which the attachment is
part. The evidence for this was extremely limited, not to say contrived. There is,
however, an alternative explanation if we assume that the exclusive construal of
arm is the default (i.e. arm1 ), then we can say that (41a) and (42b) exemplify the
inclusive construal of arm (i.e. arm2 ), but that (41c) and (42c) represent the ex-
clusive (default) construal. In support of this account of the ‘transitivity failures,’
we would claim that (41c) and (42c) improve considerably in normality if one
makes the effort to give arm the inclusive construal. The general conclusion is that
these examples give no grounds for impugning the transitivity of the part-whole
relation. However, an explanation is still required of why the inclusive construal
is not evoked in (41a) and (42c).
The account of the other transitivity failure given in Cruse 1986 is nearer the
mark, but can be more felicitously expressed using the current approach. It concerns
the following examples cited originally by Lyons as an unresolved puzzle:
There are two issues here: the first concerns the logical validity of the conclusion
expressed in (46c) and (47c) based on the respective a and b sentences as premises;
the second issue is the normality of (46c) compared with (47c). Let us look first
at the logical question. The basic principle is simple: the conclusions are valid if
the construals of all terms are consistent between premises and conclusion. This
is the case in (46), so the conclusion (46c) is valid. In (47), on the other hand,
there is a discrepancy between the construal of handle in (47b) and (47c). The
word handle is typical of those exhibiting microsense behavior: when we hear
the word handle, in most contexts, unless a hyperonymic construal is explicitly
triggered, we look for a subtype of handle that is intended (default specificity),
and the resultant subtypes manifest the properties of microsenses. The contextual
frame The X has a handle elicits an appropriate microsense, that is, ‘an X handle.’
Hence, The door has a handle elicts the construal ‘door handle,’ while The house
has a handle elicits the microsense construal ‘house handle.’ This discrepancy
Hyponymy and meronymy 159
destroys the logical validity of the conclusion. (In [46], ‘sleeve cuffs’ are identical
with ‘jacket cuffs.’)
The situation with regard to normality is not entirely clear in these examples.
Out of context, there is probably a general tendency to a slight oddness if a part-
whole statement misses out an intervening term, especially if it is salient, hence
we might expect (46c) to be a little hard to contextualize. The reason it is not may
be due to another principle, namely that one can always relate parts at any level
to the overall whole: thus, for instance, The human body has fingers and fingers
are parts of the body are not so odd as The arm has fingers. The normality of
(46c) may therefore be due to the fact that jacket is construed as the overall whole.
The oddness of (47c) is more than we would expect simply from a level-skip:
as suggested in Cruse 1986, it is more plausibly explained as being due to the
fact that houses do not usually have handles, and it is difficult to imagine what
purpose they might serve (Cruse [1986] pointed out that substituting doll’s house
for house considerably reduces the oddness, but does nothing for the validity of
the conclusion.)
6.2.2 Meronymy
Like every other sense relation, meronymy is viewed here as a relation
between contextually construed meanings (or more precisely, by pre-meanings
created by boundary construal). However, the relationship is less straightforward
than hyponymy, and it is not easy to select the optimum way of expressing it.
The problem stems from the fact that the essential relation, the part-whole rela-
tion, does not hold between construed classes of elements, but between specific
individuals belonging to those classes. Also, the relation itself is subject to con-
strual, unlike the hyponymic relation between two classes. Given two classes, the
definition of hyponymy can decide whether hyponymy holds or not: there is no
need for a separate construal of the relation. In the case of meronymy, on the other
hand, a part-whole relation between two entities is itself a construal, subject to a
range of conventional and contextual constraints. So, let us examine the following
characterization of meronymy:
Hence, finger is a meronym of hand because for every entity properly describ-
able as a finger (in the default construal), there corresponds some entity properly
describable as a hand (also in the default construal), of which it is construed as a
part. This characterization seems to capture something essential about meronymy.
160 Cognitive approaches to lexical semantics
But it is too restrictive as it stands, because it does not allow for ‘spare parts,’ which
have never participated in any whole (of the relevant sort). We might therefore relax
the characterization to admit potential part-whole relations:
This now covers spare parts, but it arguably still misses something essential
to meronymy, namely, the construal of partness as an essential element of the
construal of the part. Take the case of lake and park. Many parks have lakes
within their boundaries, and in such cases the lake would be construed as ‘part
of the park.’ Furthermore, every lake is potentially a part of some park, so by the
second definition, lake would qualify as a meronym of park. There is, however,
a difference between the relation linking finger and hand and the relation linking
lake and park. In construing lake we are under no pressure to construe it as being
part of something: in cases where a lake is part of a park, the ‘partness’ is imposed
on the construal as it were from the outside. In the case of finger, ‘partness of hand’
is an essential component of the original construal. Let us call this difference one
between an intrinsic and an extrinsic construal of partness. Then we can say that
the relation of meronymy concerns only intrinsic construals of partness:
(51) Nails had been torn from the victim’s fingers and toes.
However, the default use of nail denotes a microsense, one for each related
whole. Hence, B’s answer in (52) is acceptable as true:
(52) A: (examining B’s hands) Have you cut your nails this week?
B: (who cut his toe-nails the previous day, but had not cut his finger-nails for a
long time) No.
The second relation is illustrated by body in relation to penis and vagina. Here
it is the holonym that has the greater ‘range’: penis and vagina are parts, but in
relation to different subclasses of body. However, in contrast to the first case, there
is no evidence whatsoever that there are two microsenses of body corresponding to
‘man’s body’ and ‘woman’s body.’ This is interesting, and demands an explanation,
but the important thing for present purposes is that it is not a difference at the level
of the part-whole relation.
There are, however, difficulties that cast doubt on the utility of meronymy as a
lexical relation. Let us take the example of the word lid. The purport of this word
takes us into the conceptual area of containers, access points into containers and
means of controlling movement into and out of containers. There seems to be a very
strong conventional constraint forcing some sort of construal of incompleteness:
a lid is designed for use in conjunction with a container. However, this is not an
obligatory construal of partness, since not all lids are parts (cf. bottles, jars, teapots
etc.). There is undoubtedly a possibility of construing lid in a superordinate sense:
(53) In this box you’ll find a lot of things to put things in, and in the other box, a lot
of lids: your job is to sort out which lid goes with which container.
Here, a particular type of container is referred to. Perhaps, then, we should recruit
the notion of microsense and say that lid has microsenses (‘teapot lid,’ ‘jamjar
lid,’ ‘pan lid’ etc.), some of which involve intrinsic construal of partness (cf. Cruse
1986). However, this will not work, because, as was mentioned earlier, whether,
for instance, a teapot lid is construed as a part or not depends on factors such as
the presence of a hinge, whether sold separately and so on. In other words, we can
162 Cognitive approaches to lexical semantics
only be sure of a construal of partness at the level of the individual item. Even the
local construal of ‘lid type’ is not necessarily specified as to partness. (There is no
evidence for ‘hinged lid’ and ‘free lid’ as distinct microsenses or perspectives.) The
same phenomenon can be observed with a wide range of words. The implication
seems to be that we cannot in general deal with the part-whole relation except
at the level of the individual referent. In (53), for instance, the relation between
container and lid is indeterminate. In other words, the whole notion of meronymy
as a lexical relation is dubious.
The picture of meronymy that seems to emerge from the above discussion is
the following. There is, first, a very indeterminate purport, then a series of pre-
meaning construals that take us nearer and nearer to the target construal, and may
involve a commitment to partness at some point before the final construal, but in
many cases the part-whole relation cannot be inferred until we reach the level of
individual referents.
This is a very awkward picture from the point of view of lexical semantics:
there seems little to be said at the lexical level. One conclusion from this might
be that meronymy, unlike, say, hyponymy or antonymy, is not a relation between
meanings. After all, strictly speaking, the parallel at the individual level of the
class-based relation of hyponymy is the part-whole relation, and the strict parallel
to incompatibility is a relation between sister parts: incompatibles denote sets with
no members in common; co-parts are, in general, parts of the same whole which
have no substance in common (provided they are of the same type – segmental parts
may share substance with systemic parts). Does this mean that we can dispense with
meronymy altogether? There are reasons why this is not an entirely satisfactory
proposal.
Perhaps the most cogent arguments in favor of retaining meronymy concern
its intuitive appeal, the fact that children learn the names of parts early, the fact
that all languages have names for parts of things, and, further, that cross-linguistic
generalizations can be made regarding the naming of parts, especially parts of the
human body (see Brown 2002 for a survey).
One such generalization concerns the equivalents in various languages for the
English hand. In many languages this denotes the region of the arm from the
fingertips to the elbow. (This is the case of the Modern Greek to xeri. However,
Greek speakers assure me that it can also be construed to refer to the region
from the fingertips to the wrist only, and that there is rarely any confusion.) The
generalization is that, where a language has an ‘extended’ hand, it is highly likely
also to have word for ‘foot’ that designates the region of the leg from the toes to
the knee (this is the case with Modern Greek to podi). Another generalization, or at
least strong tendency, is that extensions from hand to foot are common, extensions
from foot to hand are much rarer. Thus, the basic meaning of the French doigt is
Hyponymy and meronymy 163
‘finger,’ but we also have doigt de pied for ‘toe’ (alongside orteil). However, this
is not an absolute rule, as the heel of the hand is well established in English.
Finally, there is a hard core of word pairs, like finger and hand, that are extremely
difficult not to construe literally with a relation that satisfies our definition (but
they are remarkably few in number).
The situation with regard to parts and wholes and their linguistic expression can
therefore be summarized as follows:
(i) The part-whole relation properly applies to individual entities. It is a
construal that is subject to variation.
(ii) The recognition of a relation of meronymy between construals is justified
by the existence of a small number of generalizations and distinctions that
only apply to classes of parts.
(iii) Every language has a range of ways of referring to parts of things. Many
of these ways involve specialized lexical items, but apart from a very
restricted core set of strict lexical meronyms, the relations between these
and expressions for whole things are very various.
7
A dynamic construal approach to
sense relations II: antonymy
and complementarity
7.1 Oppositeness
There are undoubtedly different construals of opposite involved here, but intu-
itively they belong to a family. None of them are metalinguistic. But even speakers
innocent of semantic theory have robust intuitions about lexical opposites, and
even quite young children rapidly catch on to the idea. Like all sense relations,
oppositeness is a matter of construal, and is subject to cognitive, conventional and
contextual constraints.
There seem to be two main components in a construal of oppositeness. The first
of these is binarity. Opposite meanings are construed as mutually exhausting some
domain: within the appropriate domain, there are only two possibilities. Some
domains are difficult to construe in any other way. For instance, there are only two
directions along a linear path, so up and down, and forwards and backwards are
natural opposites; likewise, there are only two extreme points on an axis, so top
and bottom, and front and back are also natural opposites. At a more abstract level,
there are only two directions of change between two states, that is, from A to B
and from B to A. This confers oppositeness onto dress:undress, tie:untie, and the
like.
These examples have a kind of built-in logical twoness. But a binary opposi-
tion does not have to be inherently logically watertight, only locally construed as
such. Take the case of town and country (urban:rural), which are often used as
164
Antonymy and complementarity 165
(c) Symmetry
For instance, large and tiny are not such good opposites as large and
small (although the sense of oppositeness is not completely absent). One possi-
ble reason for this is that large and tiny are not symmetrically disposed about
the reference point. The residual oppositeness of large and tiny may be due to
their counterdirectionality: when intensified, they move apart on the scale of
size.
Antonymy and complementarity 167
7.2 Complementarity
some entity is X, then it is not Y, and if it is not X then it is Y. The notion of domain
is essential to the understanding of complementarity. Take the pair married:single.
We can say that if someone is not married, then they are single, and if they are
not single, then they are married. But this relationship is highly dependent on
the construed domain within which the two terms are operating. First of all, we
need to specify that we are talking about humans: the inference does not work for
angels or spiders (or chairs). Second, we must restrict the domain to ‘marriageable’
persons: the logical relation does not hold for babies (in Western societies at least)
or the Pope. Third, we must exclude noncanonical forms of cohabitation. Hence,
the logical properties only appear within an appropriately construed domain. In
some cases, the appropriate domain is subject to strong conventional constraints,
as in the case of the default construals of dead:alive: if someone says that John
is not dead, it will normally be inferred that John is still alive, that is to say, we
are strongly constrained to construe the domain in such a way that the relation of
complementarity holds.
The position of the boundary between a pair of complementaries X and Y is an
aspect of particular construals. In the case of dead:alive, the location of the bound-
ary is sometimes a matter of dispute, including legal dispute, and could well differ
in different discourse domains. In everyday language, John is dead can usually be
taken to mean that John’s state is well clear of the zone of uncertainty. It could
also, however, function argumentatively to indicate that John was on the wrong
side of the speaker’s construal of the boundary, that is, as part of a dispute (imagine
a discussion as to whether attempts to revive a patient should be abandoned). No-
tice that if one construes the notion ‘life-state’ to include zombification and/or the
vampiric state, and so on, as well as ‘ordinary’ death and life, the logical relation
of complementarity will not hold.
It is important to bear in mind that complementarity is a relation between con-
struals and not between lexical items: in many cases, properties can be construed
either in absolute terms or in gradable terms. In some cases, there is a species of
alternation, according to context, between a pair of complementaries and a pair of
antonyms; we shall postpone consideration of these until antonyms have been ex-
amined in detail. But there also appear to be cases where one term of an opposition
is construed as absolute and the other as gradable:
(6) A: Is John dead?
B: No, he’s very much alive.
Here, dead is construed absolutely and alive gradably. The two construals of
alive as gradable or absolute do not seem to amount to a difference between distinct
senses. There is, for instance, no zeugma in (7):
(7) A: Is John alive?
B: Very much so.
Antonymy and complementarity 169
It appears to be possible for the absolute term to function as a zero point on the
construed scale for the other term. That is to say, in (6), dead is construed as ‘zero
aliveness.’ This does not appear to affect the logical relation between the terms;
that is to say, they are still complementaries by the ‘not-X entails Y’ criterion.
7.3 Antonymy
become apparent below. The first division is between antonyms that involve a
single scale and those that involve two separate scales working in tandem.
A simplified version of a monoscalar system is given in Figure 7.1.
LENGTH
0 short long
Figure 7.1 A simplified monoscalar system
The scale denotes a single property, in this case length; the scale has an end
point denoting zero value of the property at one end and extends indefinitely in
the other direction. One term of the opposition is associated with a higher value
of the property and the other term with a lower value. The terms move in opposite
directions along the scale when intensified.
There are three basic types of biscalar system, depending on the relative dis-
position of the two scales. In the equipollent patterns, the properties of the two
scales are fully symmetrical. There are two possibilities here: either the scales are
arranged end-to-end, and are completely disjunct, or they are completely overlap-
ping. The disjunct type is exemplified by ‘hot:cold’ in Figure 7.2. Here we have
two independent scales which meet at their zero points and extend indefinitely in
opposite directions.
COLDNESS HOTNESS
cold 0 hot
Figure 7.2 A disjunct equipollent system
The parallel type appears to be rarer. Here the two scales run parallel to one
another over their whole length (Figure 7.3).
HARDNESS hard
0
SOFTNESS
soft 0
Figure 7.3 A parallel equipollent system
In the overlapping system, there is partial overlap between the two scales. At
the same time, the scales are not equal: there is a major scale (MERIT) and a minor
scale (BADNESS), as in Figure 7.4 on page 171.
Antonymy and complementarity 171
MERIT good
0
BADNESS
bad 0
Figure 7.4 An overlapping system
(ii) the ease of construal of a determinate end point for the scale
0
X-ness
Y X
Figure 7.5 A full monoscalar system
There are two basic ways of construing a quantity of something: we can either
look at it in absolute terms (e.g. 25cm, 1.7 kg) or we can view it as more or
less than some reference value (a long pencil, a heavy suitcase). These two ways
correspond to what we shall call the absolute scale schema and the relative scale
schema respectively. In the diagram, the upper scale is the absolute scale, and the
heavy line indicates the reference value (or range) for the relative scale. (Even in
the case of a single gradable adjective we need to postulate an absolute and a relative
scale, since gradable adjectives are prototypically relative in their basic use.)
Notice that, by the same criteria, the default readings of good, bad, hot and cold
are all supras: this is how we know that they are part of biscalar systems.
(9) a. twice as good/ half as good
b. twice as bad/ half as bad
c. How good was it?
d. How bad was it?
This is not satisfactory: abundance and paucity are nominals related to much and
little – in other words, we are explaining one pair of antonyms in terms of another
pair, and the essence of antonymy is slipping through the net. There is a way of
expressing the relationship without covert resort to antonymy in the definition,
in terms of the ordering of points on the scale of length (0 = zero length; R =
contextually determined reference point; L = length of object):
long: 0, R, L
short: 0, L, R
This shows short to be parasitic on long, and avoids the risk of surreptitiously
incorporating the problem into the solution, but it does not bring out the negative
nature of short, nor its greater cognitive complexity.
Probably, the essence of the difference between long and short lies in the fact that
‘longness’ has a direct correlation with the scale of length, whereas ‘shortness’ has
an inverse correlation, which renders ‘short’ inherently more complex than ‘long’,
but it is still not clear how this might be encapsulated in a definition.
In our terms, heavier is impartial with respect to the weight of the referents
that form its arguments, in that its use is not constrained to those contexts where
heavy is appropriate. The same is true of lighter, although a slightly more complex
test frame is necessary to establish this, as a ‘heavy’ argument displays a greater
affinity for heavier than for lighter (this factor, labeled ‘pull’ in Allen 1986, also
requires explanation):
(12) This box is quite heavy, but it’s still lighter than the other one, nonetheless.
Committed uses of heavy and light are illustrated in (13) and (14):
Other examples of impartiality (the [a] examples) and committedness (the [b]
examples) are:
For languages other than English, another type of question shows impartiality
effects. Compare (19a) and (19b) in French and (20a) and (20b) in Modern Greek:
In each case the (a) question is in some sense ‘open-minded,’ compared to the
(b) question, even though the difference has no effect on the nature of the answers
to the questions – that is to say, the answer must reflect the truth value of the
implicit proposition in each case.
Antonymy and complementarity 177
(a) Comparatives
Both terms of every polar antonym, in all the languages for which data
are available, have an impartial comparative form. A committed comparative is
178 Cognitive approaches to lexical semantics
(21) Neither of them is very bright, but John is more intelligent than Bill.
Here we can see the effect of the exposure scale: slow in (22b) is more exposed
than in (22a), so (b) is committed and (a) impartial. In English, we can regard
the morphological comparative as the default variety, and the periphrastic form in
(22b) as somewhat forced. We might therefore expect the morphological variety to
attract the priority construal. Hence there are two competing explanations for (22)
(although they are not mutually exclusive). A clearer case is provided by Modern
Greek, where some adjectives have a more evenly balanced choice between a pe-
riphrastic comparative and a morphological comparative. The situation, however,
is similar to English: an adjective that has only a periphrastic form is impartial,
but where it is in competition with a morphological form, it is the latter that is
impartial, and the periphrastic form is committed.
These are quite different to ordinary comparatives. We may assume that any
mention of an explicit numeral or numerical quantity predisposes an absolute con-
strual of the adjective, at least partly because only the absolute scale can be cali-
brated. The oddness of short is presumably because there is no conventional way
of calibrating shortness. Quantified comparatives can be used with non-calibrated
scales (twice as good, only half as bad); presumably these are metaphorical ex-
tensions from the prototypical calibrated scales. Such extensions are limited as
to permissible quantifications: This stick is three and a half times as long as that
Antonymy and complementarity 179
one is normal, but ?Jane is three and a half times as pretty as Sarah is distinctly
odd.
Notice that (28) cannot be used to ask Is it long or short?, whereas this is a
possible meaning of How long is it? What is its NOM? seems to demand an absolute
quantity as an answer, and since shortness and the like cannot be quantified in this
way, that is sufficient reason for the anomaly of (29). The same goes for (30),
although here the reason is not that we are dealing with sub terms, but that we are
dealing with noncalibrated quantities (see §7.4.2.2). But why is it that this form of
question demands a quantity, whereas How ADJ is it? does not? A likely answer
is that what? demands a specific individual entity to be identified. The only way
of individuating a quantity of something is by its magnitude. That is fine once
we know that we are dealing with, say, the absolute scale of length. But that still
leaves the question of why we are limited to the absolute scale of length. One
possibility is that it is only on the absolute scale that there is the possibility of
180 Cognitive approaches to lexical semantics
(f) Is it X? questions
(32) French:
a. Elle est longue, ta nouvelle jupe?
b. Elle est courte?
(33) Modern Greek:
a. Ine makris?
‘Is it long?’
b. Ine kontos?
‘Is it short?’
Antonymy and complementarity 181
(34) Turkish:
a. O uzun mu?
‘Is it long?’
b. O kısa mı?
‘Is it short?’
Questions of this form containing the supra term (which seem not to occur
significantly in English) function in a similar way to the relative construal of
How X is it?, that is, to ask whether something is long or short. They occur
particularly in languages that do not have How-questions (e.g. French) or where
How-questions are committed (e.g. Greek and Turkish). The sub term in each
case yields a committed question. In both cases, the relative scale is involved,
and the reasons for choosing the supra term to encode the impartial question are
presumably the same in both cases, namely, that it is conceptually the simplest.
(35) This pan feels hot, but it feels cooler than that one/still feels cooler than that one,
nonetheless.
(36) This bottle feels cold, but it feels warmer than that one/still feels warmer than
that one, nonetheless.
(37) ?This pan feels hot, but it feels colder than that one/still feels colder than that one,
nonetheless.
(38) ?This bottle feels cold, but it feels hotter than that one/still feels hotter than that
one, nonetheless.
The sub term perhaps also occurs in a cool oven, a cool flame and a relatively
warm ice age. Cruse’s intuitions are that This flame is only half as cool as that one
is difficult to process, and resembles half as short. Notice, however, that reversing
the terms in (35) and (36) leads to anomaly:
(39) ?This pan feels cool, but it feels hotter than that one/still feels hotter than that
one, nonetheless.
(40) ?This bottle feels warm, but it feels colder than that one/still feels colder than that
one, nonetheless.
This suggests that the default readings of cool and warm are not subs of hot and
cold, and that the sub construal has to be coerced by contextual pressure. Wherever
the sub reading is appropriate, either inherently hot/cold items are involved (such
as ice-age, flame), or there is some other explicit indication that the domain in
question is one of hotness or coldness.
Parallel equipollent pairs are even less common than the disjunct type. They
seem to occur when (a) there is no property that can be construed as having a
Antonymy and complementarity 183
zero value in the middle of the scale, (b) each direction of construal is equally
motivated and (c) there is no dominant viewpoint, no difference that motivates
a ‘positive/negative’ construal. Hayes (2001) argues that hard:soft are plausible
candidates, and dark:light as applied to colors.
(41) John and Bill both got bad marks in the exam, although John’s marks were better
than Bill’s.
(42) ?John and Bill both got first class marks in the exam; John’s marks, however,
were worse than Bill’s.
(ii) Both members are normal in quantified comparatives; one member is system-
impartial ([43] and [44]) and the other system-committed ([45] and [46]):
(43) Both marks were admittedly bad, but Bill’s was twice as good as John’s.
(44) It’s true that Bill’s wasn’t the only bad mark, but it was only half as good as even
the other failures.
(An impartial reading of half as good is difficult to construe in [44], but it is
possible.)
(45) ?Both marks were excellent, but Bill’s was twice as bad as John’s.
(46) ?Both marks were excellent, but Bill’s was only half as bad as John’s.
gut:schlimm opposition.) Notice that although How bad is it?, worse and twice
as bad are system-committed, they are scale-impartial, since they presuppose no
particular value on the ‘bad’ scale. There is no single sub term for bad, but a
choice of contextually restricted possibilities, such as mild and slight: compare
?How mild is your gout? and ?How slight was your accident? with How bad was
your gout/accident?
Although absolute scale-schemas are necessary to explain the scale-impartiality
of better and worse, they are not calibrated, and this gives rise to differences with
polars. First, neither member yields a normal nominalization question:
(49) ?What was the goodness of the film?
(50) ?What was the cleanness of the room?
(51) ?What was John’s politeness?
(cf. What was the length of the rope?)
Given that there are no calibrated quantities with the scales of overlapping
antonyms, a relative reading seems adequately motivated.
This is what we would expect, with better showing impartiality. However, while
(58) is normal, neither (59) nor (60) is:
(58) The drought last year was worse than this year.
(59) ?The drought this year is better than last year.
(60) ?How good was the drought last year?
Antonymy and complementarity 185
It seems that certain items do not collocate with better or How good . . .?,
even though they collocate perfectly normally with worse. These items share the
property of being normally construed as ‘inherently bad’: accidents, illnesses,
famines, droughts, earthquakes and so on. The inherent badness (or whatever) of
some entity is, of course, a construal, and varies with contextual factors. Sentence
(61) is more likely in a conversation between two seismologists than between two
victims of the disaster in question:
(61) A: How was the earthquake?
B: Quite good – better than the last one.
A possible explanation of the inherentness effect utilizes the fact that the BAD-
NESS scale has only half the extent of the MERIT scale. Perhaps there is a kind
of Gricean implicature that the scale on which we place something is the smallest
that encompasses all the possibilities for a particular construed domain. Hence, if
we say, for instance, How good was the film?, we implicate that we are prepared
for an answer anywhere on the scale of MERIT. The reason we say How bad was
the famine, rather than How good was the famine?, is because the BADNESS scale
encompasses all the expected possibilities for a famine. How good was the famine
would implicate that we construe the domain of famines as potentially containing
good examples.
(Notice that both [64] and [65] imply that the shirt in question is at least slightly
dirty.) In general, a gradable construal of a property is indicated by the use of
the morphological comparative or superlative, or intensifiers such as more, fairly,
quite, rather, extremely and so on.
In (66), we have a hybrid situation, with dirty construed as relative, but clean
construed as absolute:
(66) A: How are you getting on with that very dirty pan?
B: Well, it’s almost clean – give me another ten minutes and I’ll have it clean.
The reason for this constraint, which appears to be cognitive, rather than con-
ventional, lies in the nature of the clean-dirty opposition. It represents what in
Antonymy and complementarity 187
Cruse 1986 was called a privative opposition, that is to say, one in which one
term denotes the absence of a property and its partner denotes the presence of the
property. In the clean-dirty opposition, clean denotes the absence of dirt, rather
than the converse:
(70) A: Is it dead?
B: No, look, it’s breathing – it’s still alive.
(71) You look rather more alive than you did half an hour ago!
(72) ?You look rather more dead than you did half an hour ago!
Hence, dead seems more tied to an absolute construal than alive (or, indeed,
clean). A hybrid construal is fairly normal (notice that it is not dead that is being
188 Cognitive approaches to lexical semantics
Figurative uses of dead may be less odd when construed as gradable, but this is
strictly irrelevant as far as literal construals are concerned:
(74) Every time I come back to this town it seems even more dead than the last time I
was here.
(75) I need to ask you a few questions. Are you married or single?
However, as I have noted elsewhere, Iris Murdoch has (76) in one of her novels:
Most speakers find this interpretable, but at the same time feel they are working
against palpable constraints in coming to the appropriate construal. Once one
accepts this construal, (77) presents no problems:
But it is much harder to construe (78) as representing the same relation between
Jane and Mary as (77):
This is a difficult question. From available examples, it seems that only supra
terms can support an absolute construal. However, not all supras have absolute
construals: *almost long/heavy/fast. Also necessary is some sort of natural deter-
minate value, frequently also construed as a zero. The reason for the absence of
absolute construals for sub terms is not at present obvious. It may have something
to do with the fact that they are impressionistic, and do not have a linear relation
with the absolute scale. That is to say, to get equal increments of impressionistic
shortness, we must, let’s say, successively halve the absolute length. That way, zero
will be approached asymptotically, but will never be reached. Put another way, the
default construal of the SLOW scale has no end. There is a parallel with impres-
sionistic scales of loudness and brightness, which do not have a linear relation to
absolute energy level.
Another possibility is that, in those cases mentioned, either the zero point on
the scale is of no practical communicative importance (what use have we for the
notion of zero length?) or the zero point is preempted by another lexical item (as
with stationary, weightless, free). (Of course this could work the other way round:
we have a separate lexical item for the zero point because the sub term cannot
cover it. This has some intuitive plausibility. More work is needed on this point.)
a scale of LENGTH rather than a scale of SHORTH, with long as supra and short
as sub. However, there are contexts where the roles of long and short appear to
be reversed. For instance, (85) and (86) appeared in a newspaper article about the
miniaturization of computer components:
(85) The new device is ten times smaller than anything seen previously.
(86) ‘The nanotube transistors are about 350 times smaller than a conventional silicon
transistor and faster.’ (Guardian On-line Supplement, October 10, 2002, p. 11)
And (87), in the same context, seems normal (although not actually attested):
It seems that in (85)–(87) the scale of linear extent is being construed in the
reverse direction from the default construal, as a scale of SMALLNESS rather than
a scale of BIGNESS/LARGENESS, with small as the supra term. This reversal
goes against quite strong cognitive constraints, and would not be possible unless
there were definite factors favoring it. In this case there are such factors, namely, the
fact that an increase in smallness is correlated with an increase in interestingness,
salience, desirability.
Many instances of scale reversal are correlated with a change of antonym type.
This is particularly so where the motivation for the reversal is related to evalua-
tiveness. Take the case, discussed earlier, of easy:difficult. It is not hard to con-
strue a scale of DIFFICULTY with difficult and easy as polars, difficult as supra,
and How difficult is it? as an impartial question. But an average class of under-
graduates finds it difficult to decide whether difficult or easy is the supra term,
with a significant number opting for easy, and an underlying scale of easiness.
This seems to be supported by the relative normality out of context of How easy
is it?
A significant determining factor seems to be whether ‘difficulty’ is judged neg-
atively or objectively. If ‘easy’ is positively evaluated, and ‘difficult’ negatively
evaluated, this encourages the movement of easy:difficult from the polar construal
to the overlapping construal, with both easy and difficult as supras, but easy as
the major term (parallel to good). A similar alternation in antonym type can be
observed with cheap:expensive (Hayes 2001). Where these are used simply to in-
dicate price, they behave more like polars, with cheap as sub; but where cheap has
connotations of ‘poor quality’ and expensive of ‘high quality,’ they behave more
like overlapping antonyms, with cheap being a (minor) supra term.
Antonymy and complementarity 191
A different alternation is found with thin, in opposition to thick and fat. There is
no doubt that thick:thin are polars, with thin as sub term; in other words, there is no
scale of thinness, only one of thickness (in the default construal). The majority of
native-speaker informants, however, class fat:thin as equipollents, out of context.
This involves reclassifying thin as supra, that is, construing a scale of thinness. The
motivation seems to be that there is a societal norm for girth, which is positively
evaluated, and significant deviations in either direction are negatively evaluated:
the evaluative feature has the effect of encouraging the construal of a scale of
thinness (more of the property being more salient).
Perhaps a more convincing example is provided by strong tea and strong beer.
These two coordinate happily without zeugma:
(89) John likes his tea and his beer to be very strong.
Yet (90) is much more normal than (91), the reason being that the strength of
beer is commonly measured in terms of the percentage of alcohol it contains, but
there are as yet no units for measuring the strength of tea:
(90) What is the strength of this beer?
(91) ?What is the strength of this tea?
It would be difficult to argue that the properties referred to in (90) and (91) are
different. The situation is similar in the case of hard. There is a scientific scale
for expressing the hardness of minerals, so the question What is the hardness of
quartz? is perfectly well formed in a scientific context. However, although some
types of wood are harder than others, the question What is the hardness of this
192 Cognitive approaches to lexical semantics
wood? is, outside of a specially contrived context, odd. Many other examples could
be given.
7.5 Conclusion
The investigations outlined here have shown the complex and varied
behavior of antonyms to be remarkably non-arbitrary. In comparison with the other
sense relations studied, the role of cognitive constraints is especially prominent.
8
Metaphor
(1) A myriad of ugly, dark thoughts clung to my reason and dug in with their claws.
193
194 Cognitive approaches to lexical semantics
8.2.1 Introduction
Cognitive linguists reject the so-called substitution theory of metaphor,
according to which a metaphorical expression replaces some literal expression that
has the same meaning. Metaphors (‘true’ metaphors), in general, are not literally
paraphrasable: they have a character that no literal expression has. At the same time,
although metaphorical meaning has a special character that distinguishes it from
any literal meaning, it has the same range of basic functions as literal meaning. Of
course, many metaphorical expressions have a heavy load of expressive meaning,
but so do many literal expressions. In other words, metaphorical meaning is not,
at least in basic functional respects, a special kind of meaning: it is rather the case
that metaphor is the result of a special process for arriving at, or construing, a
meaning.
One of the most influential books to emerge from the cognitive linguistic tra-
dition is Lakoff and Johnson’s Metaphors we live by (Lakoff and Johnson 1980;
see also Lakoff and Turner 1989; Lakoff 1987, 1993). Lakoff and his colleagues
use evidence from everyday conventional linguistic expressions to infer the ex-
istence of metaphorical relations or mappings between conceptual domains (in
the sense of chapter 2) in the human mind. Lakoff’s primary goal in developing
the conceptual theory of metaphor is to uncover these metaphorical mappings be-
tween domains and how they have guided human reasoning and behavior, as can
be seen by his subsequent application of metaphor theory to literature (Lakoff and
Turner 1989), philosophy (Johnson 1987; Lakoff and Johnson 1999), mathematics
(Lakoff and Núñez 2000) and even politics (Lakoff 1996).
Because of Lakoff’s aim to uncover deeply embedded conceptual relations in
the mind, for him the ideal metaphorical expressions to analyze are not the widely
discussed type of examples in (3), but rather those in (4):
The expressions in (3) differ from those in (4) in two important respects. First,
the expressions in (3) are all of the form X is Y where X and Y are both nominal ex-
pressions, a quite uncommon type of metaphorical expression in ordinary speech.
The expressions in (4) are all common, everyday constructions in which metaphor-
ically used prepositions, verbs and other expressions (typically relational in na-
ture) combine with literal phrases (typically nominals functioning as arguments
of the metaphorical relational elements). Hence the metaphorical expressions in
(3) are grammatically and semantically quite different from those in (4). This dis-
tinction is pertinent because much psycholinguistic research on metaphor (e.g.
Gentner 1983, 1988), including research purported to test Lakoff and Johnson’s
theory (e.g. Glucksberg 2001), is based on the metaphor type illustrated in (3),
not (4).
Second, and more important, the metaphors in (3) are novel creations ([3a–b]
are both from literary works, for example) while the metaphors in (4) are con-
ventionalized linguistic expressions, another aspect of their common everyday
character. Lakoff and Turner distinguish novel metaphors from conventionalized
metaphors, calling the former ‘image metaphors’ (Lakoff and Turner 1989:99;
Lakoff 1993:229; see below). Of course, many metaphors do not become conven-
tionalized. But certain metaphors do get conventionalized, and more interesting,
the same metaphors tend to become conventionalized independently across lan-
guages. There is presumably some reason why certain metaphors are conventional-
ized again and again across languages, while others are not. Lakoff and colleagues
argue that their repeated conventionalization is due to their cognitive significance,
which in turn is grounded in human experience (hence the title Metaphors we
live by). Thus, the main focus of their theory of metaphor is of conventional
metaphors, not novel metaphors; we will return to this point at the end of this
section.
The central characteristic of Lakoff and Johnson’s theory of (conventional)
metaphor is that the metaphor is not a property of individual linguistic expressions
and their meanings, but of whole conceptual domains. In principle, any concept
from the source domain – the domain supporting the literal meaning of the ex-
pression – can be used to describe a concept in the target domain – the domain
the sentence is actually about.
For example, the literal meaning of at in (4a) is locative in nature, but it has
been metaphorically extended to apply also to time. Likewise, in in (4b) has a
196 Cognitive approaches to lexical semantics
basic locative meaning, and the use in (4) is a metaphorical extension of this:
here, a state (danger) is conceived as a container that one can be inside of or
outside of. But many other locative expressions can be used to describe time, as
in (5), and many container expressions can be used for a wide range of states, as
in (6):
What constitutes the love is a journey metaphor is not any particular word
or expression. It is the ontological mapping across conceptual domains, from the
source domain of journeys to the target domain of love. The metaphor is not just
a matter of language, but of thought and reason. The language is secondary. The
mapping is primary, in that it sanctions the use of source domain language and
inference patterns for target domain concepts. The mapping is conventional; that
is, it is a fixed part of our conceptual system, one of our conventional ways of
conceptualising love relationships. (Lakoff 1993:208)
In general, metaphors are conceptual structures, and are not merely linguistic
in nature, although, of course, they are normally realized linguistically. The cor-
respondences between domains are represented in the conceptual system, and are
fully conventionalized among members of a speech community. An open-ended
range of linguistic expressions can tap into the same conceptual structure in both
conventional and unconventional ways, and be understood immediately: a concep-
tual metaphor cannot therefore be reduced to a finite set of linguistic expressions.
What Lakoff calls ‘elaborations’ involve more specific versions of a basic metaphor
whose characteristics in the source domain carry over to the target domain. For
instance, the difference in intensity between boil and simmer in reference to a
heated liquid carries over to indicate corresponding differences in degree of anger
in to boil with anger and to simmer with anger.
Another consequence of the conceptual nature of metaphor is that certain pat-
terns of reasoning may carry over from the source domain to the target domain.
Lakoff calls these ‘metaphorical entailments’ (it is not clear how metaphorical
entailments differ from epistemic correspondences). For instance, with reference
to the metaphor She demolished his argument, an example of the argument
is war metaphor: if you destroy all your enemy’s weapons, you win the war;
198 Cognitive approaches to lexical semantics
similarly, if you demolish all your opponent’s points in an argument, you win the
argument.
We may summarize Lakoff’s conceptual theory of metaphor as follows:
(i) It is a theory of recurrently conventionalized expressions in everyday lan-
guage in which literal and metaphorical elements are intimately combined
grammatically.
(ii) The conventional metaphorical expressions are not a purely linguistic
phenomenon, but the manifestation of a conceptual mapping between
two semantic domains; hence the mapping is general and productive
(and assumed to be characteristic of the human mind).
(iii) The metaphorical mapping is asymmetrical: the expression is about a
situation in one domain (the target domain) using concepts mapped over
from another domain (the source domain).
(iv) The metaphorical mapping can be used for metaphorical reasoning about
concepts in the target domain.
The examples in (10) suggest that the metaphor should be formulated more
concisely, that is, using less schematic source and target domains, in such a
way that the metaphorical mapping is valid for the concepts in the source and
target domains. Clausner and Croft propose the more specific formulation t h e
convincingness of an argument is the structural integrity
Metaphor 199
Clausner and Croft also argue that metaphors vary in productivity. Many
metaphors, such as the convincingness of an argument is the
structural integrity of a building , appear to be completely pro-
ductive, once formulated at the appropriate level of schematicity. Other metaphors
are only partially productive, in that some expressions are acceptable and others
are not. Examples of partially productive metaphors are the revelation idioms:
(13) spill the beans, let the cat out of the bag, blow the whistle, blow the lid off, loose
lips
(14) *spill the peas, *let the cat out of the house
Lakoff (1993:222) further adds that love is a journey can be grouped with
a career is a journey under a more schematic metaphor a purposeful
life is a journey , which in turn is an instance of what he calls the event
structure metaphor (roughly, action is directed motion ), which includes
the mappings in (18):
Grady’s and Lakoff’s highly schematic analyses raise the question of which
metaphors are more basic to human understanding, the more specific or the more
schematic ones? To address this question, however, we must first address another
issue, namely what conceptual structures are mapped in the metaphor.
Lakoff proposes the Invariance Hypothesis as a constraint on metaphorical map-
ping (Lakoff 1990:54):
Lakoff calls these ‘target domain overrides’ (1993:216), and illustrates them with
give a kick and give an idea. When you give someone a kick, the person does not
‘have’ the kick afterward, and when you give someone an idea, you still ‘have’ the
idea. The target domain of transfer of energy or force does not allow that energy to
continue to exist after the transmission event, hence that metaphorical entailment
does not hold. Likewise, the target domain of knowledge does not imply that
knowledge transmitted is lost: that metaphorical entailment does not hold either.
The Invariance Hypothesis and the target domain override raise a fundamental
issue about conceptual metaphors: why do they exist in the first place? If the target
domain has image-schematic structure already, which can override the metaphor,
then why do we have metaphors? Likewise, if we can isolate image-schematic
structure, or construct highly schematic metaphors such as organization is
physical structure , is it not simply a highly schematic conceptual struc-
ture that is instantiated in both source and target domains? If so, then is it really
a metaphorical mapping, or simply an instantiation of the image-schematic con-
ceptual structure in two different cognitive domains? (This latter view has been
propounded by Glucksberg [2001] and Jackendoff and Aaron [1991:328–30].)
Lakoff and Johnson present two counterarguments against these criticisms. First,
they argue that, although target domains of metaphors are structured, they are not
202 Cognitive approaches to lexical semantics
fully so: ‘they are not clearly enough delineated in their own terms to satisfy
the purposes of our day-to-day functioning’(Lakoff and Johnson 1980:118). Thus,
target domains lack at least some (image-schematic) structure. Lakoff and Johnson
argue that the linking of otherwise independent conceptual domains by metaphor
in fact creates similarity:
. . . the ideas are food metaphor establishes similarities between ideas and
food. Both can be digested, swallowed, devoured, and warmed over, and both
can nourish you. These similarities do not exist independently of the metaphor.
The concept of swallowing food is independent of the metaphor, but the concept
of swallowing ideas arises only by virtue of the metaphor. (Lakoff and Johnson
1980:147–48)
The second counterargument that Lakoff and Johnson present for why concep-
tual metaphors exist is that there is an asymmetry between source domain and target
domain. For example, love is expressed in terms of journeys, but journeys are not
expressed in terms of love (Lakoff and Johnson 1980:108). If image-schematic
structure were simply a highly schematic concept subsuming the corresponding
concepts in the source and target domains, then one would expect metaphorical
mappings to go in either direction; but they do not. Even when it appears that
there is a bidirectional metaphorical mapping, Lakoff and Turner argue that the
two mappings are different:
We can have cases like people are machines , as in
These two counterarguments are persuasive; but they also imply that the Invariance
Hypothesis and the target domain override captures only part of the nature of
metaphorical mappings, and perhaps not the most important part. The fact that
both people and machines have parts that function (or malfunction) is part of why
human engine works as a metaphor. But Eliot is certainly conveying more than
that with the metaphor, including perhaps the mechanization of twentieth-century
Metaphor 203
life and its dehumanizing effects (not to mention further images created by the
simile on the next line).
It is likely that a far richer structure than simply compatible image-schemas is
brought into the target domain from the source domain. It also suggests that Lakoff
and Johnson’s first counterargument – the target domain lacks (image-schematic)
structure that is added by the metaphorical mapping from the source domain –
makes too sharp a distinction between target domain structure and mapped source
domain structure. It implies a minimum of interaction between the target domain
structure (already there) and the source domain structure (filling in for the ab-
sence of target domain structure). Instead, many metaphor theorists argue for a
more interactive relationship between source and target domain structure, involv-
ing something like a ‘fusion’ or ‘superimposition’ of structure from both domains
(Jackendoff and Aaron 1991:334; they also cite Black’s [1979] ‘interaction’ and
Ricoeur’s [1978] ‘reverberation’). Jackendoff and Aaron suggest that the source
domain concepts are transformed as well in being metaphorically applied to the
target domain (ibid.). It is this intuition that blending theory attempts to capture
(see §8.3.3). This interactive relationship of course strengthens the first counter-
argument: the metaphor brings much more than extra image-schematic structure
to the target domain.
The final issue we wish to raise is the relationship between the conventional
metaphors that Lakoff centers his attention on and novel metaphor creation. Lakoff
and Johnson (1980:52–53) argue that some novel metaphors are extensions of
existing conventional metaphors, such as the song lyric We’re driving in the fast
lane on the freeway of love for love is a journey (Lakoff 1993:210). Lakoff
and Johnson also allow for completely novel metaphors, using as an example l ov e
is a collaborative work of art (Lakoff and Johnson 1980: 139–43).
They argue that a novel metaphor of this type can be a systematic mapping between
two conceptual domains.
Lakoff and Turner argue that other novel metaphors are more restricted; these
are image metaphors. An example they give is my wife . . . whose waist is an
hourglass ([3b] above; Lakoff and Turner 1989:90). Lakoff and Turner argue that in
image metaphors, specific and richly specified mental images are mapped, whereas
in conventional image-schematic metaphors, ‘there is no rich imagistic detail’
(Lakoff and Turner 1989:91). They also argue that image metaphors do not involve
the mapping of rich knowledge and inferential structure of conventional image-
schematic metaphors (ibid.).
For conventional image-schematic metaphors themselves, Lakoff and Johnson
argue that they ultimately originate in human bodily and cultural experience. For
example, conscious is up/unconscious is down, exemplified by wake
up and fall asleep, are based on the fact that ‘humans and most other mammals
204 Cognitive approaches to lexical semantics
sleep lying down and stand up when they awaken’ (Lakoff and Johnson 1980:15).
A cultural example is l a b o r i s a ( m at e r i a l ) r e s o u r c e, in that it can be
measured, assigned a value, used and so on. This metaphor arises from the fact
that we use material resources for various purposes; by virtue of those purposes
the resources have value; and the use of those resources requires labor (Lakoff and
Johnson 1980: 65–66).
It seems likely that many conventional metaphors – the kind found in everyday
language – have a basis in everyday human experience. However, many novel
metaphors do not, and some conventional metaphors do not either, except in terms
of very general image schemas such as those described by Grady (1997, 1998).
Also, where Lakoff & Johnson discuss truly novel image-schematic metaphors
such as love is a collaborative work of art , and Lakoff and Turner
discuss novel image metaphors, they do not describe how a speaker comes up with
the new metaphor: they only describe what the structure of the metaphor is. But
we have suggested that even conventional metaphor involves a blending of richer
structure than just image-schematic structure between source and target domains.
It seems plausible that even conventional metaphors draw on the full richness of our
encyclopedic knowledge of our bodily and cultural experience, especially when
they are first coined. Nor does there appear to be a difference in kind between the
‘rich detail’ mapped in novel image metaphors and the ‘rich knowledge’ mapped
in conventional image-schematic metaphors. If so, then there is only a difference
in degree between conventional metaphors and novel metaphors.
gets repeated sufficiently often, its character changes. First, its meaning becomes
circumscribed relative to the freshly coined metaphor, becoming more determi-
nate; second, it begins to be laid down as an item in the mental lexicon, so that in
time, it can be retrieved in the same way as a literal expression; third, it begins a
process of semantic drift, which can weaken or obscure its metaphorical origins.
At the beginning of its life, even if it is being laid down as an item in the lexicon,
speakers are very conscious of its status as a metaphor, and they can recreate easily
the metaphorical path of its derivation. As time passes, however, the sense of the
expression’s metaphorical nature fades and eventually disappears (although it can
be brought to life by means of Lakoffian elaborations etc.). Once that happens,
the expression is no different from a literal expression, and only etymologists and
historians of the language can recreate the path of derivation. At some point along
this path of change, the expression acquires a capability to act as a literal basis for
further metaphorical extensions, which is not possible for a fresh metaphor.
How does this impinge on the Lakoff account of metaphor? Let us look at some
relevant examples. Take the metaphor mentioned by Kövecses (2002:8), social
organizations are plants . Kövecses gives the following expressions (not
in the original order) that are said to exemplify this metaphor:
Each of these metaphors is well established in the language, but they are at dif-
ferent stages in their life history. The use of prune in (22) still strongly evokes the
source domain of arboriculture, together with the therapeutic function of pruning,
that is, to remove unnecessary growth and increase vigor. This is therefore still in
its youth as a metaphor. Arguably, reap in (23) is at a later stage of assimilation into
the language: the evocation of harvesting is relatively weak; the expression reap
benefits also shows a degree of frozenness – one cannot reap anything other than
benefits – and by some definitions this would count as an idiom. (The deadness of
reaped can be felt if it is contrasted with harvested in the same context.) When we
come to (24), a normal speaker will probably not activate the PLANT domain at
all, although if the connection is pointed out they will assent to the connection. The
word branch has become the normal literal term for a local office, shop or other
premises forming part of a larger organization. In this sense branch has developed
a completely independent set of syntagmatic and paradigmatic relations that have
nothing to do with the source domain. For instance, branch office contrasts with
head office, rather than, say, trunk office; one opens a new branch (one does not
grow one) and one closes it down (one does not cut it off); there is typically a
206 Cognitive approaches to lexical semantics
He admits that this is literally untrue, and thus seems to support the anomalist
position. However, he then goes on to point out that (27) is true:
(27) Man is not a wolf.
Metaphor 207
Does this undermine the anomalist case? Not really, because although Man is not
a wolf may be true, it is nonetheless, in the sort of context in which a metaphorical
reading would be normal, pragmatically or conversationally anomalous. Black
himself effectively concedes this point. It is not obvious how a metaphor can
function if there is nothing perceptibly odd whatsoever about its literal construal
(Eco 1996 gives a concise account of this argument).
[as in the ship of state] is linked to a ship in the source space and a nation in the
target, a surgeon is linked to both a surgeon and a butcher, and so forth. Intuitively
speaking, the point of metaphors is precisely that one thing is depicted or equated
with another. (Grady et al. 1999:114)
The two input spaces have different roles: the material in the target space functions
as topic, whereas the material in the source space ‘provides a means of re-framing
the first [i.e. the material in the target space] for some conceptual or communicative
purpose . . .’ (Grady et al. 1999:117).
However, the blended space does not only contain a selection of properties
drawn from the two input domains: it also contains new conceptual material that
arises from an elaboration of the conceptual blend on the basis of encyclopedic
knowledge. As an illustration of the model, consider the following metaphor:
(29) This surgeon is a butcher.
Grady et al. point out that this expression carries a strong implication that the
surgeon is incompetent, although incompetence is a normal feature neither of
surgeons nor butchers. The contents of the four mental spaces are described in
(30) (adapted from Grady et al. 1999):
Perhaps the ‘seeing as’ account of blending is the most illuminating suggested so
far, except that it is not really explicit enough.
Stern (1999) makes the lack of context sensitivity the major plank of his critique
of Lakoff. He argues that the aspects of a source domain that are relevant to a
target domain are heavily dependent not only on the domains themselves, but on
the whole context of the utterance, even of the discourse, and that any model that
depends on fixed structures in the mind is doomed to failure. He points out that
every context structures domains that are invoked in it in a characteristic way,
in terms of what is salient and what is backgrounded, patterns of inference, and
expressive or attitudinal factors. He illustrates his point by considering possible
metaphorical interpretations of the sun (with literal reference to the center of our
solar system). He begins with the following example uttered by Romeo in Romeo
and Juliet 2.2:
the first aspect. A satisfactory account of novel metaphor will have to be more
explicit on this matter.
On the other hand, (35a) and (36a) are statements of similarity, and do not
correspond to metaphors:
There is a different but related distinction between true similes and what might
be called speculations. Both may have the same form. Take the example of He
sounds like someone with a severe cold. Said of a person at the other end of a
telephone line, this could be a speculation that they indeed are suffering from a
severe cold. On the other hand, if it is said of someone known by speaker and
hearer to be perfectly healthy, it is a simile.
(51) She was gone in a flash of red, like a vengeful queen on her way to order armies
to march in on us.
(52) Cameras were already on her like a storm of hurled spears.
(53) It’s like I’m on the wrong planet.
(54) . . . now it’s like he’s a hunted animal with no place to go?
(55) I felt as if I were inside cut crystal.
(56) It was as if a wild animal had dragged her dying body off to its lair and mauled
it.
Similarly, there are apparent metaphors, where the mapping is highly restricted –
the motivation for the transfer can be reduced to one or a small number of features.
In such cases, one can say that they are more simile-like than are prototypical
metaphors. In fact, in certain cases, there appears to be no blending, either, which
would make them almost indistinguishable from similes.
(57) . . . her breath smoking out (= came out like smoke, because it was a cold day,
and it condensed)
(58) Grass was a thick, stiff carpet (because it was frozen) (change from metaphor to
simile makes very little difference)
(59) A computer mouse (any blending?)
(60) A splinter of light glinted in the dark. (appearance only?)
There are many cases where there is a striking difference in interpretation be-
tween a metaphor and its corresponding simile:
(61) . . . containers lined up at loading docks like animals feeding from troughs.
(cf. The containers were animals feeding from troughs.)
Notice how this opens up the mapping possibilities: there is no obvious way of
building in the mapping restrictions within the format of the metaphor. In (62), the
meaning is changed completely:
(62) Broken wooden packing cases littered the beach like the debris of a disordered
mind. (J. G. Ballard: Cocaine Nights)
(cf. Broken wooden packing cases littered the beach, the debris of a disordered
mind.)
(64) Bizarre, angry thoughts flew through my mind like a thousand starlings.
Here, the phrase like a thousand starlings gives a precise picture of the sort of
flying that is to serve as the vehicle of the metaphor.
(65) She was standing there, her eyes fastened to me like steel rivets.
Here, like steel rivets specifies the quality of fastened: rivets provide a particularly
strong and rigid type of fastening.
There are many different sorts of ‘twisting’ and ‘tumbling,’ but the similes function
to narrow them down.
In the second type, the second term of the simile appears completely literal in
itself, but is only metaphorical when the comparison is made with the first term:
1 While identity of domain does seem to be a factor, we agree with Feyaerts (2000) and Riemer (2001)
that on its own it remains an unreliable criterion in the absence of independent means of delimiting
domains.
Metaphor 217
(77) She caught the minister’s ear and persuaded him to accept her plan.
(78) He stopped on the sidewalk and looked into my eyes as people flowed around us
and light from shops unevenly shoved back the night.
The context makes clear that this refers to someone who has been trying to conceal
her anger, but has lost control, and has allowed it to be perceptible. A possible
interpretation of this is that slipped out of hiding is first metonymically interpreted
as ‘become visible,’ which is then metaphorically extended to apply to someone’s
anger, that is, ‘become perceptible (not necessarily visually).’ In all these cases,
metaphor and metonymy, although both present, can be seen to make separate
contributions.
Of course it is possible that the car had been left with the brakes off, and had rolled
down an incline on its own, coming to a stop in front of the bakery. This would be a
fully literal construal of (81). But in the normal course of events, we would interpret
this as referring to a car that was being driven, and that the agent of the stopping
action was the driver. This could then be taken as an example of metonymy – the
car is used to refer indirectly to ‘the driver of the car’ (whether this would be a
case of noun transfer or verb transfer is not relevant here). But there is a third
220 Cognitive approaches to lexical semantics
The action of drawing up is not something the car can do on its own, so any
interpretation must involve the driver. But (82) undoubtedly describes the motion
of the car and the driver may not have been visible to the speaker. In this case,
perhaps because of the specificity of the description of the car, it is easier to see
this as a humanization of the car. But a pure metaphor account would not give full
credit to the role of the driver.
A more revealing explication of (82) might be that a yellow Porsche denotes
a single entity that represents a kind of fusion of car and driver (notice that this
is different from the ‘unity’ observed with facets). If this were true, it would be
neither pure metonymy nor pure metaphor, nor would it be part metonymy, part
metaphor – it would be something intermediate. Another example is (83):
(83) Britain declares war on Iraq.
Once again, we seem to have a fusion – this time of country, government, final
decision-taker, monarch (perhaps) and so on, forming a single, semi-animate agent,
by a process that is neither pure metonymy nor pure metaphor. In one sense these
examples would seem to support Radden’s contention in the sense that expressions
can be placed on a scale of metaphoricity-metonymicity. In another sense, however,
the distinction between metaphor and metonymy as processes arguably remains
intact.2
8.6 Conclusion
Cognitive approaches to
grammatical form
9
From idioms to construction grammar
9.1 Introduction
The cognitive linguistic approach to syntax goes under the name of con-
struction grammar. It is not an exaggeration to say that construction grammar
grew out of a concern to find a place for idiomatic expressions in the speaker’s
knowledge of a grammar of their language. The study of idioms led to calls for a
rethinking of syntactic representation for many years before construction grammar
emerged, and some of this work will be referred to in this chapter. At least partly
independently of construction grammar, a number of researchers have emphasized
the need to represent linguistic knowledge in a construction-like fashion. But in
cognitive linguistics, these concerns led to a grammatical framework in which all
grammatical knowledge is represented in essentially the same way. This chapter
presents the arguments for a construction grammar.
Construction grammar, like any other scientific theory, did not arise in a theoreti-
cal vacuum. Construction grammar arose as a response to the model of grammatical
knowledge proposed by the various versions of generative grammar over the period
from the 1960s to at least the 1980s, and other syntactic theories that emerged as
direct offshoots of generative grammar. (These models in turn represented exten-
sions of the organization of a traditional descriptive grammar of a language, albeit
with significant changes in terminology.)
In most theories of generative grammar, a speaker’s grammatical knowledge is
organized into components. Each component describes one dimension of the prop-
erties of a sentence. The phonological component, for example, consists of the rules
and constraints governing the sound structure of a sentence of the language. The
syntactic component consists of the rules and constraints governing the syntax –
the combinations of words – of a sentence. The semantic component consists of
rules and constraints governing the meaning of a sentence. In other words, each
component separates out the specific type of linguistic information that is contained
in a sentence: phonological, syntactic and semantic. In addition, all versions of
Chomskyan generative grammar have broken down the syntactic component fur-
ther, as levels or strata (such as ‘deep structure,’ later ‘D-structure,’ and ‘surface
225
226 Cognitive approaches to grammatical form
syntactic component
semantic component
syntactic component
semantic component
The components are intended to be highly general rules that apply to all struc-
tures of the relevant type. Thus the rules of the phonological component apply
to all word forms and all phonological phrases (for prosodic and other phrasal
From idioms to construction grammar 227
phonology); the rules of the syntactic component apply to all sentences and
sentence types; and the same applies to rules for other components.
Of course, there must be some general way to map information from one compo-
nent onto another; for instance, there must be a way to map the syntactic structure
of a sentence onto the semantic structure of the meaning conveyed by the sentence.
These rules are called linking rules, and are also intended to be highly general,
applying to all sentences of the language. One might ask at this point, why are
the linking rules just a bunch of rules that link components, while the components
define the way that grammatical knowledge is divided up in the speaker’s mind?
As we will see, that is essentially the question that construction grammar asks.
The response of the generative grammarians is that the rules inside each compo-
nent are so highly intertwined and self-contained that they represent a cohesive
structure relative to the linking rules (and if they are not so highly intertwined, the
components are broken down further into levels, modules etc.).
In sum, the final model of the organization of grammatical knowledge in the sorts
of syntactic theories prevalent from the 1960s to the 1980s will look something
like the diagram in (3):
(3)
phonological component
linking rules
lexicon
syntactic component
linking rules
semantic component
One of the crucial characteristics of this model is that there are no idiosyncratic
properties of grammatical structures larger than a single word. Phrases and sen-
tences are governed by the highly general rules of the syntactic component and
their counterparts in the semantic and phonological components, and the equally
highly general linking rules. On the other hand, words represent an arbitrary and
idiosyncratic joining of form (phonological and syntactic) and meaning. The re-
striction of arbitrariness in grammar to the lexicon is a central principle of gener-
ative grammar, reiterated in recent versions of generative grammar (e.g. Chomsky
1993:3, 4).
One of the consequences of this model is the rejection of the concept of con-
struction in the traditional grammar sense of that word. In traditional grammar,
one describes a syntactic structure such as is found in the sentence in (4a) as ‘the
passive construction’:
The logical conclusion of this process of analysis is the hypothesis that all
properties of syntactic constructions – that is, a grammatical structure larger than
just a single word – can be captured with the general rules of the grammatical
components and their interfaces, and thus there is no need for constructions in
grammatical analysis. Chomsky makes this claim explicit:
a central element in the work discussed here, as in recent work from which it
evolves, is the effort to decompose such processes as “passive”, “relativization”,
etc., into more fundamental “abstract features”. (Chomsky 1981:121)
UG provides a fixed system of principles and a finite array of finitely valued
parameters. The language-particular rules reduce to choice of values for these
parameters. The notion of grammatical construction is eliminated, and with it,
construction-particular rules. (Chomsky 1993:4)
From idioms to construction grammar 229
The effect of this device is to include phrasal syntactic information in the lexicon,
under the lexical entry for each verb. However, this way of handling the distribu-
tional patterns in (8)–(9) is consistent with the general principle that idiosyncratic
information is to be found in the lexicon.
There is another class of syntactic phenomena that poses a much greater problem
for the componential model of grammar and the principle that all grammar above
the word level can be explained by highly general rules. These are idioms. Idioms
are, by definition, grammatical units larger than a word which are idiosyncratic in
some respect. Some examples of idioms are given in (11):
(11) a. It takes one to know one.
b. pull a fast one
c. bring down the house
d. wide awake
e. sight unseen
f. all of a sudden
g. (X) blow X’s nose
h. Once upon a time . . .
It is the necessary property of idioms that Nunberg et al. identify – their conven-
tionality – which is the relevant property of idioms with respect to the componential
model of grammar. If expressions such as those listed in (11a–h) are conventional,
then they must somehow be stored as such in a speaker’s mind. If so, then idioms
are part of a speaker’s grammatical knowledge. However, at least some aspects of
an idiom cannot be predicted by the general rules of the syntactic and semantic
components and their linking rules (we will leave out the phonological component
for the time being). Hence they pose a problem for the componential model. It
is possible to make certain sorts of stipulations to handle the conventionality of
idioms (cf. Nunberg et al. 1994:507). However, a more general treatment would
be preferable to such stipulations, if such a general treatment were available.
The linguists who ended up proposing the original construction grammar (in
Fillmore, Kay and O’Connor 1988) approached the problem of idioms from the
opposite direction. Instead of treating idioms as a problematic phenomenon from
the point of view of the componential model of grammar, they analyzed the wide
variety of idioms, and their analysis became the basis for a new model of grammat-
ical organization. The remainder of this chapter will take the reader from idioms
to construction grammar.1
Idioms can be characterized in many different ways. The description and clas-
sification that we will begin with is drawn from Fillmore et al. 1988, who used
their analysis to argue for a construction grammar. Fillmore et al. begin with three
features that can be used to classify idioms. The first feature they describe, drawn
from Makkai 1972, is the distinction between encoding and decoding idioms.
An encoding idiom is one that is interpretable by the standard rules for in-
terpreting sentences, but is arbitrary (i.e. conventional) for this expression with
this meaning. Examples given by Fillmore et al. are answer the door, wide awake
and bright red. These are all expressions that a hearer could figure out upon hear-
ing them. However, a speaker would not have guessed these expressions are the
natural-sounding English way to describe ‘open the door in response to some-
one knocking,’ ‘completely awake’ and ‘intense color.’ Another way of looking at
1 Fillmore et al. were not the first linguists to analyze idioms in a systematic way. There is a vast
literature on idioms, particularly in Europe where the study of idioms is called ‘phraseology.’ Another
important line of research on idiomatic expressions is the Firthian research on collocations. Nor were
Fillmore et al. the first to perceive the problem of idioms for a componential model of grammar, and
to propose an alternative; one such antecedent is Becker (1975). However, as in the case of many
scientific ideas, variants of the idea are proposed independently but commonly only one variant is
propagated through the scientific community (Hull 1988).
232 Cognitive approaches to grammatical form
encoding idioms is from the point of view of someone learning a foreign language.
For instance, an English learner of Spanish would not be able to give the correct
way of asking ‘how old are you?’, if s/he did not know it already; but if s/he heard
a Spanish speaker say Cuantos años tiene? (lit. ‘how many years do you have?’),
s/he would very likely figure out what the speaker meant.
A decoding idiom is one that cannot be decoded by the hearer: a hearer will
not be able to figure out the meaning of the whole at all from the meaning of
its parts. Fillmore et al. give the examples of kick the bucket and pull a fast one.
Any decoding idiom is also an encoding idiom: if as a hearer you cannot figure
out what it means, then you are also not going to be able to guess that it is a
conventional way to express that meaning in the language. One of the reasons that
a decoding idiom is a decoding idiom is because there are not any correspondences
between the literal and idiomatic meaning of the parts of the decoding idiom. For
example, kick the bucket is a transitive verb phrase, but its idiomatic meaning is
the intransitive ‘die,’ and there is nothing that corresponds even metaphorically to
a bucket.
The encoding/decoding idiom distinction corresponds rather closely to Nun-
berg, Sag and Wasow’s distinction between idiomatically combining expressions
and idiomatic phrases respectively (Nunberg, Sag and Wasow 1994:496–97).
Idiomatically combining expressions are idioms where parts of the idiomatic
meaning can be put in correspondence with parts of the literal meaning. For in-
stance, in answer the door, answer can be analyzed as corresponding to the action
of opening, and of course the door denotes the door. In an idiomatically combining
expression such as spill the beans, meaning to divulge information, spill can be
analyzed as corresponding to ‘divulge’ and the beans to ‘information.’ In contrast,
no such correspondences can be established for kick and the bucket in kick the
bucket (‘die’). Nunberg et al. call the latter idiomatic phrases.
The encoding/decoding distinction is not the same as the idiomatically combin-
ing expression/idiomatic phrase distinction, however. Some idioms, such as spill
the beans, are encoding idioms even though they are idiomatically combining ex-
pressions. The encoding/decoding distinction is rather vaguely defined: it refers to
how clever (or lucky) the hearer is in decoding an expression of the language. For
this reason Nunberg et al.’s distinction is preferable.
The encoding/decoding distinction, as well as the idiomatically combining ex-
pression/idiomatic phrase distinction, characterizes idioms in contrast to regular
syntactic expressions with respect to the interpretation rules linking the syntac-
tic component to the semantic component. With idiomatic phrases such as kick
the bucket, the interpretation rules cannot apply because the parts of the syntactic
phrase do not correspond to parts of the semantic phrase at all. With idiomatically
combining expressions such as spill the beans, the parts of the syntactic phrase
From idioms to construction grammar 233
(13) by and large; No can do; trip the light fantastic; kingdom come; battle royal;
Handsome is as handsome does; Would that it were . . .; every which way; Easy
does it; be that as it may; Believe you me; in short; happy go lucky; make believe;
do away with; make certain
There is also a substantive idiom that fits the pattern of the The X-er, the Y-er
schematic idiom:
The existence of the schematic idiom The X-er, the Y-er does not preclude the
existence of a substantive idiom like (15), just as the existence of a general syntactic
rule where the direct object follows the verb does not preclude the existence of an
idiom such as kick the bucket.
The substantive/schematic distinction characterizes idioms in contrast to regular
syntactic expressions on the one hand and the lexicon on the other. Both substantive
and schematic idioms have parts that are lexically completely specified, although
schematic idioms have parts that are specified in syntactic terms (that is, by a
syntactic category such as ‘noun phrase’ or ‘possessive pronoun’). In contrast,
syntactic rules make reference only to general syntactic categories such as V
(verb), NP (noun phrase) and so on, as in the phrase structure rules given in (16)
for simple active intransitive and transitive sentences:
(16) a. S → NP VP
b. VP → V
c. VP → V NP
The last distinction that Fillmore et al. give is for idioms with or without prag-
matic point (Fillmore et al. 1988:506). Idioms with pragmatic point are idioms
that, in addition to having a meaning in the usual sense of that term, also are
specifically used in certain pragmatic contexts. Obvious examples of idioms with
pragmatic point are idioms used for opening and closing conversations such as
Good morning or See you later, and for other specialized discourse contexts
such as telling a fairy tale (Once upon a time . . .; ibid.). Other idioms with
pragmatic point are those that have a certain conventional pragmatic content, as
with the schematic idiom illustrated by Him be a doctor?! On the other hand,
From idioms to construction grammar 235
many other idioms such as all of a sudden do not have any specific pragmatic
point.
The with/without pragmatic point distinction characterizes idioms with respect
to the ‘information structure’ or ‘discourse’ component that some linguists have
argued for. They demonstrate that some idioms have conventional information-
structure or discourse-contextual properties associated with them, which again
cannot be predicted from general pragmatic or discourse-functional principles.
For example, it may be a general pragmatic principle that in taking leave, one
may make reference to a future meeting; but it cannot be predicted that the
specific phrase See you later is conventionally used in English for that pur-
pose, whereas in Spanish Hasta luego (lit. ‘until later’) is used for the same
function.
Fillmore et al.’s analysis demonstrates that idioms are quite varied in their
syntactic, semantic and pragmatic properties, ranging from completely fixed ex-
pressions to more general expressions, which may be semantically more or less
opaque and may not even correspond to the general syntactic rules of the language.
The distinctions discussed above are summarized in (17):
Fillmore et al. use the features given above for a final, three-way categorization
of idioms. Their first category of idioms are unfamiliar pieces unfamiliarly ar-
ranged. The new aspect of this definition is the fact that certain words occur only
in a idiom. Examples of (substantive) idioms with unfamiliar pieces are kith and
kin ‘family and friends’ and with might and main ‘with a lot of strength.’ In other
words, such idioms are lexically irregular as well as syntactically and semantically
irregular. Unfamiliar words are by definition unfamiliarly arranged: if the words
do not exist outside the idiom, then they cannot be assigned to a syntactic category
in terms of a regular syntactic rule. Also, unfamiliar words unfamiliarly arranged
are by definition semantically irregular.
However, an idiom containing unfamiliar pieces unfamiliarly arranged does
not imply that it is an idiomatic phrase; such an idiom can be an idiomatically
combining expression. This point is made clearer by the schematic idiom of this
type given by Fillmore et al., the idiom The X-er, the Y-er illustrated in (14) above.
The unfamiliar pieces are the two occurrences of the, which are not definite articles
(in fact, they come from the Old English instrumental demonstrative þy). The
unfamiliar arrangement is the parallel syntactic structure, with a degree expression
236 Cognitive approaches to grammatical form
followed by a clause with a gap corresponding to the degree expression. In (14a) for
example, the parallel gapped structures are more . . . you practice and easier . . .
it will get ). Nevertheless, this idiom is an idiomatically combining expression,
in that the parts of the construction can be made to correspond with the parts of
its meaning (roughly, ‘the degree to which you practice determines the degree to
which it gets easy’).
Fillmore et al.’s second category of idioms are familiar pieces unfamiliarly
arranged. These idioms do not contain unique words but are extragrammatical. In
other words, such idioms are lexically regular, but syntactically and semantically
irregular. Fillmore et al. give all of a sudden and in point of fact as examples of
substantive idioms in this category. They give as an example of a schematic idiom of
this category the phrase Nth cousin (M times removed): this is a syntactically unique
construction. Again, idioms made up of familiar pieces unfamiliarly arranged may
be idiomatically combining expressions.
Fillmore et al.’s third and last category of idioms are familiar pieces familiarly
arranged. Such idioms are lexically and syntactically regular but semantically
irregular. Again, such idioms may be substantive or schematic; Fillmore et al. give
examples of both types. The substantive idioms they list are in fact not entirely
fixed expressions; they include pull X’s leg (which can have any person-denoting
noun phrase as X) and tickle the ivories ‘play the piano’ (which can be inflected
for tense/mood). Fillmore et al.’s schematic idioms are even more schematic; they
include what they call ‘fate tempting expressions’ such as Watch me (drop it, slip
etc.).
The types of idioms, and their comparison to regular syntactic expressions, are
given in Table 9.1.
some elements of the construction are lexically open on the one hand, and so the
idioms fitting the description cannot simply be listed as ‘phrasal lexical items.’ In
this respect, schematic idioms differ from substantive idioms. Fully substantive
idioms, such as It takes one to know one or The bigger they come the harder they
fall, can simply be listed as lexical items. Listing substantive idioms would require
the allowance for multiword lexical items in the lexicon. But this concession to
the linguistic facts does not conflict greatly with the principle of componential
grammar that arbitrary and idiosyncratic linguistic knowledge is found in the
lexicon (cf. the discussion of subcategorization frames in §9.2). Hence substantive
idioms do not require any drastic departure from the componential model of the
organization of grammar.
Schematic idioms, on the other hand, cannot simply be listed in the lexicon.
And schematic idioms are idioms; that is, they are semantically and possibly
also syntactically and lexically irregular. Syntactic, semantic and in some cases
pragmatic properties of schematic idioms cannot be predicted from the general
rules of the syntactic and semantic components (and the pragmatic component)
or the general rules linking these components together. Instead, the syntactic,
semantic (and in some cases pragmatic) properties must be directly associated
with the construction. Such a representation would cut across the components in
the componential model of grammatical knowledge, and hence represents a direct
challenge to that model, at least for idioms.
Fillmore et al. make the case for constructions as units of syntactic represen-
tation by examining one construction in great detail, the construction containing
the conjunction let alone, and demonstrating that it has syntactic, semantic and
pragmatic properties that cannot be described by the general rules of the language,
but is rule-governed within the context of the let alone construction and certain re-
lated constructions. The following discussion will present some of the more salient
unique properties of the let alone construction.
The syntax of the let alone construction is complex. Let alone can be described
as a coordinating conjunction; like other conjunctions, it conjoins a variety of
like constituents (Fillmore et al. 1988:514; the emphasized elements represent
prosody):
However, let alone fails in some syntactic contexts where and is fine, and vice
versa (Fillmore et al. 1988:515–16; Fillmore et al. also discuss WH-extraction and
It-clefts):
238 Cognitive approaches to grammatical form
Example (20d) shows that let alone allows sentence fragments for the second
conjunct. In this respect let alone is like certain other conjunctions, including
comparative than (Fillmore et al. 1988:517, 516):
However, unlike comparative than and ordinary conjunctions, let alone is impossi-
ble with VP ellipsis (deletion of the verb phrase excluding the auxiliary; Fillmore
et al. 1988:516):
(22) a. Max will eat shrimp more willingly than Minnie will.
b. Max won’t eat shrimp but Minnie will.
c. *Max won’t eat shrimp let alone Minnie will.
The let alone construction is a focus construction, like a number of other con-
structions of English (see Prince 1981b, discussed below), hence its characteristic
prosody. In fact, let alone is a paired focus construction, like those given in (23b–c)
(Fillmore et al. 1988:517):
The let alone construction allows for multiple paired foci (see [19a]), and in such
sentences allows multiple let alones (Fillmore et al. 1988:520):
(24) a. You couldn’t get a poor man to wash your car for two dollars, let alone a rich
man to wax your truck for one dollar.
b. You couldn’t get a poor man, let alone a rich man, to wash, let alone wax, your
car, let alone your truck, for two dollars, let alone one dollar.
In this respect, let alone is similar to the construction not P but Q (illustrated
in [25]) and to the respectively construction (illustrated in [26]); but Fillmore et
al. argue that in other respects let alone differs from both of these constructions
(1988:521–22):
(25) Ivan sent not an album but a book, and not to Anna on her anniversary but to
Boris on his birthday.
(26) Fred and Louise hated their shrimp and squid respectively.
From idioms to construction grammar 239
Let alone is a negative polarity item, not unlike any; it occurs in negative contexts
and certain other contexts (Fillmore et al. 1988:518):
However, unlike these polarity items, let alone is allowed in certain contexts where
other negative polarity items are disallowed (Fillmore et al 1988:519; example
[29a] is attested):
(29) a. You’ve got enough material there for a whole s e m e s t e r, let alone a w e e k.
b. *You’ve got enough material for any semester.
The semantics as well as the syntax of let alone is complex and not entirely pre-
dictable from more general rules of semantic interpretation from syntactic struc-
ture. As mentioned above, the let alone construction has at least one paired focus
(e.g. the pair semester and week in [29a]). The interpretation of a let alone sen-
tence requires the following steps. First the interpreter must recognize or construct
a semantic proposition in the fragmentary second conjunct that is parallel to the
proposition in the full first conjunct. Second, the interpreter must recognize or
construct a semantic scale underlying the elements in the propositions. This is not
always easy. For instance, the scale for (18a) may have to do with the assumed de-
gree of distastefulness of shrimp versus squid, or it may have to do with the relative
cost of shrimp versus squid (and Fred’s stinginess; Fillmore et al. 1988:524–25).
More specifically, the interpreter must perform the following semantic opera-
tions. The interpreter must construct a scalar model, which ranks propositions
on a scale – for example, the scale of distastefulness of eating seafood or the cost
thereof. The propositions in the two conjuncts must be from the same scalar model –
in this case, ‘Fred not eat shrimp’ and ‘Fred not eat squid.’ The two propositions
are of the same polarity (in this case, negative). Finally, the initial, full conjunct
denotes the proposition that is stronger or more informative on the scale – Fred not
eating shrimp is more informative than Fred not eating squid, on the assumption
that people who would eat squid would also eat shrimp but not vice versa. This
semantic analysis can be generalized to multiple paired focus versions of let alone
(see Fillmore et al. 1988 for details). This whole semantic apparatus is required for
the interpretation of the let alone construction, and is not necessary (as a whole)
for other constructions.
Finally, there is a specific pragmatic context in which the utterance of a let alone
construction is felicitous (Fillmore et al. 1988:532). First, the discourse context
is one such that the weaker (less informative) proposition, that is, the underlying
240 Cognitive approaches to grammatical form
proposition of the fragmentary second conjunct, is at issue – for example, the issue
of whether or not Fred eats squid. The weaker proposition accepts or rejects this
context – in this case, Fred doesn’t eat squid rejects it. But simply uttering the less
informative proposition is not cooperative since the speaker knows that the strong
proposition represented by the initial conjunct is true. So the speaker utters the let
alone sentence. Fillmore et al. note that let alone is similar pragmatically to other
conjunctions allowing sentence fragments, such as those illustrated in (21b–c)
above. However, some of these conjunctions present the stronger proposition in
the second, fragmentary conjunct, unlike let alone:
The preceding discussion has presented some of the evidence that the let alone
construction has its own syntactic, semantic and pragmatic properties that cannot be
predicted from more general rules of syntax, semantics and pragmatics. A number
of other studies done in the emerging framework of construction grammar demon-
strate that other constructions also have unique syntactic, semantic and pragmatic
properties. A reading of these studies gives rise to two general observations.
First, the construction on which attention is focused by the researcher(s) turns
out to be just one of a family of related constructions. For example, the let alone
construction turns out to be just one of a family of coordinate constructions that
allow certain kinds of sentence fragments in the second conjunct, two of which
were illustrated in (21b–c). The let alone construction also turns out to be just
one of a family of paired focus constructions, two of which were illustrated in
(23b–c). Paired focus constructions are in turn related to a family of single focus
constructions. The phrase let alone is itself related to other negative polarity items,
and let alone is also related to a number of items that require a scalar model for
their semantic interpretation, such as even, almost, few and merely (Fillmore et al.
1988:530).
Likewise, Lakoff’s seminal study of the There-construction, as in There’s a
fox in the garden, uncovered a large family of related constructions with slightly
different syntactic and semantic properties, which are illustrated in examples (31)–
(32) (see Lakoff 1987, Appendix 3 for the analysis of the differences among There-
constructions):
(34) It is against pardoning these that many protest (Philadelphia Inquirer, February
6, 1977)
(35) What you are saying is that the President was involved (Haldeman, Watergate
tapes)
Prince notes that WH-clefts and It-clefts differ syntactically, in that the former
allow clefted adverbs or prepositional phrases as well as clefted noun phrases,
and the latter commonly allow verb phrases or sentences as clefted items (Prince
1978:884). Discourse-functionally, WH-clefts can be used when the information
in the subordinate clause is in the hearer’s consciousness (Prince 1978:894). In
contrast, Prince identifies at least two distinct ‘sub-senses’ for It-clefts, illustrated
in (36)–(37) (1978:896, 898):
(36) So I learned to sew books. They’re really good books. It’s just t h e c ov e r s that
are rotten. (Bookbinder in S. Terkel, Working, 1974)
(37) It was just about 50 years ago that Henry Ford gave us the weekend. (Philadelphia
Bulletin, January 3, 1976)
Example (36) illustrates what Prince calls a stressed focus It-cleft. In the stressed
focus It-cleft, the subordinate that-clause is given but not assumed to be in the
hearer’s consciousness. The stressed focus It-cleft is interesting also in that it
has a phonological property associated with it: only the focused part (in small
capitals in [36]) has strong stress; the that-clause is weakly stressed. Example
(37) is an informative-presupposition It-cleft: the that-clause presents information
that is a general known fact, albeit not to the hearer and hence new to the hearer
(and therefore also not in the hearer’s consciousness). Informative-presupposition
It-clefts have a normal rather than weak stress on the that-clause. These examples
indicate that constructions may have phonological features associated with them
as well as syntactic, semantic and pragmatic/discourse features.
Birner and Ward (1998) analyze a wide range of preposing constructions, such
as Topicalization (illustrated in [38]; Birner and Ward 1998:51), postposing con-
structions, such as right-dislocation (as in [39]; Birner and Ward 1998:146) and
argument reversal constructions, such as inversion (as in [40]; Birner and Ward
From idioms to construction grammar 243
1998:159):
(38) As members of a Gray Panthers committee, we went to Canada to learn, and learn
we did. (Philadelphia Inquirer, June 16, 1985)
(39) It’s very delicate, the lawn. You don’t want to over-water, really. (father in the
movie ‘Honey, I Shrunk the Kids’)
(40) Behind them, moving slowly and evenly, but keeping up, came Pa and Noah.
(J. Steinbeck, The Grapes of Wrath, 1939)
Birner and Ward argue that, although there are commonalities among the different
constructions with respect to the discourse status of the preposed element and
the status of the postposed elements for syntactically similar constructions, each
has its own unique discourse properties. In other words, the various preposing,
postposing and inversion structures they discuss must each be analyzed as distinct
grammatical constructions.
Wierzbicka has discussed the properties of several families of constructions in
various publications (see the papers collected in Wierzbicka 1980, 1988, as well as
the examples discussed here). For example, Wierzbicka argues that the schematic
idiom have a V and the related types give a V and take a V, illustrated in (41)–(43),
represent rule-governed constructions (Wierzbicka 1988:293, 338):
(41) a. have a drink
b. *have an eat
(42) a. give the rope a pull
b. *give the window an open
(43) a. take a look at
b. *take a look for
Wierzbicka argues that the item following the indefinite article is a verbal in-
finitive, not a noun, and hence differs from other have constructions that do take
a noun, or more generally a noun phrase. For example, the phrase have a cough
is nominal, in that one can also have a headache/have pneumonia and so on in
which the word is indubitably a noun (Wierzbicka 1988:295–96). In this respect,
the have a V construction is syntactically unique.
Semantically, Wierzbicka argues that have a V represents an action as limited
in time but not punctual, lacking an external goal, and repeatable, and is of benefit
to the agent/experiencer (1988:297–302). Wierzbicka argues that this semantic
characterization is still incomplete, since it provides necessary but not sufficient
conditions for the occurrence of verbs in this construction. Instead, she presents
ten subtypes of the have a V construction, just as Prince offers two subtypes of the
It-cleft. One of these types she describes as ‘aimless objectless action which could
cause one to feel good,’ exemplified by have a walk/swim/run/jog/lie-down and so
on. In this subclass, the verbs are intransitive but durative and atelic (Wierzbicka
244 Cognitive approaches to grammatical form
The verb cannot describe a purposeful action (one with an external goal), other
than a recreational activity, hence the unacceptability of (45b–c):
Wierzbicka (1987) points out that in fact, the literal translation of (46) in various
languages (see examples [47]–[50]) does not have the same pragmatic meaning
as (46). Instead, different constructions are used to obtain approximately the same
pragmatic force (Wierzbicka 1987:96–97). In examples (47)–(50), the (a) sentence
is the closest literal translation to (46), the (b) sentence the one with the closest
pragmatic meaning to that of (46), and the (c) sentence is the literal translation of
the (b) sentence:
(47) French:
a. *Les garçons sont les (des?) garçons.
b. ?Les garçons seront toujours les (des) garçons. [still questionable]
c. ‘Boys will always be boys’
(48) German:
a. *Knaben werden Knaben sein.
b. ?Knaben bleiben (immer) Knaben.
c. ‘Boys remain (always) boys’
(49) Russian:
a. *Mal’čiki budut mal’čiki.
b. (Čego ty xočeš’?) oni že mal’čiki.
c. ‘(What do you expect?) They are boys’
(50) Polish:
a. *Chlopcy be da chlopcy.
b. (Jednak) co Paryż to Paryż.
c. ‘(However) what (is) Paris this (is) Paris’
From idioms to construction grammar 245
These examples indicate that the pragmatic interpretation in (46) is in fact con-
ventionally associated with the equational tautology in which it occurs in English.
In fact, Wierzbicka argues for several different equational tautological con-
structions in English, which cannot be substituted for each other (Wierzbicka
1987:104):
Wierzbicka argues that the semantic interpretations for the constructions exem-
plified in (51)–(55) can be characterized as follows: a ‘sober’ attitude toward com-
plex human activities (51); tolerance for human nature (52–53), the future subtype
indicating ‘the willful and uncontrollable spontaneity’ of the human type (1987:
107); obligation with respect to a human role, activity or institution ([54]–[55];
Wierzbicka argues that [54] has other readings as well). These semantic differences
cannot be inferred either from general rules of semantic interpretation in English
or general rules of the pragmatics of communication.
Even in the generative grammatical tradition, which is the theory most closely
identified with the componential model, there have been studies of schematic id-
ioms, in particular by Jackendoff (1990, 1997; see also Akmajian 1984 and Lam-
brecht’s 1990 reanalysis in construction grammar terms). For example, Jackendoff
(1997) analyzes the ‘time’-away construction, illustrated in (56):
Syntactically, the noun-phrase after the intransitive verb acts like a direct object
complement, and normally cannot occur with a transitive verb ([57a–b]). In some
cases the ‘normal’ direct object can appear in a with phrase, which it cannot do in
an ordinary active construction ([57c–d]; Jackendoff 1997:535):
246 Cognitive approaches to grammatical form
The particle away as an aspectual particle seems to have a meaning and behavior
similar to away in the ‘time’-away construction:
But the particle away is atelic (unbounded), while the ‘time’-away construction is
telic, as indicated by the It take NPtime construction:
(63) a. *It took a month for Lois and Clark to finally get to dance away.
b. It took a month for Lois and Clark to finally get to dance two blissful hours
away.
accounts is that if one wants to ‘preserve the assumption that the lexical verb’s argu-
ment structure always determines the argument structure of the VP’ (Jackendoff
1997:557), then one must commit oneself to the lexical rule analysis. Jackend-
off himself inclines to the constructional analysis for the ‘time’-away construc-
tion, since he believes that constructions are necessary in other contexts anyway
(ibid.).
Jackendoff’s inclination in his 1997 paper is another step away from the com-
ponential model of generative grammar toward a construction grammar model.
Jackendoff’s inclination is also a step toward the construction grammarian’s bolder
hypothesis. Since one must posit constructions in order to account for a substantial
part of a speaker’s grammatical knowledge, is it possible to generalize the concept
of construction to account for all of a speaker’s grammatical knowledge? The next
section presents construction grammar’s arguments for the bolder hypothesis.
The preceding section presented a number of case studies that argue for
the need to posit constructions as a unit of syntactic representation. A construction
is a syntactic configuration, sometimes with one or more substantive items (e.g.
the words let alone, have a . . . and away) and sometimes not (as with the focus
constructions, the exclamative constructions and the resultative construction). A
construction also has its own semantic interpretation and sometimes its own prag-
matic meaning (as with the tautological constructions). Hence a construction as a
unit cuts across the componential model of grammatical knowledge. The existence
of constructions would require a revision to the componential model in (3) that we
may represent as in (66):
(66)
phonological component
constructions
linking rules
lexicon
syntactic component
linking rules
semantic component
Constructions, like the lexical items in the lexicon, are ‘vertical’ structures that
combine syntactic, semantic and even phonological information (for the specific
words in a construction, as well as any unique prosodic features that may be
associated with a construction). As more and more constructions are discovered
and analyzed, construction grammarians came to argue that, in fact, grammatical
248 Cognitive approaches to grammatical form
Other schematic idioms have one or more open argument slots as well as inflec-
tional flexibility, such as give NP the lowdown ‘tell NP the news’:
Still other schematic idioms have open classes for all ‘content’ words, leaving
just a salient form such as the connective let alone as a substantive element:
(69) a. She gave me more candy than I could carry, let alone eat.
b. Only a linguist would buy that book, let alone read it.
Finally, a constructional analysis has been proposed for some schematic id-
ioms in which all elements are lexically open, such as the resultative construction
(Goldberg 1995:181; attested examples):
(70) a. This nice man probably just wanted Mother to . . . kiss him unconscious.
(D. Shields, Dead Tongues, 1989)
b. I had brushed my hair very smooth. (C. Brontë, Jane Eyre, 1847)
Yet the resultative construction has no lexically specific element. It can be de-
scribed only by a syntactic structure, in this case [NP Verb NP XP], with a unique
specialized semantic interpretation.
It is a very short step from analyzing the resultative construction as a construc-
tion to analyzing all the syntactic rules of a language as constructions (Fillmore
From idioms to construction grammar 249
et al. 1988:501, 534; Langacker 1999:19). After all, a syntactic rule such as
VP → V NP describes a completely schematic construction [V NP], and the seman-
tic interpretation rule that maps the syntactic structure to its corresponding semantic
structure is unique to that schematic construction. Indeed, Goldberg suggests that
there is a transitive construction just as there are more specialized schematic syn-
tactic constructions such as the resultative construction (Goldberg 1995:116–19).
Reanalyzing general syntactic rules as the broadest, most schematic constructions
of a language is just the other end of the substantive- schematic continuum for
idioms/constructions.
Turning to semantic interpretation, one can also argue that constructions and
compositional semantic rules differ only in degree, not in kind. As we noted in
§9.2, Nunberg et al. (1994) argue that most idioms are idiomatically combining
expressions. In an idiomatically combining expression, the syntactic parts of the
idiom (e.g. spill and beans) can be identified with parts of the idiom’s semantic
interpretation (‘divulge’ and ‘information’). Nunberg et al. argue that idiomatically
combining expressions are not only semantically analyzable, but also semantically
compositional.
Nunberg et al. observe that idiomatically combining expressions are only the
extreme end of a continuum of conventionality in semantic composition. The other
end of the continuum is represented by selectional restrictions. Selectional re-
strictions are restrictions on possible combinations of words which are determined
only by the semantics of the concepts denoted by the word. For example, the re-
strictions on the use of mud and car in (71)–(72) follow from the fact that mud is
a viscous substance and a car is a machine:
The restrictions on mud and car are not dependent on the conventional form in
which the concepts are expressed. If one used the word goo instead of mud or
automobile instead of car, the judgements in (71)–(72) would remain the same.
The combinations in (71a) and (72a) are semantically compositional: the meaning
of the whole can be predicted from the meaning of the parts.
An intermediate point on this continuum involves what are called collocations.
Collocations are combinations of words that are preferred over other combinations
that otherwise appear to be semantically equivalent. For example, Matthews argues
that toasted and roasted describe essentially the same process, but are restricted
in their acceptable combinations (Matthews 1981:5):
250 Cognitive approaches to grammatical form
Nunberg et al. argue that exactly the same reasoning applies to idiomatically
combining expressions. Idiomatically combining expressions are largely fixed in
their words; any substitution leads to ungrammaticality, as in (76b–c) and (77b):
However, given the meanings of the words in the idiomatically combining expres-
sion, the meaning of the whole expression is compositional:
that beans can have the meaning ‘information’ without spill. The availability of
these meanings for each constituent can be dependent on the presence of another
item without requiring that the meaning ‘divulge the information’ attach directly
to the entire VP. Rather it arises through a convention that assigns particular
meaning to its parts when they occur together. (Nunberg et al. 1994:497)
At first, Nunberg et al.’s analysis may look odd. To say that pull and strings
each have a meaning found only in pull strings, and that those meanings are
compositional in the idiomatically combining expression, seems ad hoc. The more
natural description is the traditional one, that the meaning of the idiomatically
combining expression is ‘noncompositional.’ In fact, it is sometimes said that
one of the strongest pieces of evidence for constructions as independent syntactic
objects is that there is some degree of ‘noncompositionality’ in the meaning of
the construction. But there is evidence that Nunberg et al.’s analysis is the right
one.
Some English words exist only in idiomatically combining expressions, such as
heed in pay heed. It makes sense to say that heed has a meaning, that is of course
found only in pay heed. It has been argued that heed is idiomatic, because it is
essentially synonymous with attention in pay attention, and yet does not behave
the same way (Radford 1988; see Nunberg et al. 1994:505):
Nunberg et al. argue that heed does not in fact mean the same thing as attention
does, when attention is the object of pay (Nunberg et al. 1994:505):
The semantic differences are related to the difference between the verbs attend and
heed: ‘we clearly attend to much that we do not heed . . . one can take heed but not
attention, and . . . attention but not heed can wander’ (Nunberg et al. 1994:506).
In other words, heed in pay heed does have its own meaning even though it occurs
(as a noun) in only that combination. Hence, it is reasonable to assume that other
words have specialized meanings in idiomatically combining expressions, and that
those meanings are compositional.
Another important line of evidence for the compositionality of idiomatically
combining expressions is psycholinguistic. Speakers of English recognize the
meanings of words in idiomatically combining expressions, and recognize them
as figurative meanings, even though the figurative meanings are found only in the
idiomatically combining expressions (Gibbs 1990). These two pieces of evidence
252 Cognitive approaches to grammatical form
point to Nunberg et al.’s conclusion that ‘The dependency among the parts of
idiomatically combining expressions is thus fundamentally semantic in nature’
(Nunberg et al. 1994:505).
From a construction grammar perspective, Nunberg et al.’s analysis of idiomat-
ically combining expressions looks more natural. An idiomatically combining ex-
pression such as spill the beans is a construction. As a construction, it has unique
syntax: the verb must be spill and its object must be the beans. It also has a se-
mantic interpretation, namely ‘divulge information.’ All Nunberg et al. are saying
is that this construction has its own semantic interpretation rules, mapping spill
onto ‘divulge’ and the beans onto ‘information.’ The constructional analysis is
presented in the diagram in (80), using lowercase to describe form and uppercase
to describe meaning, boxes to represent the construction and its parts, and dotted
lines to indicate the syntax-semantics mapping (see §10.1):
(80)
spill the beans
DIVULGE INFORMATION
The English predicate adjective construction has the form [NP be Adj]. It differs
from the ordinary verbal construction in requiring the copula verb be. One can ana-
lyze the semantics of the predicate adjective construction as follows. The members
of the Adjective category have a meaning that requires them to be combined with
the copula be in order to be interpreted as ascribing a property to a referent (unlike
verbs). The copula be has a meaning that requires combination with a member
of the Adjective category in order to be interpreted as doing the job of ascribing
(a property) to the subject NP. This analysis is in fact essentially the semantic
analysis that Langacker argues for (Langacker 1987:214–22; 1991a:204–5). In
Langacker’s terminology (see §3.5), Adjective symbolizes an atemporal relation,
and the copula be symbolizes a process that Adjective meanings must be combined
with in order to be predicated.
In like fashion, semantic interpretation rules can be provided for any schematic
construction describing the most general syntactic structures of the language. In
other words, all syntactic expressions, whatever their degree of schematicity, have
rules of semantic interpretation associated with them, although some substantive
idioms appear to inherit their semantic interpretation rules from more schematic
syntactic expressions such as [Verb Object] (see §10.2.1). Hence, the difference
between regular syntactic expressions and idiomatically combining expressions
is not that the former are ‘compositional’ and the latter are ‘noncompositional.’
Instead, the former’s rules of semantic composition are more general and the
254 Cognitive approaches to grammatical form
2 We ignore here the fact that the plural stem for the plural in -en is distinct from the singular stem.
From idioms to construction grammar 255
these observations suggest that in fact morphology is very much like syntax, and
that a constructional representation is motivated for morphology as well.
Lastly, the lexicon differs only in degree from constructions. Words in the lexi-
con are pairings of syntactic form (and phonological form) and meaning, includ-
ing pragmatic meaning. Constructions are also pairings of syntactic form (and
phonological form, for the substantive elements) and meaning, including prag-
matic meaning. The only difference is that constructions are complex, made up of
words and phrases, while words are syntactically simple. Some words are morpho-
logically complex, of course. But we have just argued that construction grammar
would analyze morphologically complex words as constructions whose parts are
morphologically bound. Morphologically simple words are atomic, that is, they
cannot be further divided into meaningful parts. But a word is again just the limiting
case of a construction (Fillmore et al. 1988:501).
The end point of this argument is one of the fundamental hypotheses of construc-
tion grammar: there is a uniform representation of all grammatical knowledge
in the speaker’s mind, in the form of generalized constructions. Table 9.2 com-
pares the different types of grammatical entities found in the componential model
of grammar and their analysis as constructions in construction grammar.
Table 9.2 The syntax-lexicon continuum
construction3
construction2
[etcetera]
lexicon
syntactic properties
semantic properties
257
258 Cognitive approaches to grammatical form
CONSTRUCTION
syntactic properties
linking rules
Figure 10.2 The relation between form and function in a componential syntactic
theory
An overview of construction grammars 259
word NP VP Quant N
grammar
Pred Pred(arg) Qx: Pred(x)
constructions
Figure 10.3 The relation between form and function in construction grammar
In the componential model, the various syntactic structures are organized inde-
pendently of the corresponding semantic structures, as represented by the high-
lighted boxes in Figure 10.2. In construction grammar, the basic linguistic units are
symbolic, and are organized as symbolic units, as represented by the highlighted
boxes in Figure 10.3.1 As a consequence, the internal structure of the basic (sym-
bolic) units in construction grammar is more complex than that of basic units in
the componential model.
The internal structure of a construction is the morphosyntactic structure of sen-
tences that instantiate constructions. For example, a simple intransitive sentence
such as Heather sings is an instance of the Intransitive construction. If we compare
a simplified representation of Heather sings in generative grammar to a simpli-
fied representation of the same in construction grammar, we can see that they
are actually rather similar except that the construction grammar representation is
symbolic.
The box notation used in Figure 10.4b (on page 260) is simply a notational
variant of the bracket notation used in Figure 10.4a (Langacker 1987; Kay and
Fillmore 1999). Thus, we can see that both the generative grammar representation
and the construction grammar representation share the fundamental part-whole or
meronomic structure of grammatical units: the sentence Heather sings is made
up of two parts, the Subject Heather and the Predicate sings.
The brackets in Figure 10.4a are labeled with syntactic category labels, while the
corresponding boxes in the syntactic structure of Figure 10.4b are not labeled. This
does not mean that the boxed structures in Figure 10.4b are all of the same syntactic
type. Construction grammarians assume, of course, that syntactic units belong to
1 Other theories that share construction grammar’s basis in symbolic units are Head-driven Phrase
Structure Grammar (HPSG; Pollard and Sag 1987, 1993), and Semiotic Grammar (McGregor 1997).
However, these theories are not explicitly construction based, although HPSG and Fillmore and Kay’s
version of construction grammar have converged in many respects.
260 Cognitive approaches to grammatical form
construction
Figure 10.4 Simplified generative grammar and construction grammar
representations of Heather sings
a variety of different syntactic categories. The boxes have been left unlabeled
because the nature of those categories is one issue on which different theories
of construction grammar diverge. That is, we may ask the following question of
different construction grammar theories:
(i) What is the status of the categories of the syntactic elements in construction
grammar, given the existence of constructions?
for clarity’s sake; but all three types of parts of constructions are illustrated in
Figure 10.5 (compare Langacker 1987:84, Fig. 2.8a; Figure 10.5 suppresses links
between parts of the construction for clarity).
(syntactic)
Heather sings element
(semantic)
HEATHER SING component
(symbolic) unit
Figure 10.5 Elements, components and units of a construction
Figure 10.4b has two other relations apart from the symbolic relation: one join-
ing the two syntactic elements and one joining the two semantic components. The
link joining the two semantic components describes a semantic relation that holds
between the two components, in this case some sort of event-participant relation.
Thus, the semantic structure of a construction is assumed to be (potentially) com-
plex, made up of semantic components among which certain semantic relations
hold.
The link joining the two syntactic elements in Figure 10.4b is a syntactic rela-
tion. The syntactic relation does not obviously correspond directly to anything in
the generative grammar representation in Figure 10.4a. This is because the repre-
sentation of syntactic relations in most syntactic theories is more complex than a
simple syntactic link. One layer is the syntactic relation itself, such as the subject-
verb relation holding between Heather and sings in the construction grammar
representation in Figure 10.4b. A second layer is the means of representing syn-
tactic relations. Different syntactic theories use different means for representing
abstract syntactic relations. For example, generative grammar uses constituency
to represent syntactic relations, while Word Grammar (Hudson 1984) uses depen-
dency. The third layer is the overt manifestation of syntactic relations, such as
word order, case marking and indexation (agreement). We strip away the latter
two layers in comparing construction grammar theories.
An important theoretical distinction is made regarding the internal structure
of constructions (Kay 1997). The analysis of syntactic structure is unfortunately
confounded by an ambiguity in much traditional syntactic terminology. We can
illustrate this with the example of the term ‘subject’ in the Intransitive Clause
construction in Figure 10.5, illustrated once again by the sentence Heather sings.
The term ‘subject’ can mean one of two things. It can describe the role of a
262 Cognitive approaches to grammatical form
Subject Predicate
The subject role defines a grammatical category. But the term ‘subject’ can
also describe a syntactic relation between one element of the construction – the
subject – and another element of the construction – the Verb. This is the sense in
which one says that Heather is the ‘subject of the Verb’ sings. In other words, the
term ‘subject’ confounds two different types of relations in a construction: the role
of the part in the whole, and the relation of one part to another part. The difference
between the two is illustrated in (2):
(2)
Intransitive Construction
role
idiom The X-er, the Y-er, and also of a substantive idiom The bigger they come,
the harder they fall. We noted (following Fillmore et al. 1988) that the existence
of the schematic idiom was not incompatible with the existence of the substantive
idiom. A construction grammarian captures the fact that the substantive idiom is an
instance of the schematic idiom by representing this relationship with a taxonomic
link, as in (3):
SBJ sleep SBJ run SBJ kick OBJ SBJ kiss OBJ
I didn't sleep
The sentence I didn’t sleep thus has multiple parents in the taxonomy of con-
structions to which it belongs. This is a consequence of each construction being a
partial specification of the grammatical structure of its daughter construction(s).
For example, the Negative construction only specifies the structure associated with
the subject, verb and auxiliary; it does not specify anything about a verb’s object
(if it has one), and so there is no representation of the object in the Negative
construction in (6).
A construction typically provides only a partial specification of the structure of
an utterance. For example, the Ditransitive construction [Sbj DitrVerb Obj1
Obj2], as in He gave her a book, only specifies the predicate and its arguments.
It does not specify the order of elements, which can be different in different
constructions: compare the Simple Declarative example just given above with
An overview of construction grammars 265
the It-Cleft construction It was a book that he gave her. Nor does the Ditransitive
construction specify the presence or position of other elements in an utterance, such
as modal auxiliaries or negation, whether in a declarative sentence (where they are
preverbal; see [7a]) or an interrogative sentence (where the auxiliary precedes the
subject; see [7b]):
(7) a. He won’t give her the book.
b. Wouldn’t he give her the book?
Finally, the taxonomic hierarchy appears to represent the same or similar infor-
mation at different levels of schematicity in the hierarchy. For example, the fact
that the habit is the direct object of kick in kick the habit is, or could be, repre-
sented in the idiom construction itself [kick the habit], or at any one or more of the
schematic levels above the hierarchy, all the way up to [TrVe r b O b j]. Different
theories of construction grammar have offered different answers to the question of
how information is to be represented in the taxonomic hierarchy of constructions:
(iv) How is information stored in the construction taxonomy?
theory. The different answers to the four questions bring out some current issues of
debate in construction grammar. It should be noted that the different theories tend
to focus on different issues, representing their distinctive positions vis-à-vis the
other theories. For example, Construction Grammar explores syntactic relations
and inheritance in detail; the Lakoff/Goldberg model focuses more on catego-
rization relations between constructions; Cognitive Grammar focuses on semantic
categories and relations; and Radical Construction Grammar focuses on syntactic
categories and typological universals. Finally, the last three theories all endorse
the usage-based model, which is described in chapter 11.
(8) [cat v]
[role head]
[lex +]
[role filler]
[loc +] +
[gf ¬subj]
An overview of construction grammars 267
(9)
cat v
role filler
role head loc + +
lex +
gf subj
The equivalent diagrams in (8)–(9) are read as follows. The two inner boxes
(feature structures) indicate the features of the verb and its complements (if any).
The first box specifies that the first constituent of the VP construction is its head
and that it must be lexical. For example, in found her bracelet the first constituent
is the head of the VP and it is a word, not a larger constituent. The feature-value
pair [cat v] above it is actually a simplification of a more complex feature structure
(Kay and Fillmore 1999:9, n. 13), which specifies that the syntactic category of
the head of the VP, in this case found, must be ‘verb.’ The second box specifies the
complements, if any, of the verb. The + sign following the second box (‘Kleene
plus’) indicates that there may be zero, one or more complements in the VP. In the
VP found her bracelet, her bracelet is the one and only complement. In the VP
construction, the complements are given the role value ‘filler’ (see below). The
feature [loc(al) +] indicates that the complement is not extracted out of the VP.
An example of an extracted, [loc −], complement of find would be the question
word What in the question What did he find?
The internal structure of a construction in Construction Grammar can be most
easily understood by working from the parts to the whole. Minimal parts are words
(or more precisely, morphemes; we will ignore this distinction for now). Each
part has syntactic features, grouped under the feature syn, and semantic features,
grouped under the feature sem. Construction Grammar separates the phonological
features under a feature phon if the construction is substantive. The syn and
sem features are themselves grouped under the feature ss (formerly synsem),
which represents the symbolic structure of that part of the construction. The basic
symbolic structure for Construction Grammar is given in (10):
(10)
syn ...
ss
sem ...
...
phon <...>
268 Cognitive approaches to grammatical form
(11)
Intransitive Construction
'Heather sings'
The role feature is used to represent the role of the syntactic element in the
whole. The role feature is associated with each part of a complex construction
and defines syntactic roles such as mod(ifier), filler and head. For instance, the
subject-predicate (Intransitive) construction in (11), Heather sings, has the roles
head for sings and filler for Heather (Kay and Fillmore 1999:13).
An overview of construction grammars 269
2 The version of the feature geometry given in Kay and Fillmore 1999:9, n. 10 includes another syntactic
feature under rel, namely case.
270 Cognitive approaches to grammatical form
wh + Inherit SAI
272 Cognitive approaches to grammatical form
Goldberg treats the first sense (the one in [17]) as the central, prototypical sense
and the other senses as extensions from the prototype. The extensions from the
prototype inherit the syntactic construction schema from the prototype.
Goldberg does not explicitly argue for a schema subsuming all of the senses
of the ditransitive construction. However, inheritance is a characteristic of taxo-
nomic links, so Goldberg’s analysis suggests that there is a schematic syntactic
ditransitive construction, even if there is not a semantic schema (Goldberg does not
propose a schematic meaning of the ditransitive). In other words, there is a syntac-
tic construction schema that has as instantiations the six senses in (17)–(22), and
the actual successful causation meaning in (17) is the prototype for the ditransitive
construction.
In fact, however, the syntactic schemas for the six subsenses are all slightly
different, because each sense has a distinct subclass of verbs associated with it
(Croft 2003a). Thus, each subsense has associated with it a syntactic schema
specifying the verbs or verb classes that each subsense applies to. In general,
semantic differences in the grammatical constructions such as the ditransitive or
the perfect are likely to have syntactic consequences, and so distinct constructional
senses are likely to also have distinct syntactic schemas.
It is still possible to posit a superordinate ditransitive construction specifying
what is common to all of the subordinate constructions. Such a construction would
have a syntactic schema such as [Sbj DitrV Obj1 Obj2]. The D i t r V category
in the ditransitive construction would be a polysemous category, that is, there are
no necessary and sufficient conditions to describe all and only the ditransitive
verbs in the language. But as we saw in chapter 3, this is true of many linguistic
categories, and so it is not surprising that a syntactic category turns out to be the
same. Likewise, the semantics of the superordinate ditransitive construction would
also be polysemous: although there would be a necessary condition that some
modulated transfer of possession is involved (actual, intended, future etc.), this
condition is not a sufficient condition defining the ditransitive construction’s use.
An overview of construction grammars 275
The most important property of the polysemy analysis is that one construction
sense is central and another is an extension from it. A clear case of extension
from a central sense in constructions is a metaphorical extension, another type
of link proposed by Goldberg, following Lakoff (1987) in his analysis of There-
constructions.
Lakoff argues that many of the extensions of the central There-construction
involve metaphorical extension. For example, the Perceptual Deictic There-
construction, illustrated in (23), involves a number of metaphorical extensions
from the Central Deictic There-construction illustrated in (24) (Lakoff 1987:511,
509):
(23) a. Here comes the beep.
b. There’s the beep.
(24) There’s Harry.
The Perceptual Deictic describes the impending (23a) or just-realized (23b) acti-
vation of a nonvisual perceptual stimulus, for example an alarm clock that is about
to go off. To express this meaning, the Presentational Deictic uses the metaphor of
deictic motion of a physical entity in physical space. The extension of the Central
Deictic to the Perceptual Deictic requires the following metaphorical mappings
(Lakoff 1987:511):
(25) Perceptual Deictic domain Central Deictic domain
nonvisual perceptual space is physical space
percepts are entities
realized is distal
soon-to-be-realized is proximal
activation is motion
A metaphorical extension (or any other semantic extension, for that matter)
need not establish a schema of which the basic construction and the metaphorical
extension are both instantiations. Lakoff’s based-on link, like Goldberg’s polysemy
link, involves inheritance of both syntactic and semantic properties, and so is not
unlike a taxonomic link. Lakoff, however, does not posit a superordinate Deictic
There-construction schema. On the other hand, Goldberg argues that there is a
superordinate schema subsuming both a central construction and its metaphorical
extension (Goldberg 1995:81–89).
category. For example, we know that most birds fly, to the point that if we hear
reference to ‘a bird,’ we will assume that it can fly. Of course, if we are fur-
ther informed that the bird in question is an ostrich or a penguin, or that it has
a broken wing or it is dead, we would cancel that assumption. One model for
representing this information is to store the information FLIES with the category
BIRD, instead of with the many instances of bird species and individual birds
that can fly. The property FLIES is inherited in those cases, but inheritance can
be blocked if it conflicts with information in the more specific case, such as pen-
guins, ostriches, a bird with a broken wing, a dead bird and so on. This is normal
inheritance.3
Lakoff uses normal inheritance in his analysis of There-constructions. Nor-
mal inheritance is part of Lakoff’s based-on link between constructions (Lakoff
1987:508); so does Goldberg (1995:74). For example, Lakoff argues that the
Presentational Deictic There-construction in (26) is based on the Central Deictic
There-construction in (27) (Lakoff 1987:520, 482):
One of the properties of the Central Deictic is that the verb must occur in
the simple present tense, because the semantics of the Central Deictic is to point
out a referent in the speech act situation (Lakoff 1987:490–91). The Presentational
Deictic is based on the Central Deictic but also specifies that the verb may appear in
a variety of tenses as expressed in auxiliaries (Lakoff 1987:521). This specification
blocks the inheritance of the simple present tense requirement from the Central
Deictic.
Goldberg also allows for the representation of information at all levels in the
taxonomic hierarchy of constructions. Goldberg describes such a model as a full-
entry model (Goldberg 1995:73–74). She gives an example of a situation that
virtually requires a full-entry representation, namely a conflict in multiple inher-
itance. If there are multiple parents, then there will be inheritance from multiple
‘parents’ in the taxonomic network. It may be that the multiple parent nodes have
conflicting specifications of some properties, and this conflict has to be resolved
for the specific instance. Normal inheritance cannot handle this problem. Normal
inheritance adjudicates a conflict in specification between parent and child nodes
in the taxonomy (the child always wins). In multiple inheritance the conflict is
between the two parent nodes, and there is no principled way to choose which
parent would win in a conflict.
3 Construction Grammar eschews default inheritance; instead, default values are left unspecified and
default constructions fill in unspecified values (Fillmore 1999:115, n. 3; Kay 2002:470).
An overview of construction grammars 277
Goldberg notes (following Bolinger 1971) that some resultatives allow for word
order variation of the same type as is found in the verb-particle construction:
Goldberg proposes that the class of resultatives illustrated with break open in
(30) are instances of the verb-particle construction as well as of the resultative
construction. However, the two parent constructions have conflicting properties.
While the verb-particle construction allows word order variation (compare [29]),
the resultative construction does not:
In this case, break open inherits the word order variation of the verb-particle con-
struction, not the fixed word order of the resultative. Conversely, while the resulta-
tive allows for a simple predication (compare [28b]), the verb-particle construction
does not:
In this case, break open inherits the predicability of resultatives, not the ungram-
matical predication of the verb-particle construction.
The two parent constructions of (30) give conflicting specifications as to whether
the word order change is acceptable or not, and whether the predication of the result
phrase is acceptable or not. Goldberg suggests that, in this case, the information
about the specific construction types is provided in the specific construction, even if
it is redundant with the information contained in (one of) the parent constructions;
then the problem of how to resolve the conflict of multiple inheritance does not
arise. In other words, Goldberg argues that a full-entry model in this situation is
desirable.
Is full entry plausible when the information could be represented nonredun-
dantly by inheritance? It might appear to the reader that, a priori, the inheri-
tance model is to be preferred over the full-entry model for reasons of parsimony.
278 Cognitive approaches to grammatical form
However, most cognitive linguists argue that a cognitively based grammar should
not be constructed in an a priori fashion, because grammatical knowledge is a psy-
chological phenomenon. Clearly, speakers do not store a representation of every
utterance they have ever used or heard. Speakers form schemas that generalize
over categories of utterances heard and used. But it does not necessarily follow
from this observation that speakers store every piece of grammatical knowledge
only once. It does not even necessarily follow that actual speakers form a more
schematic category for every linguistic generalization that clever linguists have
found (see Croft 1998c and §11.2.5).
The principle that information should not be stored redundantly is motivated by
the desire for parsimony in representation. But parsimony in representation simply
pushes complexity to the processes of language use. A complete inheritance model
maximizes storage parsimony, that is, it minimizes the redundant storage of infor-
mation. A complete inheritance model thus requires maximum on-line processing
in order to access and use the information in the production and comprehension of
utterances (see Goldberg 1995:74; Barsalou 1992b:180–81). A full-entry model
maximizes computing parsimony: as much information as possible is stored in
multiple places, so that on-line computation is minimized during production and
comprehension (Barsalou 1992b:180–81; see §§12.1, 12.2.5).
On the whole, the psychological evidence suggests that ‘concepts and properties
in human knowledge are organized with little concern for elegance and [storage]
parsimony’ (Barsalou 1992:180). This does not mean that full entry is to be pre-
ferred in all situations, however: such a model is just as a priori as the inheritance
model. Instead, Goldberg, following Langacker (1987, chapter 10) and other cog-
nitive linguists, advocates a usage-based model, in which patterns of language use
are taken as evidence for the independent representation of grammatical informa-
tion (see especially Goldberg 1995:133–39). In chapter 11, we will examine some
suggested criteria for positing schematic constructions and the degree to which
information is stored redundantly in the mind.
4 Langacker also allows for independent phonological and semantic units, but not independent syntactic
units.
5 Cognitive Grammar and Construction Grammar, like Head-driven Phrase Structure Grammar (Pollard
and Sag 1993), eschews the use of phonologically ‘null’ or ‘empty’ elements (see also Kay 2002).
Construction Grammar replaces the concept of a null element with the concept of null instantiation,
that is, some constructions have a feature that indicates that there is a (semantic) argument that is
not formally instantiated (Fillmore 1986b; Fillmore and Kay 1993, chapter 7). Fillmore and Kay
distinguish three types of null instantiation: definite (equivalent to null anaphora), indefinite (as in
The dog ate) and free (corresponding to unspecified adjuncts). Fillmore and Kay argue that null
instantiation is associated with either constructions or individual words; Croft (2001:275–80) argues
that instantiation is associated with constructions only.
280 Cognitive approaches to grammatical form
have an essentially semantic basis, but in terms of the construal of experience, not
in terms of semantic classes. As described in chapter 3, Langacker has developed
semantic construal analyses of a wide range of syntactic categories, including
parts of speech, grammatical roles (subject and object), the count/mass distinction,
various English tense/aspect inflections and auxiliaries, the English possessives -’s
and of, ergativity, English complementizers and complement types, Cora locatives
and the Yuman auxiliary (see Langacker 1987, 1991a, b, 1999).
One question that can be raised about the Cognitive Grammar analysis of gram-
matical categories is the relationship between the abstract semantic construal def-
initions and the variation in both formal distribution and semantic polysemy of
such categories across languages. It has been suggested that cross-linguistic vari-
ation in putatively universal semantic categories can be accommodated in terms
of conventionalized construal: the same semantic category is found everywhere,
but the construal of specific experiences as belonging to the semantic category is
language-specific:
For example, the English root sick is construed as an adjective or atemporal relation,
that is, summarily scanned (see §3.2), and requires a copula verb be for predica-
tion/sequential scanning; but the equivalent Russian root bol(e)- is construed as a
verb (sequentially scanned) and requires an adjectival derivational suffix to be con-
strued atemporally. But it is not clear that there is any difference between positing
a universal semantic category plus language-specific conventionalized construal
for specific cases on the one hand, and positing a polysemous category with a
semantic prototype and language-specific semantic extensions on the other.
(33)
HEATHER e-site SINGS
elaboration
elaborates a not very salient substructure of read. The relative strength of the two
relations is illustrated in (35):
(35)
FIGURE READER
TRAIN TEXT
ON LOCATION
dependent
READ
autonomous
Langacker use the terms autonomous and dependent to describe the gradient
reinterpretation of the predicate-argument distinction: ‘one structure, D, is depen-
dent on the other, A, to the extent that A constitutes an elaboration of a salient
substructure within D’ (1987:300). Conversely, A is autonomous relative to D to
the extent to which it does not elaborate a salient substructure of D. In (35), on the
train is dependent on read because read elaborates the highly salient figure role
of the locative relation on the train. Conversely, read is autonomous relative to on
the train because on the train elaborates only the not very salient substructure of
the location of the reading event.
The Cognitive Grammar analysis of concepts of ‘head,’ ‘modifier’ and so on are
both similar and different from the analysis in Construction Grammar. In Construc-
tion Grammar, the roles represent a relation between the parts of a construction
and the whole, and are defined syntactically. In Cognitive Grammar, the analo-
gous concepts also represent a relation between the parts of a construction and the
whole, but they are defined semantically and symbolically.
Cognitive Grammar defines a semantic relation between part and whole as the
profile determinant: the profile determinant is the part of the construction whose
semantic profile the whole construction ‘inherits’ (Langacker 1987:289). The pro-
file is the concept designated by the unit, against the background knowledge pre-
supposed by that concept (see chapter 2). Langacker combines the concepts of
profile determinacy and autonomy/dependence to define ‘head,’ ‘complement’
and ‘modifier’ in the intuitively expected way (1987:309). A head is a dependent
predication that is a profile determinant; a complement is an autonomous predi-
cation that is not a profile determinant; and a modifier is a dependent predication
that is not a profile determinant.
(i) What is the status of the categories of the syntactic elements, given
the existence of constructions?
The standard analysis of meronomic relations between syntactic struc-
tures is reductionist (§11.2.1): a construction such as the intransitive or transitive
construction is made up of parts, and those parts are themselves defined inde-
pendently of the constructions in which they occur. For example, various clausal
constructions have verbs, which are analyzed as belonging to the same part of
speech no matter what construction they occur in. This analysis is motivated in
part because they have the same inflections (present in third person singular -s and
non-third person singular zero, past in -ed or other allomorphs):
In other words, the same units occur as the parts of many different constructions.
Ultimately, the decomposition of a construction will lead to a set of basic or
primitive elements that cannot be analyzed further, and out of which constructions
are built. These atomic elements include syntactic categories such as Verb or Noun
and relations such as Subject or Object and so on. A model of grammatical structure
of this type is a reductionist model: more complex structures are treated as built up
out of primitive and ultimately atomic units. In the example given here, the atomic
units are the basic categories and relations.
The reductionist model has a significant shortcoming: it does not capture certain
empirical facts about the distribution of words. For example, while many English
verbs occur in either the transitive or intransitive constructions, many others do
not:
One solution is to divide Verbs into Transitive Verbs and Intransitive Verbs.
If so, then a decision must be made about verbs such as dance, which occur
in both constructions: do they simultaneously belong to both subclasses? or do
they form a third distinct class? One effect of dividing Verbs into Transitive
Verbs and Intransitive Verbs is that one is essentially defining the categories in
terms of the construction they occur in, Transitive or Intransitive. These problems
are multiplied in cross-linguistic comparison, where the variation found is more
extreme (Croft 2001).
One can deal with such problems in the reductionist model by adding syntactic
features that prevent certain category members from occurring in the unacceptable
constructions, as in (40b) and (41a). Again, the effect is that one is introducing a
feature that specifies the category in terms of the construction it occurs in/does not
occur in (in this case, Transitive and/or Imperative and/or VP Conjunction).
Radical Construction Grammar takes a different approach to the relations of
constructions to their parts. It takes the constructions as the basic or primi-
tive elements of syntactic representation and defines categories in terms of the
An overview of construction grammars 285
constructions they occur in. For example, the elements of the Intransitive construc-
tion are defined as Intransitive Subject and Intransitive Verb, and the categories are
defined as those words or phrases that occur in the relevant role in the Intransitive
construction. In other words, Radical Construction Grammar rejects the existence
of atomic schematic units (see Table 9.2 in §9.4), because atomic schematic units
are defined independently of constructions. This differentiates Radical Construc-
tion Grammar from reductionist theories.
Radical Construction Grammar is a nonreductionist model because it takes the
whole complex structure as basic and defines the parts in terms of their occur-
rence in a role in the complex structure. In effect, Radical Construction Grammar
takes to its logical conclusion one of the strategies for handling these problems
in reductionist theories, namely the subdividing of classes and the employment of
syntactic features that essentially specify which constructions a particular word or
phrase occurs in (see Croft 2001, chapter 1).
Constructions are individuated like any other conceptual object, by categoriza-
tion. Constructions possess formal features, including word order, patterns of con-
tiguity and specific morphemes (or very small classes of morphemes) in particular
roles. Constructions are also symbolic units, and typically possess discrete mean-
ings. Radical Construction Grammar assumes a nonclassical category model, and
allows for prototypes and extensions of constructions, as well as the possibility of
gradience between construction types.
(42)
the song
s s s
r SONG
DEF
286 Cognitive approaches to grammatical form
(43)
IntrSbj IntrV TrSbj TrV TrObj
(45) The children [patient = Subject] were taken [distinct verb form] to school by
their parents [agent = Oblique].
(46) Upriver Halkomelem: patient = Subject, agent not Oblique, distinct verb form
tə́s -1 əm θ úλ́ ’à tə swı́yə qə
bump.into -ACCID -3SG.PASS she ART man
‘She was bumped into by the man’
(47) Bambara: patient = Subject, agent = Oblique, verb form not distinct:
o fo’ra dugutigi fè
3sg greet’compl.intr chief with
‘S/he was greeted by the chief.’
(48) Maasai: patient not Subject, agent prohibited, distinct verb form:
aa- dɔ l -i
1sg.obj- see -pass
‘I am seen.’
10.3 Conclusion
291
292 Cognitive approaches to grammatical form
usage-based model contrasts with the traditional structuralist and generative mod-
els of grammatical representation. In the structuralist and generative models,
only the structure of the grammatical forms determines their representation in
a speaker’s mind. For example, the traditional models make a sharp distinction
between regular and irregular word forms. Regular inflected word forms, such
as the English plural form boy-s, are derived by a highly general rule forming the
plural from the singular, because the structural relationship between boy and boys,
namely addition of -s, allows this possibility. Irregular word forms, such as the
plural form feet, do not have a straightforward structural relationship linking the
singular and the plural. Since they cannot be derived by a general rule, irregular
plural word forms are therefore listed in the lexicon.
In the usage-based model, properties of the use of utterances in communica-
tion also determine the representation of grammatical units in a speaker’s mind. In
particular, two usage-based properties are assumed to affect grammatical represen-
tation: the frequency of occurrence of particular grammatical forms and structures,
and the meaning of the words and constructions in use. In §11.2, we discuss four
hypotheses of the usage-based model derived from this assumption, and the evi-
dence for those hypotheses in morphology, the area of grammar most intensively
studied in the usage-based model (Bybee 1985, 1995, 2001). In §11.3, we discuss
how the hypotheses of the usage-based model can be applied to the representation
of syntactic constructions, their acquisition and syntactic change.
may be entrenched even though it is a regular instance of the noun plural schema
[Noun-s] because it has a high token frequency. In contrast, it is less likely that
the plural form cornices is entrenched, because its low frequency of occurrence is
probably insufficient to lead to the storage of this form independently of the base
form cornice and the noun plural schema [Noun-s]. This first hypothesis of the
usage-based model is summarized in (1):
(1) Hypothesis 1: the storage of a word form, regular or irregular, is a function of its
token frequency.
Bybee and Slobin observe that the forms that retain the irregular past in t are all
frequent, while the others are rare (in fact, one has dropped out of the language).
Further dynamic evidence that irregularity is correlated with frequency is the
evidence that low-frequency irregular forms are more likely to be regularized in
production. For example, Bybee and Slobin discovered that there was a signifi-
cant rank order correlation between the likelihood of regularization of irregular
past tense forms by preschool children and the token frequency of the verb in the
adult caretaker’s speech, such that lower token frequency correlated with a high
regularization rate (Bybee and Slobin 1982:270). Bybee and Slobin found signif-
icant correlations between token frequency and regularization for some though
not all irregular verb classes in experimental production tasks given to third-grade
children and adults (1982:270–71).
A more indirect piece of evidence that regular forms are stored independently
under some circumstances is that a regularly inflected word form, or a regularly
derived word form, may diverge semantically from its parent word. For exam-
ple, ‘something can be dirty without involving real dirt at all . . . someone can
soil an item without being anywhere near real soil’ (Bybee 1985:88). Examples
of divergence of a former inflectional form are clothes, formerly the plural of
cloth (Bybee 1985:91) and shadow, formerly an Old English oblique case form
of shade (Croft 2000:36). Presumably, semantic divergence presupposes the in-
dependent representation of the inflected form, which then is free to diverge in
meaning.
This evidence regarding irregular word forms has led some generative linguists,
including Pinker and colleagues, to accept that frequency effects associated with
degree of entrenchment are found with irregulars. Hence, they accept that the usage-
based model is valid for irregularly inflected word forms. However, Pinker and
colleagues argue that regularly inflected word forms are not sensitive to frequency
effects. Instead, regularly inflected word forms are represented by a grammatical
rule based only on the structure of the word forms and not any properties of their
use (Pinker and Prince 1994; Marcus et al. 1992). This model is called the dual-
processing model of grammatical representation.3
Evidence from an experiment on the more regular third person present inflec-
tion in English indicates that low frequency regular forms are not stored in the
lexicon, because they do not exhibit gang effects (Stemberger and MacWhinney
1988:111–12). Gang effects are effects on stored word forms of phonologically
3 In fact, the traditional model, in which regularly inflected forms are generated by a structural rule
and irregular forms are listed in the lexicon, is also a dual-processing model. The only innovation
in the model proposed by Pinker and colleagues is that the irregulars conform to the usage-based
model’s predictions, instead of merely being listed in the lexicon.
The usage-based model 295
similar stored word forms; if word forms are not stored, gang effects would not
occur. The absence of gang effects has been taken to imply that regularly inflected
forms are not stored (see, for example, Prasada and Pinker 1993). Other experi-
ments, however, do suggest that high frequency regular inflected forms are stored.
Stemberger and MacWhinney conducted an experiment in which subjects were
required to produce past tense forms of regular verbs at high speed, and errors
occurred significantly less often on high frequency regular past tense forms, im-
plying that the high frequency regular past tense forms are stored (Stemberger and
MacWhinney 1988:106). Bybee reports an experiment conducted by Losiewicz
(1992) which provides some evidence of a frequency effect for regular forms
(Bybee 1995:450–51). Losiewicz observed that the acoustic duration of word-
final /t/ or /d/ is shorter if it is part of the word than if it is the regular past tense
ending (e.g. rapt vs. rapped). If the difference in acoustic duration is due to storage
of the word form, then high-frequency regulars should have shorter final /t/ or /d/
than low-frequency regulars. In a sentence reading task, subjects had an average
7 msec difference in duration between high-frequency and low-frequency final /t/
or /d/ duration, which was highly significant.
The evidence reported in the last paragraph is compatible with a model in which
high-frequency regular word forms are stored but low frequency regular word
forms are not. The results suggest that frequency affects the storage of regular
word forms, and supports Hypothesis 1.4
4 It is possible that the absence of a frequency effect for regulars is what one would expect even in a
usage-based model; activation network models trained on regular inputs do not display a gang effect
(Daugherty and Seidenberg 1994).
296 Cognitive approaches to grammatical form
Type frequency is the number of different word forms that are instances of a
particular schema. The type frequency of the English past tense schema [Ve r b-
ed] is thus the number of regular past tense verbs in English. The English past
tense suffix -ed is highly productive under this account. There is a vast number of
lower frequency verbs with the -ed past tense suffix which reinforce the past tense
schema [Ve r b -ed]. In fact, of course, the past tense schema has three allomorphs,
/t/, /d/ and / d/. But each of these phonologically defined schemas has a high type
frequency of low token frequency instances, so each allomorph is highly productive
for its phonologically defined class.
There is another important aspect of productivity: one must be able to form a
coherent schema (Bybee 1995:430). That is, there must be enough resemblance
between the types that contribute to the entrenchment of the schema that one
The usage-based model 297
can form a schema in the first place. The closer the resemblance, the more en-
trenched is the schema. Each allomorph of the English productive past tense is a
coherent schema, that is, for the past tense suffix each allomorph defines a phono-
logically and semantically coherent category, namely [-t/PAST], [-d/PAST] and
[- Sd/PAST]. Moreover, the three allomorphs possess a phonological family re-
semblance, reinforced by their complementary phonological distribution (in terms
of the final segment of the verb stem) and identity of meaning. Because of the
family resemblance a speaker may form a superordinate category, notated here
[-ed/PAST] with the orthographic representation standing in for the phonological
schema.
Bybee argues that instances of a schema that have a high token frequency will
not contribute to the productivity of a schema (1985:132–34). Instances with a
high token frequency are strongly entrenched (1995:434). Only the entrenched
specific word form will be activated in language use and thus will not reinforce the
superordinate schema. On the other hand, word forms with a low token frequency
will not be as strongly entrenched (if they are entrenched at all; see §11.2.1).
Bybee argues that low frequency word forms will contribute to the entrenchment
of a schematic representation of the inflectional ending that applies across many
different word forms, including new forms. However, the examples that Bybee
gives in support of this hypothesis do not fully separate token frequency and type
frequency. Instances of a productive schema with a high token frequency, such as
the most common regular English past tense forms, are swamped by the number of
instances with a low token frequency. One would have to find a conjugation class
in a language where excluding the high token frequency instances results in too
low a type frequency to make the schema productive, but including them would
result in a high enough type frequency to make the schema productive. Hence,
it is not clear that high token frequency instances in fact do not contribute to the
productivity of a schema.
The usage-based definition of productivity is gradient, because type frequency
is gradient. Thus, the usage-based model predicts that productivity might vary in
degree. Forms for which there is a low type frequency may exhibit a minor degree
of productivity. Evidence for this is found among the irregular English past tense
forms. Most of the irregular English past tense forms involve an internal change to
the stem, usually a change to the stem vowel. One particular class has a relatively
high type frequency of relatively low token frequency verbs. This class is Bybee
and Slobin’s Class VI. Class VI verbs fall into two subclasses, those with a past
tense form with /æ/ and a past participle with / / (Class VIa in example [4]) and
those with a past tense form with / / (Class VIb; Bybee and Slobin 1982:288,
Appendix).
298 Cognitive approaches to grammatical form
English does not provide a good case to differentiate these two hypotheses for
productivity, type frequency vs. open schema. The regular English past tense has
both a much higher type frequency than any other past tense form and a very open
phonological schema. Thus, both hypotheses predict (correctly) that the regular
English past tense is highly productive. However, there are languages in which
the open schema for an inflection does not have a high type frequency. Examples
include the German plurals (Marcus et al. 1995) and Arabic plurals (McCarthy
and Prince 1990).5
The German and Arabic plural cases both dissociate type frequency from open
schema. Both German and Arabic have a range of plural formation processes,
none of which has an overwhelmingly greater type frequency such as is found
with the English plural -s. Both German and Arabic have an open schema plural,
the German -s and the Arabic sound plural (a suffix instead of an internal stem
change). Both German and Arabic open schema plurals are used as the plural
schema with items other than standard common nouns, such as proper names,
new borrowings and derived nouns and adjectives (Bybee 1995:440–42; Janda
1990:146–48). The open schema plurals are open schema precisely because they
must be applicable to noncanonical nouns; but they are also of low type frequency
for the same reason.
5 Clahsen and Rothweiler (1992) argue that the German past participle ending -t has a lower type
frequency than -en and yet is more productive. However, Clahsen and Rothweiler used only the first
1,000 verbs of a 4,314-verb frequency list of German verbs, thereby leaving out a very large number
of regular verb types, and they counted the verb stems multiple times if they occurred with multiple
productive prefixes, which again artificially increases the type frequency of -en (Bybee 1995:438).
300 Cognitive approaches to grammatical form
Marcus et al. and Pinker and Prince argue that the applicability of the open
schemas to noncanonical nouns is evidence of their high productivity. Bybee, on
the other hand, argues that this is evidence only of their ‘emergency’ or default
status by virtue of their open schema, and that investigating the full range of
common nouns indicates that the default schema is no more productive than plural
schemas with a high type frequency. For Arabic, the iambic broken plural has
the highest type frequency, and is productive with any noun (including borrowed
nouns) that fit its canonical phonological shape criteria. The iambic broken plural
is overgeneralized by children as well as the default sound plural (Omar 1973,
cited in Bybee 1995:442), indicating that both the iambic broken plural and the
default sound plural are productive, the former especially so.
The German plural situation is more complicated. No single plural formation
pattern is highly productive. However, the evidence for productivity apart from
noncanonical nouns indicates some degree of productivity of several different
plural endings, including -s (see Bybee 1995:440–41 and references cited therein).
Children overgeneralize -en most frequently, and in nonce-probe tasks, different
endings were preferred for different noun genders, particularly if the nonce forms
were identified as common nouns. Even recent borrowings use -en and to a lesser
extent -e especially for masculine nouns; -s is used about half of the time, and
some loans have given up -s for -en upon integration into the language. Also,
nouns ending in vowels favored -s in a nonce-probe task (Köpcke 1988) and in
acquisition (Köpcke 1998:313–15). Among ordinary German common nouns -s
is associated with nouns ending in a full vowel (Janda 1990:145–46). These facts
suggest that -s is not truly a default schema.
A default schema does not require a high type frequency to arise. A default form
can arise if the non-default forms form relatively narrow and phonologically well-
defined classes, while the instances of the default schema are scattered across the
remaining phonological space, even if those instances have a low type frequency
(Hare, Elman and Daugherty 1995:626–27; they also simulate this effect in a
connectionist network). The German and Arabic plurals conform to this pattern
(compare Hare et al. 1995:608). If the phonological unity of the non-default classes
breaks down through phonological change, as happened with the Old English
past tenses, then the system becomes unstable, leading to a reorganization of
the irregular forms around new phonological classes or survival by high token
frequency (Hare and Elman 1995).
can be equally well captured by rules deriving one form from another (or both
forms from a common underlying form). Instead, we have focused on the fact that
the usage-based model can capture generalizations about morphological patterns
that are based on token frequency and type frequency, whereas rules in the dual-
processing model do not imply the existence of any generalizations linked to
frequency, at least for regular inflections. However, Bybee and others have argued
that rules and schemas in fact make slightly different predictions about what sort
of generalizations can be made over related word forms.
Structuralist and generative morphological rules are what Bybee calls source-
oriented. Source-oriented rules specify the basic word form, such as the present
tense of a verb like wait, and describe a single operation with a single set of
conditions that produce the derived form, such as addition of / d/ to a stem ending
in t or d to form waited. In the usage-based model, a source-oriented schema is a
schema with a systematic structural relationship to another schema (the source in a
rule-based model). For instance the regular past tense schema [Ve r b- d] contains
the same stem as the present tense schema [Ve r b(- z)], and so the past and present
tense verb form schemas can be uniformly represented across all stems. The term
‘source-oriented’ is somewhat misleading in the usage-based model because there
is no derivation of one schema from another by a rule in the usage-based model. The
term ‘source-oriented’ simply indicates that the ‘source’ schema is as coherent,
phonologically and semantically, as the ‘product’ schema.
Bybee argues (following Zager 1980) that in addition to source-oriented
schemas, there also exist product-oriented schemas:
(7) Hypothesis 4: strength of connection between word forms, and thus forces influ-
encing their phonological shape (among other things), is a function of similarity.
Similarity is measurable by comparing words to each other in both meaning and
form; similarity in meaning is much stronger than similarity in form.
In contrast, changing the person who is doing the eating does not make a very
substantial change in the nature of the eating event itself:
(10) valence changing < voice < aspect < tense < mood < person/number agreement
The usage-based model 305
Bybee puts forward several types of typological and diachronic evidence for
the ranking in (10). For example, more relevant inflectional categories of the verb
occur closer to the verb stem. The reasoning behind this prediction is that the
greater the meaning change, the more intimately associated with the stem meaning
is the semantic category of the inflection. Bybee tested the hypothesis on the
four most common verbal inflectional categories, aspect, tense, mood and person
agreement, in a fifty-language sample (Bybee 1985:33–35). There were virtually
no counterexamples to the ordering of aspect, tense and person/mood with respect
to the other categories; mood and person agreement were more equivocal.
The notion of relevance is a further refinement of Hypothesis 4: that seman-
tic similarity to different degrees influences formal similarity of word forms to
different degrees. That is, greater semantic similarity will favor greater phono-
logical similarity (and thus increase symbolic similarity). Semantic distinctions
expressed lexically will be more phonologically different than semantic distinc-
tions expressed inflectionally, on the whole. It should be remembered that other
factors such as degree of entrenchment also affect phonological similarity: a higher
degree of entrenchment weakens the connection between word forms and thus may
lead to greater phonological differences.
Thus, another prediction from Hypothesis 4 is that, other things being equal,
a stronger semantic connection between words will imply a greater phonological
similarity of those words. Also, one would expect to find that phonological similar-
ity can be increased through analogical change of semantically strongly connected
words. There is considerable evidence for this prediction as well. Bybee reports
that data from her survey and from Rudes (1980) indicate that suppletion in verbal
paradigms is most likely along aspectual distinctions, then along tense, and least
likely along mood. There is also some suppletion along person distinctions, but
only in very high frequency forms; this can be explained by the principle given in
the preceding paragraph, that a high degree of entrenchment weakens connections
between words.
Finally, when paradigms are leveled analogically, they are most likely to be
leveled among semantically closely related forms, in particular different per-
son/number forms of the same tense-aspect-mood paradigm. Also, the direction of
leveling will be most likely towards the most frequent form (third person singular,
followed by first person singular), because forms with weak connections will give
way to analogical formations from stronger forms. Bybee presents a number of
examples of such leveling within person-number forms (Bybee 1985, chapter 3;
see also Bybee and Brewer 1980 for Spanish). For example, the Old Provençal
person-number forms for the preterite indicative in (11) were reformed analogi-
cally on the third person form including that form’s -t person/number suffix, in a
number of Modern Provençal dialects, such as the Charente dialect in (12) (Bybee
1985:55; Charente data from Meyer-Lübke 1923:352):
306 Cognitive approaches to grammatical form
In (12), the original third person singular stem cantét has become the base for
the analogical reformation of the other person-number forms. The only form to
resist the analogical change is the first singular form, which has the highest token
frequency of the person-number forms after third singular.
Bybee and Pardo (1981) used a nonce-probe task with Spanish speakers to com-
pare the effect of semantic connection on phonological production. Many Spanish
verbs have a vowel stem alternation in the present vs. preterite forms, for example
third singular present comienza ‘s/he begins’ with a diphthong vs. third singular
preterite comenzó ‘s/he began’ with a simple mid vowel. Bybee and Pardo pre-
sented a third singular present form of a nonce verb with a diphthong followed by
either the infinitive or the third singular preterite form with a mid vowel, and then
asked subjects to produce a first singular preterite form. Subjects produced more
mid vowel variants when presented with the semantically closer related third sin-
gular preterite than when presented with the more distant infinitive.
The evidence presented by Bybee and others (e.g. Andersen 1980) implies that
semantic similarity of grammatical units such as words plays an important role
in the organization of grammatical knowledge in a speaker’s mind. Given a set
of word forms, each of which can be subsumed under several more schematic
categories (indicative, present, third person, singular), one can postulate a ranking
of those schematic categories in terms of the network connections between words.
One could go so far as to restructure the multiple-parent representation into a
hierarchy, with the semantically most relevant (and hence most weakly connected)
distinctions at the top of the hierarchy, as in (13) (compare the display of the
Spanish verb paradigm in Bybee 1985:61, Table 1):
(13)
Present Past
1sg 2sg 3sg ... 1sg 2sg 3sg ... 1sg 2sg 3sg ... 1sg 2sg 3sg ...
The usage-based model 307
11.2.5 Conclusion
The empirical data that we have discussed in support of the details of
the usage-based model are drawn from the processes of language use (as tested in
psycholinguistic experiments) and language change. These data provide evidence
supporting four hypotheses about the effects of language use on grammatical
representation. The independent representation of an inflected word form is a
function of its token frequency in language use. The productivity of a rule/schema
is a function of a high type frequency of low token frequency instances, not of
the structural openness of the schema. Product-oriented schemas exist, that is,
schemas can be formed from members of an inflectional category that cannot
be described by rules deriving the members of the category from a source form.
Finally, the organization of inflected word forms is influenced by the degree of
semantic similarity between word forms.
The hypotheses of the usage-based model can be accounted for by an interactive
activation network for the representation of knowledge (Elman and McClelland
1984). The storage of word forms is determined in part by patterns of activation of
the network as a result of language use (§11.2.1). The phenomena described in Hy-
potheses 2–4 of the usage-based model are all analyzed in terms of the interaction
of activation patterns, such that a schema activates an instance and vice versa, and
a structure’s activation can result from the activation of formally and especially se-
mantically related structures. The result of the interactive activation is manifested
not only in the conventional production and comprehension of word forms, but
also in ‘errors’ in certain contexts, and innovations in language acquisition and
language change.
308 Cognitive approaches to grammatical form
The usage-based model contrasts with the complete inheritance model in two
important respects. As we noted in §10.2.2, in the complete inheritance model
information is stored only at the most schematic level possible. In the usage-based
model information can be represented redundantly in less schematic construc-
tions, if activation levels lead to entrenchment. The second contrast follows from
the first. In the complete inheritance model, information flows down from the
most schematic constructions in the processing of an utterance. In the usage-based
model, processing involves activation of the entrenched construction(s) whose
structure(s) most closely matches those of the utterance. Since more specific con-
structions match utterances more closely than more schematic constructions, the
former are more activated than the latter. It is therefore possible that speakers will
not have the most schematic constructions represented in their minds, if they are
not activated sufficiently (Croft 1998c).
(15)
prevaricate lie
(16) ew -ed
The network structure for low vs. high token frequency (entrenchment) and
low vs. high type frequency (productivity) is the same: the higher the number
of instances, the more entrenched the superordinate category is. We may then
formulate the following generalized definition of productivity:
Bybee and Thompson also discuss a large scale corpus study of the French sub-
junctive by Poplack (1992, 1996). The French subjunctive form is disappearing
from the spoken language, but is still variably used. However, the French subjunc-
tive has survived largely in the complements of the highly frequent main clause
verb falloir ‘have to,’ and/or in the most highly frequent complement verbs, in-
cluding avoir ‘have,’ être ‘be,’ aller ‘go’ and faire ‘make, do,’ as predicted by the
usage-based model.
The example of English auxiliaries in questions and negative sentences il-
lustrates the maintenance of irregularity in more entrenched constructions that
have resisted changes in the more schematic constructions of the language.
Another respect in which more entrenched constructions are irregular is that
they undergo changes that less entrenched constructions do not undergo. These
changes commonly involve reduction, a typical concomitant of high frequency. In
schematic constructions, reduction applies of course to the substantive unit(s) of the
construction.
An example of reduction in a highly entrenched construction is the contraction
of not with the auxiliary in the [Sbj Aux -n’t . . .] construction. This contraction
is recognized in written English, and includes the fused form won’t. Of course,
such reduction originated in the spoken language, and one would expect to find
reduction of other negative-auxiliary contractions in the spoken language. Bybee
and Scheibman (1999) investigate the reduction of one specific negative auxiliary
form, don’t, in spoken American English conversation. They demonstrate that the
phonetic reduction of don’t is strongly correlated with the frequency of the verb
and of the subject with which don’t is combined in the [S b j don’t Ve r b . . .]
construction. The highest frequency subject is the first person singular I, and the
highest frequency verb in this construction is know; in fact, this reduction is so
salient that it is loosely represented orthographically as I dunno. But I dunno is
only the extreme end of a continuum of phonetic reduction that spans the full range
of verbs and subjects used in this construction. Bybee and Scheibman further note
that the reduction applies across the substantive units in the constructions as a
whole, regardless of their internal constituent structure.
In addition to syntactic irregularity as a consequence of high frequency, one
would expect to find degrees of syntactic productivity. Of course, maximal syntactic
productivity is the characteristic of the major, most schematic constructions of
the language, such as the transitive construction [S b j Ve r b O b j]. The high
productivity of this highly schematic construction is due to the very high frequency
of instances of this construction, due to the high number of transitive verbs, the
vast majority of which have a relatively low token frequency. However, even with
completely schematic syntactic constructions, one can find varying degrees of
productivity.
312 Cognitive approaches to grammatical form
The two constructions overlap in their distribution: some verbs allow both con-
structions, with approximately the same meaning, as in (22)–(23). The Ditransitive
construction is used with many fewer verbs than the Caused Motion construction,
and hence is much less productive than the Caused Motion construction (Goldberg
1995:124):
(24) Sally whispered some terrible news to him.
(25) *Sally whispered him some terrible news.
Finally, one must also differentiate between productive syntactic schemas and
default syntactic schemas, used, for example, in borrowing. Many languages, par-
ticularly languages with complex verbal morphology, do not directly incorporate
borrowed verbs into the productive native syntactic constructions of the language.
Instead, they use a default construction, combining an invariant form of the bor-
rowed verb with an inflected native verb (often meaning ‘make, do’) in a construc-
tion [B o r r Ve r b ‘make’ . . .]. An example of this phenomenon can be found
in K’iche’ Mayan, which uses the verb ban ‘make’ with the infinitive form of
borrowed Spanish verbs, as in (28) (Mondloch 1978:117):
(29) a. Active: [S b j i Ve r b k -t n s O b j j ]
b. Passive: [S b j j be-t n s Ve r b k -pp by O b j i ]
the outputs of different ‘rules’ and ‘input constructions.’ Of course, since the
syntactic constructions in question are themselves schematic, rules can be devised
for each construction type. But this is no different than having a ‘rule’ for each word
in morphology; it is the same phenomenon but at a higher level of schematicity.
The crucial point is that there is a higher degree of structural coherence defining
the product-oriented schema than its counterpart ‘source’ schema(s).
There are some strong candidates for product-oriented syntactic schemas in
English. One is the pair of interrogative and negative construction types discussed
in §11.3.1. A very general schema can be formed for each: [Au x S b j . . . ?]
and [Sbj Aux -n’t . . .]. However, there would have to be at least two rules
linking a source schema to the interrogative or negative product schema: a rule
inserting the auxiliary do for sources without an auxiliary, and a rule applying
to the (first) auxiliary for sources with an auxiliary. The different source-product
correspondences are illustrated for the question schema in (30)–(32):
Other analyses have been proposed for the constructions in (33b–c): some have
argued that it in (33b) refers to the general ambience, and others have argued that
there is only one ‘surface’ subject phrase in (33c), the ‘underlying’ subject having
been extraposed to a different syntactic position. However, whatever rule-based
The usage-based model 315
7 We disregard here the existence of constraints on the ‘path of movement’ (as it would be described
in a rule-based transformational model). Although it is widely assumed that the constraints must
be formulated in syntactic terms (beginning with Ross 1967), there are many counterexamples,
and alternative accounts have been formulated in semantic/pragmatic terms (see Deane 1991 and
references cited therein).
316 Cognitive approaches to grammatical form
different arguments. In K’iche’, for example, questions, relatives and focus con-
structions formed on the ergative (transitive subject) argument require the -Vn
focus antipassive verb form (see example [38]; Mondloch 1978:74); those formed
on the absolutive (intransitive subject or transitive object) argument retain the ac-
tive voice verb form ([39]; Larsen and Norman 1979:357); those formed on the
instrument require the -bej focus antipassive form ([40]; Norman 1978:462); and
those formed on locative/directional phrases retain the active voice form but leave
the demonstrative pronoun wih in the normal oblique position ([41]; Mondloch
1978:42):
(38) jachin x- Ø- cun -an lē yawab?
who pf- 3sg - cure -antipass the sick.one?
‘Who cured the sick one?’
(39) jachin x- Ø- u- ch’ay -Ø lē achi
who pf- 3s g . a b s- 3s g . e r g- hit -act the man
‘Who did the man hit?’
(40) jas x- Ø- u- rami -bej lē achih
what pf- 3s g . a b s - 3s g . e r g - cut -inst.pass the man
r- ē le chēʔ
3s g . p o s s - g e n the tree
‘What did the man use to cut the tree?’
(41) jawiʔ c- at- bē wi?
where impf- 2sg.abs - go wi
‘Where did you go?’
Nevertheless, in K’iche’ as in English, there is a coherent product-oriented schema
such that the questioned, relativized or focused, phrase is put in initial position,
whatever happens to the rest of the sentence. This generalization is also productive,
and is another example of a product-oriented schema.
Another example of a product-oriented schema is the Japanese passive. The
Japanese passive construction is characterized by: (i) a subject NP which may be
marked with ga (‘subject’), wa (‘topic’), or may be absent altogether if highly
topical; (ii) an oblique agent NP marked with ni, which is optional; and (iii) a
verb form in -(r)are. The subject of a Japanese passive need not be merely the
object of the corresponding active, as in (42) (Tsukiashi 1997:18; all examples are
attested):
We may describe the overall schema as [Sbj QuasiModal . . . Calv ə VP]. This
pattern seems to be mildly productive, in that almost all of the grammaticalizing
quasimodals have reduced to forms ending in [. . . Calv ə ]; of course, there is a
relatively low type frequency to this construction.
It should be pointed out that in the past two and a half decades of generative
syntactic research, emphasis has shifted from the description of rules to the de-
scriptions of constraints (‘principles’) on the output of rules (see, for example,
Chomsky 1981:3–4; 1993:5). To the extent that the principles and constraints of
generative grammar describe the structure of the ‘product’ schema, then genera-
tive grammar constraints are handling essentially the same kind of phenomena as
product-oriented schemas in the usage-based model. However, current generative
syntactic models utilize abstract syntactic structures and derivational processes,
and generate a wide range of outputs, many of which are invalid (‘crash’; Chomsky
1993:5) and only a few of which are actually occurring linguistic expressions (‘con-
vergence’; ibid.). In contrast, the usage-based model represents schemas abstracted
inductively from actually occurring utterances; in syntax, as in morphology, the
usage-based model does not posit underlying structures or nonexistent structures
that are filtered out (see also footnote 7).
We first offer semantic arguments for the relevance ranking in (49). Relevance
of sentences pertains to the meaning of the utterance in context. The illocutionary
force of an utterance has the greatest semantic effect on the meaning of a sentence,
since it alters the speaker’s intention and the hearer’s response to a proposition if
it is presented as an assertion, question, command or other speech act. Predicate
The usage-based model 319
and so on; and only later will the child learn to use different arguments with the
verb, such as Break cup, Mommy break cup, Break with stick and so on. These
results imply that children generalize across different participant types quickly and
early, but only later generalize across predicate types and the argument structure
constructions that characterize them.
The evidence presented here, if it is borne out by further studies, would al-
low us to restructure the taxonomic organization of sentence-level constructions.
Although a taxonomic organization for constructions must allow for multiple par-
ents, we can use degree of semantic similarity to rank the syntactic distinctions and
thus form a hierarchy, as was done in example (13) (§11.2.4) for morphological
paradigms. Such an organization is illustrated in (51) for the Imperative half of the
basic Declarative-Imperative split:
(51) Imperative: (POL) PREDPHRASE
,
Positive: PREDPHRASE Negative: Don t PREDPHRASE
, ,
VERB be ADJ etc. Don t VERB Don t be ADJ etc.
, ,
Jump! etc. Be happy! etc. Don t jump! etc. Don t be cruel! etc.
(52)
Subjectless Clause
,
Don t
positive negative
action action
imperative imperative
positive negative
property property
imperative imperative
Copula
represented by relative distance in the conceptual space: (in [52], for example, the
negative functions are closer to each other than either is to their positive counter-
parts). These relative distances impose constraints on the taxonomic organization
of construction (in this case, requiring the positive and negative forms to be grouped
together first in the taxonomy; compare [52] to [51]). Further structures must be
imposed on conceptual space to allow for the formulation of constraints on the
grammatical expression of conceptual structures (see, e.g., Croft 2001:160–61,
163–64, 169–70; 2003b:140–43).
The exploration of semantic relations between constructions and their con-
straints on formal properties of constructions is still in its infancy. Presumably,
further research will allow construction grammarians to impose further structure
on the network organization of syntactic as well as morphological knowledge.
[Determiner Noun ] noun phrase category early on. Pine and Lieven (1997)
found that at the earliest stage of learning nouns and determiners, children also
proceed in a piecemeal fashion. In their study, Pine and Lieven found that, although
children use a variety of nouns with both a and the, the nouns they use with a and the
nouns they use with the overlap very little at first. Instead, it appears that children
learned nouns with one determiner, or that the determiner use was associated
with larger structures in which the noun and determiner occur, such as [in the X]
or [That’s a X].
Pine, Lieven and Rowland (1998) studied the first six months of twelve chil-
dren’s multiword speech and found evidence that children begin with lexically
quite specific constructions, but that it was not always verbs that functioned as the
‘islands’ around which utterances were learned. For example, children produced
utterances with the auxiliaries can, do, be and have, constituting an average 90.3%
of all children’s utterances (Pine et al. 1998:818). However, there was very lit-
tle overlap between the verbs used with each auxiliary for each child (only one
child with one pair of auxiliaries had an overlap significantly different from zero;
1998:819). This suggests that the children are learning lexically specific auxiliary-
verb combinations and have not yet developed a productive [Au x] or [Ve r b ]
category in their utterances.
A still more fine-grained study confirms that early acquisition begins piecemeal
and indicates that acquisition is sensitive to token frequency in the input. Rubino
and Pine (1998) conducted a longitudinal study of a child learning Brazilian Por-
tuguese, and argued that the acquisition of sentence constructions with subject-verb
agreement began in a piecemeal fashion, with the acquisition of singular and plural
agreement beginning independently (in fact, in succession). Later, the child began
to overregularize the third singular agreement affixes at the time that the child
began to produce the third plural agreement affixes (Rubino and Pine 1998:51).
This developmental sequence suggests that an initial stage of rote learning was
followed by ‘joining the islands’ of singular and plural agreement to induce a
system of number agreement in the third person.
The overall average of errors produced by the child was quite low, which is
what one would expect in a model of conservative, inductive language learning. A
breakdown of error rates by person and number indicated that frequency of forms
in the input defined the course of acquisition of subject-verb agreement. The child
acquired correct subject-verb agreement for the most frequent agreement forms in
the input first, and the first correct productions of the less frequent subject-verb
agreement combinations appeared with high frequency verbs in the input (Rubino
and Pine 1998:53). However, frequency in the input does not appear to be the only
factor determining acquisition. Gathercole, Sebastián and Soto (1999) examined
the acquisition of Spanish verbal forms in two children, and observed the same
The usage-based model 325
All of the way construction examples given in (53) use a possessed direct object way
and require a complement describing the path of motion. Example (53a) describes
a means of achieving the motion along the path; (53b) describes a manner of
motion along the path, and example (53c) describes an incidental activity of the
subject as she travels along the path. The way construction is also syntactically
and semantically idiosyncratic: the verbs in the way construction are normally
intransitive, and their meaning does not normally entail motion.
Using data from the Oxford English Dictionary and the Oxford University Press
corpus of contemporary English, Israel argues that the modern way construction
grew gradually from two different, more narrowly used way constructions, the
means and manner constructions (a third source, the acquisition or continued
possession of a path, shrank rather than expanded, although it remains in certain
common instances such as find one’s way; Israel 1996:221, n. 3). The manner
construction began as a special case of the Middle English [go one’s Pat h]
construction, and was originally found with only the most common general motion
verbs, no more than sixteen verbs before 1700 (Israel 1996:221). Verbs encoding
326 Cognitive approaches to grammatical form
manner and path of motion began to be used with the manner way construction,
and in the nineteenth century, expanded particularly in the domain of laborious
motion (plod, scramble, fumble) and tortuous path (thread, worm, insinuate). At
the end of the nineteenth century, the manner way construction expanded to verbs
expressing noise accompanying motion (crunch, crash, toot).
The means way construction does not emerge until around 1650, and began
with verbs describing path clearing (cut, furrow out) and road building (pave,
smooth), as well as forcible motion (force out; Israel 1996:223). Expansion begins
with the cutting verbs and extends to fighting verbs starting around 1770. In the
nineteenth century, progressively more indirect means of reaching the goal, as in
He . . . smirked his way to a pedagogal desk (Israel 1996:224, from New Monthly
Magazine VII.386, 1823). At this point the means and manner way constructions
appear to merge, and in the late nineteenth century one finds the first examples of
incidental activity, which is quite distantly related to motion, as in He . . . whistled
his way to the main front door (Israel 1996:225, from Blackmore, Cradock Nowell
xvi, 1866).
At the same time that the class of verbs in the way construction is expanding,
the overall syntactic form of the construction becomes narrower, from allowing
other nouns than way and an optional path expression to obligatory way and path
expression (Israel 1996:221, 226). This (common) pattern in syntactic change
illustrates how a new construction emerges from an often highly specific instance
of an existing construction schema and then expands in its own direction. A usage-
based model can account for this pattern in that it allows for the entrenchment of
specific instances of construction schemas, which function as ‘islands’ from which
a new construction expands, establishing and generalizing a new construction
schema with its own syntactic and semantic peculiarities.
11.4 Conclusion
328
Conclusion 329
Akmajian, Adrian. 1984. Sentence types and the form-function fit. Natural Language and
Linguistic Theory 2:1–23.
Allan, Keith. 1986. Interpreting English comparatives. Journal of Semantics 5:1–50.
Andersen, Henning. 1980. Russian conjugation: acquisition and evolutive change. Papers
from the 4th International Conference on Historical Linguistics, ed. Elizabeth
Closs Traugott et al., 285–302. Amsterdam: John Benjamins.
Ariel, Mira. 1990. Accessing noun phrase antecedents. New York: Routledge.
Armstrong, Sharon Lee, Lila R. Gleitman and Henry Gleitman. 1983. What some concepts
might not be. Cognition 13:263–308.
Aronoff, Mark. 1993. Morphology by itself: items and inflectional classes. Cambridge,
Mass.: MIT Press.
Atran, Scott. 1990. Cognitive foundations of natural history: towards an anthropology of
science. Cambridge: Cambridge University Press.
Barlow, Michael and Suzanne Kemmer (eds.). 2000. Usage-based models of language.
Stanford: Center for the Study of Language and Information.
Barsalou, Lawrence W. 1983. Ad hoc categories. Memory and Cognition 11: 211–27.
1987. The instability of graded structure: implications for the nature of concepts. Concepts
and conceptual development: ecological and intellectual factors in categorization,
ed. Ulric Neisser, 101–40. Cambridge: Cambridge University Press.
1992a. Frames, concepts, and conceptual fields. Frames, fields and contrasts: new essays
in semantic and lexical organization, ed. Adrienne Lehrer and Eva Feder Kittay,
21–74. Hillsdale, N.J.: Lawrence Erlbaum Associates.
1992b. Cognitive psychology: an overview for cognitive scientists. Hillsdale, N.J.:
Lawrence Erlbaum Associates.
Becker, Joseph. 1975. The phrasal lexicon. Bolt, Beranek and Newman Report No. 3081,
Artificial Intelligence Report No. 28.
Berlin, Brent and Paul Kay. 1969. Basic color terms: their universality and evolution.
Berkeley: University of California Press.
Berlin, Brent, Dennis E. Breedlove and Peter H. Raven. 1973. General principles
of classification and nomenclature in folk biology. American Anthropologist
75:214–42.
Birner, Betty J. and Gregory Ward. 1998. Information status and noncanonical word
order in English. (Studies in Language Companion Series, 40.) Amsterdam: John
Benjamins.
Black, Max. 1993[1979]. More about metaphor. Metaphor and thought (2nd ed.), ed. An-
drew Ortony, 19–41. Cambridge: Cambridge University Press.
330
References 331
Bolinger, Dwight. 1971. The phrasal verb in English. Cambridge, Mass.: Harvard University
Press.
1980. Language, the loaded weapon. London: Longmans.
1992. About furniture and birds. Cognitive Linguistics 3:111–18.
Braine, Martin D. S. 1976. Children’s first word combinations. (Monographs of the
Society for Research in Child Development 41, no. 1.) Chicago: University of
Chicago Press.
Bresnan, Joan (ed.). 1982. The mental representation of grammatical relations. Cambridge,
Mass.: MIT Press.
Brown, Cecil H. 2002. Paradigmatic relations of inclusion and identity I: Hyponymy.
Lexicology: an international handbook on the nature and structure of words and
vocabularies (Handbooks of Linguistics and Communication Science, 21), ed.
D. Alan Cruse, Franz Hundsnurscher, Michael Job and Peter Lutzeier, ch. 47. Berlin:
Walter de Gruyter.
Brown, Roger. 1958. How shall a thing be called? Psychological Review 65:14–21.
Bybee, Joan L. 1985. Morphology: a study into the relation between meaning and form.
Amsterdam: John Benjamins.
1995. Regular morphology and the lexicon. Language and Cognitive Processes
10:425–55.
2001. Phonology and language use. Cambridge: Cambridge University Press.
Bybee, Joan L. and Mary Alexandra Brewer. 1980. Explanation in morphophonemics:
changes in Provençal and Spanish preterite forms. Lingua 52:271–312.
Bybee, Joan and Paul Hopper. 2001. Frequency and the emergence of linguistic structure.
(Typological Studies in Language, 45.) Amsterdam: John Benjamins.
Bybee, Joan L. and Carol Lynn Moder. 1983. Morphological classes as natural categories.
Language 59:251–70.
Bybee, Joan L. and Elly Pardo. 1981. On lexical and morphological conditioning of alter-
nations: a nonce-probe experiment with Spanish verbs. Linguistics 19:937–68.
Bybee, Joan L. and Joanne Scheibman. 1999. The effect of usage on degrees of constituency:
the reduction of don’t in English. Linguistics 37:575–96.
Bybee, Joan L. and Dan I. Slobin. 1982. Rules and schemas in the development and use of
the English past tense. Language 58:265–89.
Bybee, Joan L. and Sandra A. Thompson. 1997. Three frequency effects in syntax. Proceed-
ings of the 23rd Annual Meeting of the Berkeley Linguistics Society, ed. Matthew
L. Juge and Jeri O. Moxley, 378–88. Berkeley: Berkeley Linguistics Society.
Chafe, Wallace. 1994. Discourse, consciousness and time: the flow and displacement of
conscious experience in speaking and writing. Chicago: University of Chicago Press.
Chafe, Wallace (ed.). 1980. The pear stories. New York: Ablex.
Chomsky, Noam. 1981. Lectures on government and binding. Dordrecht: Foris.
1993. A minimalist program for linguistic theory. The view from Building 20, ed. Kenneth
Hale and Samuel Jay Keyser, 1–52. Cambridge, Mass.: MIT Press.
1995. The minimalist program. Cambridge, Mass.: MIT Press.
Clahsen, Harald and M. Rothweiler. 1992. Inflectional rules in children’s grammar: evidence
from the development of participles in German. Yearbook of Morphology 1992,
1–34.
Clark, Eve V. and Herbert H. Clark. 1979. When nouns surface as verbs. Language
55:767–811.
332 References
1990. Prototype theory and lexical semantics. Meanings and prototypes: stud-
ies in linguistic categorization, ed. Savas L. Tzohatzidis, 382–402. London:
Routledge.
1992a. Antonymy revisited: some thoughts on the relation between words and concepts.
Frames, fields and contrasts: new essays in semantic and lexical organization, ed.
Adrienne Lehrer and Eva Feder Kittay, 289–306. Hillsdale, N. J.: Lawrence Erlbaum
Associates.
1992b. Cognitive linguistics and word meaning: Taylor on linguistic categorization.
(Review article on John R. Taylor, Linguistic categorization: prototypes in linguistic
theory.) Journal of Linguistics 28:165–83.
1994. Prototype theory and lexical relations. Rivista di Linguistica 6:167–88.
2000a. Lexical “facets”: between monosemy and polysemy. Sprachspiel und Bedeu-
tung: Festschrift für Franz Hundsnurscher zum 60. Geburtstag, ed. S. Beckmann,
P. P. König and T. Wolf, 25–36. Tübingen: Max Niemeyer Verlag.
2000b. Aspects of the micro-structure of word meanings. Polysemy: theoretical and
computational approaches, ed. Yael Ravin and Claudia Leacock, 30–51. Oxford:
Oxford University Press.
2002a. Microsenses, default specificity and the semantics-pragmatics boundary.
Axiomathes 1:1–20.
2002b. Hyponymy and its varieties. The semantics of relationships: an interdisciplinary
perspective, ed. Rebecca Green, Carol A. Bean and Sung Hyon Myaeng, 3–21.
Dordrecht: Kluwer.
Cruse, D. Alan., Franz Hundsnurscher, Michael Job and Peter Lutzeier (eds.). 2002.
Lexicology: an international handbook on the nature and structure of words and
vocabularies. (Handbooks of Linguistics and Communication Science, 21.1.)
Berlin: Walter de Gruyter.
Cruse, D. Alan and Pagona Togia. 1995. Towards a cognitive model of antonymy. Lexicology
1:113–41.
Daugherty, K. and Mark Seidenberg. 1992. Rules or connections? The past tense revisited.
Proceedings of the 14th Annual Meeting of the Cognitive Science Society 259–64.
Hillsdale, N. J.: Erlbaum.
1994. Beyond rules and exceptions: a connectionist modeling approach to inflectional
morphology. The reality of linguistic rules, ed. Susan D. Lima, R. L. Corrigan and
Gregory K. Iverson, 353–88. Amsterdam: John Benjamins.
Deane, Paul D. 1988. Polysemy and cognition. Lingua 75:325–61.
1991. Limits to attention: a cognitive theory of island phenomena. Cognitive Linguistics
2:1–63.
DeLancey, Scott. 1981. An interpretation of split ergativity and related patterns. Language
57:626–57.
Denison, David. 1993. English historical syntax. London: Longman.
1998. Syntax. The Cambridge history of the English language, vol 4: 1776–1997, ed.
Suzanne Romaine, 92–329. Cambridge: Cambridge University Press.
Dreyfus, Hubert L. 1991. Being-in-the-world: a commentary of Heidegger’s Being and
Time, Division I. Cambridge, Mass.: MIT Press.
Eco, Umberto. 1996. Metaphor. Philosophy of language: an international handbook of
contemporary research (Handbooks of Linguistics and Communication Science,
7.2), ed. Marcelo Dascal, Dietfried Gerhardus, Kuno Lorenz and Georg Meggle,
1313–23. Berlin: Walter de Gruyter.
334 References
Fillmore, Charles J., Paul Kay and Mary Kay O’Connor. 1988. Regularity and idiomaticity
in grammatical constructions: the case of let alone. Language 64:501–38.
Fillmore, Charles J., Paul Kay, Laura A. Michaelis and Ivan A. Sag. In preparation.
Construction Grammar. Stanford, Calif.: CSLI Publications.
Flickinger, Daniel, Carl Pollard and Thomas Wasow. 1985. Structure-sharing in lexical rep-
resentation. Proceedings of the 23rd Annual Meeting of the Association for Compu-
tational Linguistics, 262–67. Chicago: Association for Computational Linguistics.
Gadamer, Hans-Georg. 1989. Truth and method, trans. William Glen-Doepel, ed. John
Cumming and Garrett Barden, translation rev. Joel Weinsheimer and Donald G.
Marshall. London: Sheed & Ward.
Gathercole, Virginia C. Mueller, Eugenia Sebastián and Pilar Soto. 1999. The early acqui-
sition of Spanish verbal morphology: across-the-board or piecemeal knowledge?
International Journal of Bilingualism 3:133–82.
Gazdar, Gerald, Ewan Klein, Geoffrey Pullum and Ivan Sag. 1985. Generalized Phrase
Structure Grammar. Oxford: Basil Blackwell.
Geckeler, Horst. 1971. Strukturelle Semantik und Wortfeldtheorie. Munich: Fink.
Geeraerts, Dirk. 1993. Vagueness’s puzzles, polysemy’s vagaries. Cognitive Linguistics
4:223–72.
Geertz, Clifford. 1973. The interpretation of cultures. New York: Basic Books.
Gentner, Dedre. 1983. Structure-mapping: a theoretical framework for analogy. Cognitive
Science 7:155–70.
1988. Metaphor as structure mapping: the relational shift. Child Development 59:47–59.
Gibbs, Raymond W., Jr. 1990. Psycholinguistic studies on the conceptual basis of idiomatic-
ity. Cognitive Linguistics 1:417–51.
1994. The poetics of mind: figurative thought, language and understanding. Cambridge:
Cambridge University Press.
Gibbs, Raymond W. and J. O’ Brien. 1990. Idioms and mental imagery: the metaphorical
motivation of idiomatic meaning. Cognition 36:35–68.
Glucksberg, Samuel. 2001. Understanding figurative language. Oxford: Oxford University
Press.
Goldberg, Adele E. 1995. Constructions: a construction grammar approach to argument
structure. Chicago: University of Chicago Press.
Goossens, Louis. 1990. Metaphtonymy: the interaction of metaphor and metonymy in
expressions of linguistic action. Cognitive Linguistics 1:323–40.
Grady, Joseph E. 1997. Theories are buildings revisited. Cognitive Linguistics
8:267–90.
1998. The “conduit metaphor” revisited: a reassessment of metaphors for communica-
tion. Bridging the gap: discourse and cognition, ed. Jean-Pierre Koenig, 205–18.
Stanford, Calif.: Center for the Study of Language and Information.
Grady, Joseph E., Todd Oakley and Seanna Coulson. 1999. Blending and metaphor.
Metaphor in cognitive linguistics, ed. Raymond W. Gibbs Jr. and Gerard J. Steen,
101–24. Amsterdam: John Benjamins.
Gundel, Jeannette K., Nancy Hedberg and Ron Zacharski. 1993. Cognitive status and the
form of referring expressions in discourse. Language 69:274–307.
Haegeman, Liliane. 1994. Introduction to government and binding theory (2nd ed.). Oxford:
Basil Blackwell.
Hahn, Ulrike and Nick Chater. 1997. Concepts and similarity. Knowledge, concepts and
categories, ed. Koen Lamberts and David Shanks, 43–92. Hove: Psychology Press.
336 References
Jespersen, Otto. 1942. A modern English grammar on historical principles, vol. 6. London:
Allen & Unwin.
Johnson, David E. and Paul M. Postal. 1980. Arc pair grammar. Guildford, N. J.: Princeton
University Press.
Johnson, Mark. 1987. The body in the mind. Chicago: University of Chicago Press.
1993. Moral imagination: implications of cognitive science for ethics. Chicago: Univer-
sity of Chicago Press.
Jolicoeur, Pierre S., Martin A. Gluck and Stephen M. Kosslyn. 1984. Pictures and names:
making the connection. Cognitive Psychology 19:31–53.
Kant, Immanuel. 1790 [1952]. The critique of judgement, trans. J. C. Meredith. Oxford:
Clarendon Press. (Kritik der Urteilskraft. Berlin: Bey Lagarde und Friederich.)
Katz, Jerrold J. and Jerry A. Fodor. 1963. The structure of a semantic theory. Language
39:170–210. (Reprinted in The structure of language: readings in the philosophy of
language, ed. Jerry A. Fodor and Jerrold J. Katz, 479–518. Englewood Cliffs, N. J.:
Prentice-Hall, 1964.)
Kay, Paul. 1997. Construction grammar feature structures (revised). Available at
http://www.icsi.berkeley.edu/∼kay/bcg/FSrev.html.
2002. English subjectless tagged sentences. Language 78:453–81.
Kay, Paul and Charles J. Fillmore. 1999. Grammatical constructions and linguistic gener-
alizations: the What’s X doing Y? construction. Language 75:1–33.
Kimenyi, Alexandre. 1980. A relational grammar of Kinyarwanda. (University of California
Publications in Linguistics, 91.) Berkeley: University of California Press.
Kleiber, Georges. 1996. Cognition, sémantique et facettes: une “histoire” de livres et
de . . . romans. Les formes du sens, ed. Georges Kleiber and Martin Riegel, 319–31.
Louvain: Duculot.
Koffka, Kurt. 1935. Principles of Gestalt psychology. New York: Harcourt, Brace & World.
Köpcke, Klaus-Michael. 1988. Schemas in German plural formation. Lingua 74:303–35.
1998. The acquisition of plural marking in English and German revisited: schemata versus
rules. Journal of Child Language 25:293–319.
Kövecses, Zoltán. 2002. Metaphor: a practical introduction. Oxford: Oxford University
Press.
Kövecses, Zoltán and Günter Radden. 1998. Metonymy: developing a cognitive linguistic
view. Cognitive Linguistics 9:37–77.
Kuno, Susumu. 1987. Functional syntax: anaphora, discourse and empathy. Chicago: Uni-
versity of Chicago Press.
Kuno, Susumu and Etsuko Kaburaki. 1977. Empathy and syntax. Linguistic Inquiry
8:627–72.
Labov, William. 1973. The boundaries of words and their meanings. New ways of analyzing
variation in English, ed. Joshua Fishman, 340–73. Washington D. C.: Georgetown
University Press.
Lakoff, George. 1973. Hedges: a study in meaning criteria and the logic of fuzzy concepts.
Journal of Philosophical Logic 2:458–508.
1987. Women, fire and dangerous things: what categories reveal about the mind. Chicago:
University of Chicago Press.
1990. The Invariance Hypothesis: Is abstract reason based on image-schemas? Cognitive
Linguistics 1:39–74.
1993. The contemporary theory of metaphor. Metaphor and thought, ed. Andrew Ortony,
202–51. Cambridge: Cambridge University Press.
338 References
Poplack, Shana. 1992. The inherent variability of the French subjunctive. Theoretical anal-
yses in Romance linguistics, ed. Christiane Laeufer and Terrell A. Morgan, 235–63.
Amsterdam: John Benjamins.
1996. The sociolinguistic dynamics of apparent convergence. Towards a social science
of language: papers in honor of William Labov, vol. 1: variation and change in
language and society, ed. Gregory Guy, John Baugh and Deborah Schiffrin, 295–
309. Amsterdam: John Benjamins.
Prasada, Sandeep and Steven Pinker. 1993. Generalization of regular and irregular morpho-
logical patterns. Language and Cognitive Processes 8:1–56.
Prince, Ellen F. 1978. A comparison of WH-clefts and it-clefts in discourse. Language
54:883–906.
1981a. Toward a taxonomy of given-new information. Radical pragmatics, ed. Peter
Cole, 223–56. New York: Academic Press.
1981b. Topicalization, Focus-movement and Yiddish-movement: a pragmatic differentia-
tion. Proceedings of the Seventh Annual Meeting of the Berkeley Linguistics Society,
ed. Danny K. Alford et al., 249–64. Berkeley: Berkeley Linguistics Society.
Pulman, Steven G. 1983. Word meaning and belief. London: Croom Helm.
Pustejovsky, James. 1995. The generative lexicon. Cambridge, Mass.: MIT Press.
Quine, Willard van Orman. 1951 [1961]. Two dogmas of empiricism. From a logical point
of view (2nd ed.), 20–46. New York: Harper. (Originally published in Philosophical
Review 60:20–43, 1951.)
Radden, Günter. 2000. How metonymic are metaphors? Metaphor and metonymy at the
crossroads: a cognitive perspective, ed. Antonio Barcelona, 93–108. Berlin: Mouton
de Gruyter.
Reddy, Michael J. 1979 [1993]. The conduit metaphor – a case of frame conflict in our
language about language. Metaphor and thought (2nd ed.), ed. Andrew Ortony,
164–201. Cambridge: Cambridge University Press.
Reinhart, Tanya. 1984. Principles of gestalt perception in the temporal organization of
narrative texts. Linguistics 22:779–809.
Richards, I. A. 1936. The philosophy of rhetoric. London: Oxford University Press.
Ricoeur, Paul. 1978. The metaphorical process as cognition, imagination and feeling. On
metaphor, ed. Sheldon Sacks, 141–57. Chicago: University of Chicago Press.
Riemer, Nick. 2001. Remetonymizing metaphor: hypercategories in semantic extension.
Cognitive Linguistics 12:379–401.
Rosch, Eleanor H. 1973. Natural categories. Cognitive Psychology 4:328–50.
1978. Principles of categorization. Cognition and categorization, ed. Eleanor Rosch and
Barbara Lloyd, 27–48. Hillsdale, N. J.: Lawrence Erlbaum Associates.
Rosch, Eleanor H. and Carolyn B. Mervis. 1975. Family resemblances: studies in the internal
structure of categories. Cognitive Psychology 7:573–605.
Rosch, Eleanor H., Carolyn B. Mervis, Wayne Gray, David Johnson and Penny
Boyes-Braem. 1976. Basic objects in natural categories. Cognitive Psychology
8:382–439.
Ross, John R. 1967. Constraints on variables in syntax. Ph.D. dissertation, Massachusetts
Institute of Technology.
Rubino, Rejane B. and Julian M. Pine. 1998. Subject-verb agreement in Brazilian
Portuguese: what low error rates hide. Journal of Child Language 25:35–59.
Rudes, Blair A. 1980. On the nature of verbal suppletion. Linguistics 18:655–76.
References 341
Sadock, Jerrold M. and Arnold Zwicky. 1985. Speech act distinctions in syntax. Language
typology and syntactic description, vol. 1: clause structure, ed. Timothy Shopen,
155–96. Cambridge: Cambridge University Press.
Schank, Roger C. and Robert P. Abelson. 1977. Scripts, plans, goals and understanding.
Hillsdale, N. J.: Lawrence Erlbaum Associates.
Searle, John. 1979. Literal meaning. Expression and meaning, ed. John Searle, 117–36.
Cambridge: Cambridge University Press.
Seidenberg, Mark S., Michael K. Tanenhaus, James M. Leman and Marie Bienkowski. 1982.
Automatic access of meanings of ambiguous words in context: some limitations of
knowledge-based processing. Cognitive Psychology 14:489–532.
Slobin, Dan I. 1991. Learning to think for speaking: native language, cognition, and rhetor-
ical style. Pragmatics 1:7–26.
Smith, Edward E., Edward J. Shoben and Lance J. Rips. 1974. Structure and process in
semantic memory: a featural model for semantic decisions. Psychological Review
81:214–41.
Smith, Linda B. and Larissa K. Samuelson. 1997. Perceiving and remembering: category
stability, variability and development. Knowledge, concepts and categories, ed. Koen
Lamberts and David Shanks, 161–95. Hove: Psychology Press.
Sperber, Dan and Deirdre Wilson. 1986. Relevance: communication and cognition. Oxford:
Basil Blackwell.
Stassen, Leon. 1997. Intransitive predication. Oxford: Oxford University Press.
Stemberger, Joseph P. and Brian MacWhinney. 1988. Are inflected forms stored in the
lexicon? Theoretical morphology, ed. Michael Hammond and Michael Noonan,
101–16. San Diego, Calif.: Academic Press.
Stern, Joseph. 2000. Metaphor in context. Cambridge, Mass.: MIT Press.
Sweetser, Eve. 1990. From etymology to pragmatics: metaphorical and cultural aspects of
semantic structure. Cambridge: Cambridge University Press.
Swinney, David A. 1979. Lexical access during sentence comprehension: (re)con-
sideration of context effects. Journal of Verbal Learning and Verbal Behavior
18:645–69.
1982. The structure and time-course of information interaction during speech compre-
hension: lexical segmentation, access and interpretation. Perspectives on mental
representation: experimental and theoretical studies of cognitive processes and ca-
pacities, ed. Jacques Mehler, Edward C. T. Walker and Merrill Garrett, 151–67.
Hillsdale, N. J.: Lawrence Erlbaum Associates.
Sylestine, Cora, Heather K. Hardy and Timothy Montler. 1993. Dictionary of the Alabama
language. Austin: University of Texas Press.
Talmy, Leonard. 1972. Semantic structures in English and Atsugewi. Ph.D. dissertation,
Department of Linguistics, University of California, Berkeley.
1976. Semantic causative types. The grammar of causative constructions (Syntax and
Semantics, 6), ed. Masayoshi Shibatani, 43–116. New York: Academic Press.
1977. Rubber sheet cognition in language. Papers from the Thirteenth Regional Meeting,
Chicago Linguistic Society, ed. Woodford A. Beach et al., 612–28. Chicago: Chicago
Linguistic Society.
1978. The relation of grammar to cognition: a synopsis. Proceedings of TINLAP-2:
theoretical issues in natural language processing, ed. David Waltz, 14–24. Urbana:
University of Illinois Coordinated Science Laboratory.
342 References
1983. How language structures space. Spatial orientation: theory, research and appli-
cation, ed. Herbert L. Pick, Jr. and Linda P. Acredolo, 225–82. New York: Plenum
Press.
1988a. The relation of grammar to cognition. Topics in cognitive linguistics, ed. Brygida
Rudzka-Ostyn, 165–205. Amsterdam: John Benjamins.
1988b. Force dynamics in language and cognition. Cognitive Science 12.49–100.
2000. Toward a cognitive semantics, vol. 1: concept structuring systems. Cambridge,
Mass.: MIT Press.
Taylor, John R. 1989 [1997]. Linguistic categorization: prototypes in linguistic theory (2nd
ed.). Oxford: Oxford University Press.
2002. Cognitive grammar. Oxford: Oxford University Press.
Togia, Pagona. 1996. Antonyms in English and Modern Greek: a cognitive approach. Un-
published Ph.D. thesis, University of Manchester.
Tomasello, Michael. 1992. First verbs: a case study of early grammatical development.
Cambridge: Cambridge University Press.
2000. Do young children have adult syntactic competence? Cognition 74:209–53.
2003. Constructing a language: a usage-based theory of language acquisition.
Cambridge, Mass.: Harvard University Press.
Tomasello, Michael, Nameera Akhtar, Kelly Dodson and Laura Rekau. 1997. Differential
productivity in young children’s use of nouns and verbs. Journal of Child Language
24:373–87.
Tsukiashi, Ayumi. 1997. A usage-based analysis of the Japanese passive construction. M.A.
dissertation, University of Manchester.
Tuggy, David. 1993. Ambiguity, polysemy and vagueness. Cognitive Linguistics 4:273–90.
Turner, Mark. 1987. Death is the mother of beauty: mind, metaphor, criticism. Chicago:
University of Chicago Press.
1990. Aspects of the invariance hypothesis. Cognitive Linguistics 1:247–55.
2001. Cognitive dimensions of social science. Oxford: Oxford University Press.
Ungerer, Friederich and Hans-Jürgen Schmid. 1996. An introduction to cognitive linguistics.
London: Longman.
Vallduvı́, Enric. 1992. The informational component. New York: Garland.
Wertheimer, Max. 1923 [1950]. Laws of organization in perceptual forms (abridged and
translated by Willis D. Ellis). A source book of Gestalt psychology, ed. Willis D.
Ellis, 71–88. New York: Humanities Press. (Untersuchungen zur Lehre von der
Gestalt. Psychologische Forschung 4:301–50.)
Whittlesea, Bruce W. A. 1997. The representation of general and particular knowl-
edge. Knowledge, concepts and categories, ed. Koen Lamberts and David Shanks,
335–70. Hove: Psychology Press.
Wierzbicka, Anna. 1980. Lingua mentalis: the semantics of natural language. New York:
Academic Press.
1985. Oats and wheat: the fallacy of arbitrariness. Iconicity in syntax, ed. John Haiman,
311–42. Amsterdam: John Benjamins.
1987. Boys will be boys. Language 63:95–114.
1988. The semantics of grammar. Amsterdam: John Benjamins.
1996. Semantics: primes and universals. Oxford: Oxford University Press.
Wilson, Deirdre. 1975. Presuppositions and non-truth-conditional semantics. New York:
Academic Press.
References 343
Aaron, David 201, 203 Clausner, Timothy C. 22–23, 44–45, 68, 198–99,
Abelson, Robert P. 7, 17 318–19
Akhtar, Nameera 323 Coleman, Linda 31–32
Akmajian, Adrian 245 Collins, Allan M. 76
Allan, Keith 176 Cornwell, Patricia 193, 210, 213, 218, 219, 221
Andersen, Henning 306 Coseriu, Eugenio 141
Ariel, Mira 51 Coulson, Seana 38, 207–9, 210
Armstrong, Sharon Lee 88 Croft, William 4, 22–23, 26, 44–45, 57–58, 59,
Aronoff, Mark 226 64, 66, 68, 70–71, 73, 97, 102, 198–99, 242,
Atkins, Beryl T. 11–12 257, 274, 278, 279, 283, 285–87, 289, 294,
Atran, Scott 86 319, 321–23
Cruse, D. Alan 4, 79, 80, 83, 84, 109, 116, 117,
Baldwin, Gillian 323 120, 130, 136, 141, 142, 143, 147–50, 152,
Barlow, Michael 291 154, 156, 158, 159, 160–61, 165, 173–74,
Barsalou, Lawrence W. 87, 92, 96, 278 182, 187
Bates, Elizabeth A. 329
Becker, Joseph 231 Daugherty, K. 295, 300
Berlin, Brent 81, 82, 83 Deane, Paul D. 109, 315
Bienkowski, Marie 303 DeLancey, Scott 62
Birner, Betty J. 242–43 Denison, David 310, 320
Black, Max 203, 206 Dodson, Kelly 323
Bolinger, Dwight 85, 277 Dreyfus, Hubert L. 59
Boyes-Braem, Penny 82, 84
Braine, Martin D. S. 323 Eco, Umberto 207
Breedlove, Dennis E. 82, 83 Elman, Jeffrey L. 300, 307, 329
Bresnan, Joan 229
Brewer, Mary Alexandra 305 Fauconnier, Gilles 33–39, 71, 123, 207, 328
Brinkmann, U. 294, 299–300 Feyaerts, Kurt 216
Brown, Cecil H. 86, 87, 162 Fillmore, Charles J. 7, 8–14, 15, 16, 17–20, 28,
Brown, Roger 77 30, 60, 76, 95, 231–40, 248–49, 255, 257,
Bybee, Joan L. 291, 292–94, 296–306, 310–12, 259, 263, 266–71, 276, 279
313 Flickinger, Daniel 275
Fodor, Jerry A. 76
Carling, Christine 97
Chafe, Wallace 46, 50 Gadamer, Hans-Georg 21
Chater, Nick 76 Gathercole, Virginia C. Mueller 324
Chomsky, Noam 226, 227–29, 318 Gazdar, Gerald 229
Clahsen, Harald 299 Geckeler, Horst 141
Clark, Eve V. 47 Geeraerts, Dirk 109
Clark, Herbert H. 13, 18, 47, 60, 102–3 Geertz, Clifford 21
344
Author index 345
Bold page numbers indicate definitions or major discussions. Category values (‘future’) are found
under category entries (‘tense’).
347
348 Subject index
auxiliary 228, 238, 264, 265, 268, 271, 276, 280, cognitive linguistics 1, 40, 42, 45, 105, 225, 291,
286, 310, 311, 315, 317, 324 325, 327, 328–29
cognitive psychology 3, 7, 17, 28, 30, 45, 46, 54,
background 59 75, 86, 328, 329
background assumptions 29–30 collocations 12, 18, 249–50, 252
Bambara 289 common ground 60–61, 102–3
Bantu 317 communication 19, 74, 99, 103, 193, 291, 326,
base 7, 15–16, 19, 25, 132 329
based-on link 275 comparative 177–79, 183, 186, 228
Being-in-the-world 58–59 quantified comparative 178–79, 183
binarity 164–65, 166 comparison 44, 54–58, 68
biological kinds 17, 86 complement 267, 281, 325
biscalar antonym/system 170, 172, 174, 181–85 complement constructions 280, 316
disjunct 170, 172, 182 complementaries 165–66, 167–69, 185,
equipollent 170, 172, 174, 181–83, 188, 191 188
overlapping 170, 172, 183–85, 188, 190 complementizer 280
parallel 170, 182–83 componential model (of a grammar) 225–29,
blending 38–39, 193, 213–15, 216, 221 231, 232, 237, 245, 247, 248, 255,
see also space, mental 258–59, 263
Blending Theory (BT) 38–39, 203, 207–9, 210, component, semantic (of a construction)
328 138–40, 260, 286
‘Blessings-Wishes-Curses’ construction 271 compositionality, semantic 105, 120, 177, 179,
borrowing 296, 299, 300, 312 249–54
boundary (of category/sense) 75, 76, 89–91, compound 31
93–95, 97, 102, 104, 105, 109–15, 122, comprehension 99, 100, 278, 285, 307
143, 146, 151, 153, 155–56, 159, 168, concepts 7, 14–15, 24–27, 30, 37, 47, 48, 88,
322 92–93
see also fuzziness (of boundaries) generic concepts 74–75
bounded/unbounded 64, 70, 71 individual concepts 74–75
bridging 13 Conceptual Metaphor Theory (CMT) 194–204,
207, 209
calibration 171–72, 178, 179, 180, 184, 191–92 conceptual space 109, 288, 321–23, 327
case (marking) 261, 286, 319 conceptual structure 2, 3, 15, 30, 34, 39, 46, 197,
Categorial Grammar 229 328
categorical 61 conceptualization see construal
categorization/category 3, 17, 46, 53, 54–55, conditional 228
74–106, 282–83, 285, 326 conjunction 237–38
ad hoc category 92 connection 303–4
see also boundary (of category); classical see also phonological connection; semantic
model; dynamic construal model; frame, connection; symbolic connection
semantic; levels (categories); prototype consciousness 46, 75
model constituency (syntactic) 261, 311
Caused Motion construction 312 constitution see Gestalt
change see language change constraints (conceptual) 100–1, 101–3, 109
circumstantial phrase 228 constraints (syntactic) 225, 315, 318
classical model (categorization) 76–77 construal 1, 19, 28, 40–69, 75, 79, 80, 93–98,
see also categorization 103–4, 109–10, 122, 127, 138, 140, 144,
clause 273 145, 150, 151, 153, 155, 158, 160, 161,
coercion 43 164–65, 167–68, 169, 182, 185–92, 216,
cognitive abilities/capacities 2, 3, 45 218, 221, 279–80, 328, 329
Cognitive Grammar 72, 257, 266, 278–83, 285, absolute construal (antonyms) 177, 178,
287, 288 179–81, 185–89
Subject index 349
default construal 71–72, 83, 102, 103–4, 126, declarative sentence construction 264, 265, 314,
140, 144, 146–47, 158, 159, 166, 168, 319–21
185, 189, 190, 191, 216 default construal see construal, default
hybrid construal (antonyms) 186–87 default schema see schema, default
relative construal (antonyms) 175, 177, default specificity 127, 129–33, 134, 135, 158
179–81, 184, 185–89 degree of membership (DOM) 79–81
construal operations 40–73 deictic center 60
interaction of 69–70 deixis 10–11, 44, 46, 58, 59–62, 63
construction grammar 4, 76, 225, 227, 229, 231, epistemic deixis 46, 60–61, 63
240, 245, 247–48, 252, 255–56, 257–90, demonstrative 51, 235
302, 313, 317, 326, 329 dependency 261
Construction Grammar (Flllmore, Kay et al.) dependent (grammatical structure) 282
257, 266–72, 279, 280, 282, 285, 288, derivation, grammatical 40, 41–42
291 determiner 324
Construction Grammar (Lakoff, Goldberg) 257, diachrony 111, 305
266, 272–78, 283, 288 dictionary meaning 30
constructions, grammatical 4, 8, 14, 34, 41–42, dimension 25, 69
53, 73, 177, 227–29, 236–49, 251–56, direct object see object
257–90, 295, 302, 308, 313, 319, discourse function see information structure
321–27 disjunction 34
container/containment image schema 80, 89, ditransitive construction 264–65, 273–74, 312
104, 142, 151, 201 domain, semantic 15–16, 17–32, 39, 44, 47, 65,
contextual constraints/pressure 102–3, 109, 68–69, 70, 79, 131–32, 164–66, 167–69,
122–23, 127–28, 130, 131, 134, 135, 172, 194–216, 221
136, 150, 159, 164, 165, 182, 193, 204, abstract domain 24
221 basic domain 24, 25, 26
contextual modulation 128, 129, 130, 135, image-schematic domains 68–69
140 source domain 55, 195–204, 207, 210, 215,
contextual pressure see contextual constraints 221
contiguity (syntactic) 286 target domain 55, 195–204, 207, 210, 221
controller (of agreement) 286 domain, social
convention, conventionality 31, 43, 72–73, 156, domain matrix 25, 27, 31, 47, 69, 122, 132, 216
195–98, 199, 203–4, 230, 231, 249–52, domain structure 26
258, 279, 280 dominion 46, 51–52
conventional constraints 102, 103–4, 109, dual-processing model 294, 299
111–12, 114, 117, 131, 135, 139, 144, durative adverbial 246
159, 161, 164, 166, 193, 209, 216 dynamic construal model (categorization) 4, 75,
conventional imagery 72 92–104, 141
conventional universalist position 73 see also categorization
conversational implicature see implicature
conversion 43 economy 74
conversive 166 elaboration 281–82
coordinate construction 240 elaboration site (e-site) 281
copula 253, 319, 321–22 element (syntactic) 260, 264, 285, 286, 287
Cora 280 embodiment 44
core (intensional) 150 empathy 46, 61, 62, 63
correspondence see metaphorical emphatic negative imperative construction 271
correspondence; symbolic encyclopedic knowledge/meaning 30, 86, 148,
correspondence 196, 204, 208
countability (count/mass) 71, 280 English (Modern) 14,21, 24, 41, 42, 64, 72,
see also noun 90–91, 110, 178, 181, 228, 235, 253,
counterfactual 36, 38, 39 280, 293, 310, 314, 316, 320–22
350 Subject index
English (Modern) (cont.) type frequency 296–300, 301, 307, 308, 309,
Middle English 310 312, 327
Old English 235, 294, 300 full-entry model 276–78
entailment 13, 104, 143, 145–46 functionalism (linguistics) 242, 329
entity/interconnection 44, 67–68 fuzziness (of boundaries) 77, 91, 94, 95
entrenchment 111–12, 131–33, 135, 136, 139,
292–95, 296, 297, 304, 305, 308–11, 327 games 33
epistemic correspondences 196–97, 201 gang effects 294–95
epistemic deixis see deixis gender (grammatical) 300
equality/inequality constructions 179–80 Generalized Phrase Structure Grammar 229
ergativity 280, 316 generative grammar 1, 2, 225–29, 259–60, 261,
split ergativity 62 263, 292, 293, 294, 302, 313, 317, 328,
e-site see elaboration site 329
evaluative terms 18–19 geometric structure 63, 64–65, 70
event 57–58, 70 German 10, 20, 21, 53, 183–84, 299, 300
evolution of humankind 328 Gestalt 46, 63–69, 75, 100, 101, 115, 116, 175,
exclamative constructions 241, 247, 258 286
experience 19, 24, 28, 44, 45, 54, 63, 68, 69, Gestalt psychology 56, 63, 329
71–73, 74, 101, 172, 195, 201, 203–4, goodness of exemplar (GOE) 77–79, 80–81, 92,
326 153, 166
expert systems 86 Government and Binding theory 229
explicature 100 gradability 71
exposure 177–81 graded centrality 3, 32, 75, 77–81, 88, 91
extraction constructions 237, 315–16, 317 grammar/grammatical knowledge 1, 3, 12,
106, 225–27, 229, 231, 247, 254,
facets 47–48, 101, 116–26, 131, 132, 137, 138, 255–56, 257, 263–65, 271, 278, 279,
140, 216, 220 287, 288, 291, 292, 296, 306, 326–27,
familiarity 78 328, 329
family resemblance 78, 82, 85 see also Arc-Pair Grammar; Categorial
feature Grammar; componential model (of a
grammatical 266–70, 284, 285 grammar); construction grammar;
semantic 7, 8–10, 76, 78, 87, 88, 91, 100, 148, Construction Grammar (Fillmore, Kay et
150 al.), Construction Grammar (Lakoff,
feature structure 266–67, 269, 271 Goldberg); generative grammar;
fictional situations 33 Generalized Phrase Structure Grammar;
fictive motion 46, 53, 54 Government and Binding theory;
figurative meaning 193, 230, 251 Head-driven Phrase Structure Grammar;
figure/ground 44, 46, 56–58, 59, 62, 71, 101, 281 Lexical-Functional Grammar;
filler 268 Minimalist theory; organization,
focal adjustments 43–44, 47, 58, 59 grammatical; Radical Construction
focal orientation 149–50 Grammar; Relational Grammar;
focus antipassive 316 representation, grammatical; rules;
focus constructions 238, 240, 247, 315–16 schemas; Semiotic Grammar;
folk classification 86 usage-based model; Word Grammar
force dynamics 43, 46, 66–67, 69 grammaticalization 63, 317–18
foreground 59 granularity 52
frame, semantic 8–22, 34, 37, 39, 46, 47, 53, 55, Greek, Modern 162, 172, 178, 180–81, 182
87, 91–92, 95–96, 167, 272 ground see figure/ground
French 20, 72, 90–91, 136, 178, 180–81, 311
frequency 78, 80, 133, 292, 304, 305, 309 have a X constructions 243–44, 247, 248
token frequency 292–95, 301, 306, 307, 308, head 267, 268, 270, 282, 282
309–10, 324–25 hearer 100, 286, 318, 329
Subject index 351
Head-driven Phrase Structure Grammar 259, inflections, grammatical 40, 41–42, 293–307,
266, 279 324–25
Hebrew, Modern 72 information structure 61, 226, 235, 242
historical linguistics 1 inherentness (in antonyms) 184–85, 189
see also change, language; diachrony inheritance 76, 270–72, 273, 274, 275–78
holonym 160 complete inheritance 270–71, 278, 308
homonymy 100, 111, 217, 303 default inheritance see inheritance, normal
homophony 303 multiple inheritance 264, 276–77
hyperonym 84, 114, 117, 120, 121, 127, 128, 129, normal inheritance 275–76
130, 131–32, 133, 134, 135, 148, 149, see also full-entry model
158 innate capacity for language 2–3
hyponym 14, 117, 120, 121, 127, 128, 132, 143, instance (of a construction) 259
144, 146, 147, 148, 152, 175 instrument 316
hyponymy 3, 7, 104, 141–50, 159, 162 integration (of components of a meaning) 125–26
intention 88
iconicity 175, 286 interactive activation network 307
ideal 80 interconnection see entity/interconnection
Idealized Cognitive Model (ICM) 28–32, 92, 95 interface (grammatical) 228, 229
cluster ICM 31, 92, 95 interpretation 98–100, 101, 109–10
identity 36 interrogative constructions 265, 310, 314, 319
idiom 199, 205, 225, 2300–37, 270 see also questions
decoding idiom 231, 232, 235 intransitive construction 234, 259, 260–62, 264,
encoding idiom 231–32, 235, 250 268, 283–85, 287
extragrammatical idiom 233, 235, 236 Invariance Hypothesis 201–2
formal idiom 233, 235 inversion constructions 242–43
see also idiom, schematic irregularity see regularity
grammatical idiom 232, 235 It-cleft construction 237, 242, 265, 315
schematic idiom 234, 235, 236–37, 241, 243, iteration 70
244, 247, 262–63
substantive idiom 233, 234, 235, 236, 237, Japanese 316–17
247, 253, 262–63 Javanese 21
see also pragmatic point judgement see comparison
idiomatic phrase 232, 235, 252, 253
idiomatically combining expression 232–33, K’iche’ Mayan 254, 312, 316, 317
235, 236, 249–53, 263 Kinyarwanda 317
illocutionary force 318–20 knowledge of language 1, 2, 3–4, 225, 328, 329
image 44
image nouns 33 landmark 58
see also picture nouns language 2, 71–73, 99, 328–29
image schemas 44–46, 62, 64–65, 68, 80, 104, language change 291, 292, 293–94, 298, 307,
167, 169, 172, 201–4 325–26
imaging systems 43 language use 1, 2, 3–4, 278, 307, 326, 328, 329
immediate scope, see scope of predication latency 134–37
imperative constructions 319–22 learning 74, 78, 325
implication 8 see also acquisition
see also entailment left isolation (LI) construction 271
implicature 10, 185, 268 let alone construction 237–40, 247, 248
incompatibility 104, 117, 126–27, 133, 141, 145, levels (categories) 75, 82–87, 96–97, 130
147, 152, 162 basic 82, 83–84, 96, 97, 130, 148, 175
individuation 64, 70 generic see level, basic
induction 4, 323–25, 327 subordinate 82, 84, 85–86, 96, 130
infinitive 247 superordinate 82, 84–85, 96, 97, 130, 175
352 Subject index
quantifier 48 scale 22–23, 46, 65, 69, 104, 166, 167, 169–92,
quasimodals 317–18 201
questions 311, 314, 315–16 scale-committedness see partiality,
How X is it? 173–74, 176, 179, 180, 183, 184 scale-partiality
Is it X? 181–82 scale-impartiality see partiality, scale-partiality
What is its NOM? 173, 176, 179–82, 184 scale-schema 173
see also interrogative constructions absolute scale-schema 173, 175
relative scale-schema 173
radial category structure 272 see also construal, absolute; construal, hybrid;
Radical Construction Grammar 257, 266, construal, relative
283–89, 321 scanning 44, 46, 53–54, 67, 280
range congruence 160 sequential scanning 46, 53–54, 67, 72, 280
reality 101–2 summary scanning 46, 53–54, 67, 280
reductionism 1 schema 295–302, 207, 308–11, 314, 317, 327
reductionist model 268, 272, 284–85 default schema 300, 312
reference 35–36, 54 open schema 299–300
reference point 51 product-oriented schema 301–2, 307, 313–18,
referential opacity 35–36 327
referential/attributive 36 source-oriented schema 301, 313, 327
regular syntactic expressions 232, 233, 234, 236, schematic relation 24, 26–27
252 schematic systems 43
regularity/irregularity (grammatical) 292–300, schematicity 175, 198–201, 202, 253, 254, 263,
303, 311, 313, 318 265, 309, 312
relation see semantic relation; symbolic relation; schematization 4, 44, 52–53
syntactic relation scope of attention 46
Relational Grammar 229 scope of predication 23–24, 46, 154
relationality 46, 48, 58, 67–68, 195, 269, 281 script 8, 17
see also entity/interconnection search domains 50–51
relative clause 315–16 selection 44, 46, 47–50, 65
relativity (semantic/linguistic) 72–73 selectional restrictions 249
relevance 304, 318–19, 322, 327 semantic connection 303, 304, 307, 318, 321,
see also similarity 327
Relevance Theory 100 semantic interpretation rules 232–33, 237, 245,
representation, grammatical 2, 257 249, 252, 253, 254, 257
representation, knowledge 291, 326 Semantic Map Connectivity Hypothesis 322
resultative construction 246, 247, 248, 249, 277 semantic map model 283, 288, 321–22, 327
reversive 165–66, 169 semantic pole 279, 286
right-dislocation construction 242–43 semantic relation 261, 286
roles (in mental spaces) 34–36 semantics 1, 2, 4, 12–13, 34, 40, 226, 297, 301,
roles, syntactic see syntactic roles 328
rules 225, 229, 237, 246, 292, 294, 295–96, frame semantics 8–32
300–2, 313–18, 327 lexical semantics 3
Russian 42–43, 280 semantics of understanding 4, 8, 13, 99
structural semantics 7, 76
salience 47 truth-conditional semantics 1, 2, 7, 8, 12, 33,
see also attention 38, 40–43, 64, 328, 329
sanction 55 see also lexical field theory
scalar adjustment 46, 51–53, 64, 65 semelfactive 43
qualitative 46, 52–53, 64 Semiotic Grammar 259
quantitative 46, 51–52, 64 sense unit 109
scalar model 239, 240 full sense unit 112, 115
Subject index 355