Language Mind and Brain Dabrowska Libros
Language Mind and Brain Dabrowska Libros
Language Mind and Brain Dabrowska Libros
E WA DA
B ROWSK A
A CIP record for this book is available from the British Library
Contents
Chapter 1 Introduction 1
Chapter 7 Words 81
1. The semantics of locative terms 83
1.1 Conventional construal 85
1.2 Polysemy 89
1.3 Cross-linguistic variation 91
2. The acquisition of locative terms 100
2.1 Warlpiri 101
2.2 Tzotzil 101
2.3 Japanese 102
2.4 Korean 103
2.5 General discussion 104
3. Innate structure, yes but of what kind? 105
3.1 Semantic primitives 106
3.2 Perceptual primitives 107
4. Lexical learning in a constrained connectionist network 111
5. Conclusion 112
Contents vii
Bibliography 230
Index 259
DABROWSKA pre Make up 26/3/04 4:32 pm Page viii
List of figures
List of tables
Acknowledgements
1 Introduction
Introduction 3
Part I
The Basic Specications
DABROWSKA text Make up 26/3/04 4:34 pm Page 6
DABROWSKA text Make up 26/3/04 4:34 pm Page 7
If you were to hear this sentence instead of reading it, what would reach
your ears would be a series of air vibrations; yet almost instantaneously you
would be able to work out its meaning: a woman picks up a piece of tissue,
covers her nostrils with it, and clears her nose by vigorously expelling air
through it. To arrive at this interpretation, you must have performed a number
of operations: identied the words; accessed their relevant senses; determined
the grammatical relationships between them; and integrated lexical, syntactic,
and contextual information with real-world knowledge to construct an inter-
pretation of the sentence in context. Each of these operations involves the
integration of multiple and only partially reliable cues.
Let us begin with the rst stage: word recognition. The major problems
here are variability of the speech signal and segmentation. The acoustic prop-
erties of speech segments vary according to the context in which they occur:
the /b/ sounds in blew, bet, bear and rob, for example, are phonetically quite
distinct. Furthermore, since the acoustic properties of speech depend on the
shape of the vocal tract and everybodys vocal tract is shaped slightly dier-
ently, what we perceive as the same sound is actually quite dierent when
produced by dierent speakers, or even by the same speaker on dierent occa-
sions. The segmentation problem is a consequence of the fact that speech is
DABROWSKA text Make up 26/3/04 4:34 pm Page 8
continuous (speakers dont normally pause between words), and hence word
boundaries are not given in the acoustic input but must be inferred by the
listener.
Because of the lack of invariance and the segmentation problem, even this
low-level processing poses a considerable challenge, which is perhaps best
appreciated in the context of current developments in automatic speech rec-
ognition. Despite decades of intensive research, we are only beginning to see
computer applications that are reasonably accurate at identifying words in
continuous speech. Most of the existing programs are trained to recognise the
voice of a particular speaker in a particular acoustic environment, with little
background noise; and although their performance is good enough to make
them commercially viable, they still make many errors.
Once the listener has segmented the speech signal and identied the words,
he or she must access the relevant sense of each word and determine how they
combine to form constituents. Both of these operations are fraught with di-
culty, since most words are ambiguous, and there is often more than one way
of constructing a well-formed syntactic structure from a given string of words.
Let us begin with lexical ambiguity. The rst word in our sentence, Tess, is
a proper name, and hence has no conventional linguistic meaning apart from
specifying that the referent is female and probably human. However, the next
word, blew, is multiply ambiguous. As the past tense form of the verb blow,
blew could have any of the following meanings:
(This is a selection of some of the most frequent meanings; any sizeable dic-
tionary will list many more.) The same phonological form, when spelt blue, is
also associated with the following meanings:
The next content word, the noun nose, can designate any of the following:
1. substance manufactured from wood bre used for writing, drawing etc.;
2. newspaper;
3. a set of examination questions;
4. essay;
5. to cover with paper;
and the same is true of tissue, which can refer to any of the following:
1. woven fabric;
2. a mass of cells, as in muscular tissue;
3. a small piece of thin soft absorbent paper;
4. web, as in a tissue of lies.
Finally, the determiner a, in addition to various specialised uses which will not
be discussed here (but see Quirk et al. 1985: 26597), can be used to indicate
that the speaker wishes:
To resolve these ambiguities, the listener must use grammatical, lexical, col-
locational, and pragmatic information. Let me illustrate this process using the
second word in our example, the phonological form /blu/. In the context of
our sentence, /blu/ is the past tense of the verb blow rather than the adjective
blue, since (1) there is no noun for it to modify and (2) it comes immediately
after the subject, and hence is likely to be the rst element of the verb phrase,
which is usually a verb or an auxiliary. The fact that the subject NP is animate
eectively eliminates senses (1), (6), (7) and (11), which normally co-occur
with subjects designating wind, a wind instrument, a tyre, and a fuse, respec-
tively. Senses (4), (5), (8), (9) and (10) are also ruled out for collocational
reasons: they normally take direct objects which designate a gaseous forma-
tion such as smoke rings (sense 4), a wind instrument (5), a tyre (8), money or
equivalent (9), or an opportunity (10). Our sentence could mean that she blew
into a nose-shaped piece of plastic, or that she sold her nose in order to buy a
paper tissue but such interpretations are extremely unlikely. Since noses
DABROWSKA text Make up 26/3/04 4:34 pm Page 11
usually come attached to faces, sense (2) is ruled out on the basis of our knowl-
edge of the world (although the sentence could, in theory, mean that she pro-
pelled a plastic mould of a nose, or a real nose which had been severed from
someones face, by blowing until it landed on a paper tissue either by exhal-
ing air from the lungs or by using a device such as a bellows). The direct object
her nose is incompatible with sense (3), which is used in intransitive construc-
tions or with an NP designating the substance that comes out of the lungs such
as air or smoke. This leaves (12), the nose-clearing sense, as the preferred inter-
pretation of the phonological form /blu/ in the context of our example.
Let us now briey look at syntactic processing. English is a fairly rigid SVO
language, so the listener can safely assume that the preverbal NP (Tess) is the
subject. If we consider just lexical category information, the sequence blew her
nose could be interpreted either as a verb followed by a direct object, where
her is a possessive determiner in the object NP (analogous to ate his lunch) or
a verb with two objects (analogous to gave him lunch). However, her cannot be
the object of the verb, for two reasons. First, blow is monotransitive (it only
allows one object), so if her were its object, there would be no way of attach-
ing the NP nose to the verb phrase. Secondly, since nose is a singular count
noun, it must be preceded by a determiner. Therefore, the entire expression her
nose must be the object of blew.
The next four words, on a paper tissue, clearly constitute a prepositional
phrase. To interpret the sentence, the listener must determine how it combines
with the preceding words. There are three possibilities here: the PP could be a
complement to the verb, an adjunct, or a modier of the noun. In the rst case,
the sentence would have the same structure as Tess put the milk in the fridge.
This possibility is ruled out by the fact that blew is monotransitive, and hence
takes only one complement, the direct object. In the second case, on a paper
tissue would specify either the location or the direction of Tesss doings, and
the syntactic structure would parallel that for Tess read the book on the train.
Finally, the PP could postmodify the noun nose. Imagine that Tess has two
noses, and has put one of them on a paper tissue. In this context the sentence
could mean that she blew the nose which was on the paper tissue, and would
be similar in structure to Tess sold the house in Canada.
Our example is a very simple sentence, and hence the amount of syntactic
ambiguity that the reader or listener must cope with is fairly modest.1 However,
syntactic ambiguity becomes a major problem when the processor is con-
fronted with longer, more complex sentences characteristic of written dis-
course. It is well known from the natural language processing literature that the
number of candidate parses increases exponentially with sentence length.
Consider the following example, borrowed from Martin et al. (1987). Their
grammar assigns three structural analyses to the sentence List the sales of prod-
ucts in 1973 (the prepositional phrase in 1973 can modify the verb list, the
expression sales of products, or just the noun products). If we add a single word
to the sentence, making it List the sales of products produced in 1973, the
DABROWSKA text Make up 26/3/04 4:34 pm Page 12
number of analyses increases to 10; and if we add with the products produced
in 1972 at the end of the last sentence, the number of alternative structural
descriptions (assuming the grammar developed by Martin et al.) is exactly 455.
Returning to Tess and her nose: to determine which of the two grammati-
cally acceptable readings identied above is the one that the speaker intended,
the listener must again use pragmatic information. Since noses are usually
attached to their owners and people dont normally have more than one, the
easiest way to specify which nose you mean is by identifying the owner by
means of a possessive modier; hence, a prepositional phrase after nose is
unlikely to be a postmodier. It must therefore be an adjunct specifying either
where Tess was when she blew her nose or where what she blew out of her nose
ended up. Stating that Tess was on a paper tissue is not a very good way of
locating her in space, since tissues are rather small in relation to a person, and
do not have a xed location. On the other hand, paper tissues are excellent for
blowing your nose on, so by far the most likely interpretation of the sentence
is that Tess used the tissue to clear her nose.
The listener must also decide whether her refers to Tess or to another female
not mentioned in the sentence. English grammar allows both possibilities, so
again, the correct interpretation depends on pragmatic knowledge. Since it is
impossible to clear another persons nose by exhaling air on a paper tissue,2 we
must conclude that her refers to Tess.
I have deliberately chosen to illustrate the complexity of the processing task
with a very simple (if somewhat insalubrious) sentence: one with no embed-
ding, no dislocated constituents, no long-distance dependencies. The example
illustrates rather nicely the amount of ambiguity that listeners must cope with,
as well as the fact that successful comprehension requires the integration of
lexical, syntactic and real-world knowledge. The discussion so far, however,
has completely ignored another important feature of ordinary language use
which introduces a whole new dimension of complexity: the fact that much of
the linguistic input that reaches our ears is degraded.
First, the acoustic signal itself is often distorted or partially inaudible
because of either background noise, or various modications that speakers
make in rapid speech, or both. In a now-classic study, Pollack and Pickett
(1964) spliced words out of a recorded conversation and asked speakers to
identify them. Only about 50 per cent of the words presented in isolation were
intelligible, but performance improved dramatically when the same words
were heard in a short phrase. By varying how the linguistic context was pre-
sented (visually or visually and auditorily), Pollack and Pickett were able to
show that the improvement was partially attributable to listeners adjusting to
the specic quality of the speakers voice and partially to the lexical and syn-
tactic constraints introduced by the additional material, a conclusion which
was corroborated by later research (see Massaro 1994; Samuel 1986). Thus, in
addition to all the complexities already discussed, listeners are also frequently
forced to guess what the speaker actually said.
DABROWSKA text Make up 26/3/04 4:34 pm Page 13
2. Speed
(Grosjean 1980; Marslen-Wilson 1987) well before the speaker has nished
articulating them!
Lexical ambiguity is also resolved very quickly. In fact, one of the most hotly
debated issues in psycholinguistics is whether the contextually irrelevant mean-
ings of ambiguous words are accessed at all, or whether access is selective from
the very beginning. This is a question of considerable theoretical signicance. If
comprehenders are able to use context to access only the relevant sense, we would
have to conclude that language processing is highly interactive. On the other
hand, initial activation of all senses is usually taken as supporting modular
accounts of processing, according to which language comprehension and pro-
duction is performed by a series of relatively autonomous components one spe-
cialising in word recognition, another in syntactic analysis, and so on.
Unfortunately, the large number of studies investigating the processing of
ambiguous words have produced mixed results: some studies support the selec-
tive access account, and others do not (see Simpson 1994 for a review). There is
an emerging consensus that the reason for these conicting ndings is that lexical
access is not an all-or-nothing phenomenon: the dierent senses of an ambig-
uous word are probably activated to varying degrees, depending on their relative
frequencies (with frequent meanings being inherently easier to activate) and the
strength of the contextual constraints. Thus, highly entrenched senses are di-
cult to suppress regardless of how inappropriate they may be in a given context;
but contextual factors seem to have a considerable eect on the activation of less
frequent senses.
However this debate is eventually resolved, the important point is that even
when several meanings are initially accessed, the contextually irrelevant ones
are deactivated very quickly. Tanenhaus et al. (1979) and Seidenberg et al.
(1982), for example, found that if an ambiguous word was presented immedi-
ately after a word which was semantically related to one of its senses, both
senses were primed that is to say, people were able to make lexical decisions
faster. If, however, the target word was presented after a short delay (about 200
ms), the facilitatory eect was found only when the prime was related to the
relevant sense of the target word, which suggests that the irrelevant sense had
already been eliminated at that point. This was true for noun/verb ambiguities
(He bought a watch/They decided to watch), which can be resolved using
morphosyntactic information, as well as noun/noun ambiguities (You should
have played the spade/Go to the store and buy a spade), which require the inte-
gration of linguistic and real-world knowledge.
These early studies have shown that semantic and pragmatic knowledge are
integrated with the incoming acoustic information extremely quickly. This was
conrmed by more recent studies which attempted to determine the point at
which participants establish reference by monitoring their eye movements as
they examined a visual display accompanying a linguistic stimulus. When
people are presented with a visual context for a spoken sentence, their eye
movements are closely synchronised with the unfolding spoken sentence. For
DABROWSKA text Make up 26/3/04 4:34 pm Page 15
example, if they hear an instruction such as Put the apple on the towel while
viewing an array consisting of an apple, a towel and a few distractor objects,
they look at the apple on hearing apple; then, on hearing towel, they shift their
gaze to the towel. Thus, tracking a persons eye movements gives us a window
onto the mental processes underlying comprehension.
Using this methodology, Eberhard et al. (1995) were able to demonstrate
that when the visual array does not contain a distractor object with a similar-
sounding name, people are able to integrate acoustic, linguistic and visual
information and pick out the intended referent on average 55 ms before the
speaker has nished articulating the word. Altmann and Kamide (1999) went
one better and showed that in a suciently constraining context, listeners can
identify the probable referent before the speaker even begins the correspond-
ing word. In their experiment, people listened to sentences which contained
verbs with fairly specic selectional restrictions (e.g. The boy will eat the cake)
as well as more general verbs (e.g. The boy will move the cake). Both types of
stimulus were accompanied by a visual scene containing the objects men-
tioned in the sentence (in this case, a boy and a cake) as well as some distrac-
tors (e.g. a train set, a ball and a toy car). In this context, the verb eat is highly
predictive, since there is only one edible object in the scene, whereas move is
not. Listeners were able to use this information online and, in the eat condi-
tion, looked at the cake shortly after the oset of the verb.
Syntactic analysis and the integration of grammatical, semantic and con-
textual information also proceed at lightening pace so fast, in fact, that it is
very dicult to determine whether these are distinct processing stages, or
whether the processor immediately integrates all the information available.
Evidence for the former position comes from the so-called garden-path
eects. When confronted with structurally ambiguous input, readers and lis-
teners often have a strong preference for one analysis over the other (Ferreira
and Clifton 1986; Frazier 1987; Rayner et al. 1983). For example, on encoun-
tering the sentence fragment put the book on the shelf, people have a strong ten-
dency to interpret the prepositional phrase as a complement of the verb (a
phrase specifying where the book was put) rather than a modier of the noun
(specifying where the book was located before it was moved). Sentences in
which this expectation is violated (e.g. She put the book on the shelf in the box)
are more dicult to process, and the diculty manifests itself in longer
reading times in the disambiguating region (i.e. the second prepositional
phrase). The traditional interpretation of this nding is that the processor has
been lead up the garden path to an incorrect analysis, and additional process-
ing time is needed to reject it and compute a new one.3 Crucially, such eects
sometimes arise even when context or pragmatic considerations favour the dis-
preferred reading (Ferreira and Clifton 1986; Rayner et al. 1983). This has
been taken as supporting a modular view of language processing in which syn-
tactic analysis is informationally encapsulated and precedes semantic and
pragmatic processing. This conclusion, however, is controversial, since many
DABROWSKA text Make up 26/3/04 4:34 pm Page 16
studies have found that semantic and pragmatic processing operate on the
same time scale as parsing, and, if the contextual bias is strong enough, can
even guide syntactic analysis (Altmann et al. 1998; Marslen-Wilson and Tyler
1987; Ni et al. 1996; Spivey et al. 2002; Trueswell et al. 1994). Moreover,
regardless of whether there is a purely syntactic stage, it is clear that other
types of processing are also performed very rapidly: the processing cost of a
garden path (the time required to reject the initial analysis and compute a new
one) is usually quite short: 100 ms or thereabouts.
The view that semantic and pragmatic processing begins immediately and
proceeds as quickly as parsing appears to be corroborated by research on
event-related brain potentials (ERPs). This involves measuring electrical
activity at the scalp while a person is engaged in some linguistic activity often
listening to sentences containing some kind of anomaly, since anomalies elicit
distinctive ERPs which can be linked to a specic point in the sentence. There
are three main ERP proles associated with language processing:
Thus, ERP responses to semantic anomalies are observed at the same time
as (and possibly earlier than) responses to syntactic anomalies (Brown et al.
2000; Osterhout and Nicol 1999). Both begin while the word is still being
heard and are completed at the end of the word or shortly afterwards. In fact,
a latency of 200 ms post-onset suggests that the brain begins to react to the
anomaly even before there is enough acoustic information to identify the word
uniquely. This is conrmed by a study by Van Petten et al. (1999), who timed
the beginning of the N400 eect with respect to the isolation point the point
in the word at which most people are able to recognise it. Van Petten and col-
leagues identied the isolation point for a number of English words using the
gating technique. (This involves presenting people with very short initial frag-
ments of a word: e.g. the rst 50 ms, 100 ms, 150 ms and so on, until they have
DABROWSKA text Make up 26/3/04 4:34 pm Page 17
identied the word correctly.) They then presented another group with sen-
tences which ended in a word which was either fairly predictable from the
context (e.g. He spilled some of the oil onto the engine because he didnt have a
funnel) or incongruous (He spilled some of the oil onto the engine because he
didnt have a tunnel) and measured ERPs for the nal word. They found, in line
with previous research, that the incongruous completions elicited a much
larger N400 than the congruous completions. The proles for the two condi-
tions began to diverge about 200 ms before the isolation point. This shows that
the participants in the experiment must have either anticipated likely comple-
tions for the sentences and matched them against the incoming signal, or used
the rst phonemes of the word to activate a cohort of candidate words and
compared their semantic representations with the semantic representation of
the sentence being processed. Whatever they did, the experiment shows that
the incoming acoustic information is immediately integrated with the repre-
sentation of the sentence at multiple levels.
3. Processing shortcuts
provide the correct solution but are very demanding computationally). In the
following three subsections, we will explore some of these in relation to lan-
guage processing.
frequency forms, even when the frequency of the root is controlled for. This
nding has been reported for the Dutch third person singular (Schreuder et al.
1999), the Dutch plural (Baayen et al. 1997), and English noun morphology
(Sereno and Jongman 1997).5
Bod (2001) used a similar method to show people may also store simple sen-
tences. He presented people with high-frequency sentences (e.g. I like it), low-
frequency sentences (e.g. I keep it), and ill-formed strings consisting of the
same words (e.g. It I like). The high- and low-frequency sentences were iden-
tical in syntactic structure, were equally plausible and contained lexical items
of similar frequencies. The participants task was to decide as quickly as pos-
sible whether the word string was an English sentence or not; and they made
this decision faster for the high-frequency sentences.
Additional evidence suggesting that people store regularly inected forms
comes from studies which examined pronunciation. Word-nal stops in
certain consonant clusters are usually shorter and articulated less clearly than
in other positions, or they may be deleted altogether, particularly in rapid
speech: for example, band and guest may be pronounced [bn] and [s]
respectively. Such simplications are usually restricted to consonant clusters
which are part of the same morpheme: the nal clusters in banned and guessed
tend not to be simplied, in spite of the fact that these words are homopho-
nous with band and guest (New 1972). However, the nal [t] and [d] in high-
frequency past-tense forms behave like the corresponding segments in
monomorphemic words: they are often omitted in rapid speech, and, when
present, they tend to be pronounced less clearly (Bybee 1995, 2000), which
suggests that, like the monomorphemic forms, they are stored.
As indicated earlier, there is very little work investigating whether adults
store regular (i.e. non-idiomatic) combinations of words, although we do
know that children do, and that such units play an important role in language
acquisition (see Chapter 9). Bybee and Scheibman (1999) have demonstrated
that similar reductions occur in high-frequency word combinations: the form
dont in the expressions I dont know, I dont think, why dont you is phonetically
very dierent from the dont we nd in novel combinations, which, by the logic
outlined above, suggests that it is part of a stored chunk. There is also some
less direct evidence from language change. It is well known that frequently
occurring word combinations often fuse into single words. For example, in
English, going to, want to, have to are often pronounced gonna, wanna, hafta
(note, too, the conventionalised spelling); auxiliaries are often contracted
when they follow pronominal subjects (Im, weve, shell) and the negative par-
ticle not may attach to the preceding auxiliary (havent, doesnt, etc.). It is also
quite common for prepositions to fuse with the following article, as in the
German am (an + dem), zur (zu + der), aufs (auf + das) etc. For such phono-
logical reductions to occur, the participating words must have become single
chunks of structure in speakers minds, since they are not systematic but
restricted to specic combinations of words.
DABROWSKA text Make up 26/3/04 4:34 pm Page 22
combine prefabs with words. For example, in order to form an utterance such
as Why dont you stay at home tonight?, the speaker might integrate two pre-
constructed units (why dont you VP, stay at home) and the adverb tonight; in
order to construct our original example, Tess blew her nose on a paper tissue,
the speaker will need three chunks, X blow XPOSS-DET nose, on a N and
paper tissue, as well as the pronoun her and the proper noun Tess. Such piecing
together requires grammatical knowledge, since the component chunks must
be combined in a specic way if the result is to be grammatical. For example,
only units of a particular kind (viz., VPs with an untensed verb) can be
inserted into the slot after why dont you; the possessive determiner in the
second sentence must be coreferential with the subject; and so on. However,
the processing required to accomplish this is relatively shallow: the speaker
must attend to the relationships between units and the underspecied parts of
other units, but not to every detail of their internal structure.
There is growing evidence that the processing involved in ordinary language
comprehension is, in fact, fairly shallow: that is to say, people may integrate
the meanings of the most salient lexical items, some easily accessible grammat-
ical cues, and contextual information to arrive at a semantic interpretation
without fully parsing the sentence.
One of the earliest studies demonstrating this was conducted by Fillenbaum
(1974), who asked a group of students to paraphrase some sentences as accu-
rately as possible. Although they were specically instructed not to improve
the sentences or make them more sensible, the participants in the study often
normalised pragmatically anomalous sentences. For example, perverse
threats such as Clean up the mess or I wont report you were given a more plau-
sible paraphrase (e.g. If you dont clean up the mess I will report you) 54 per cent
of the time. The participants were then asked to reread each sentence and its
paraphrase in the presence of the experimenter and asked whether they could
see some shred of dierence between the two and even then, they failed to
spot the discrepancies 34 per cent of the time.
We cannot be sure what caused Fillenbaums subjects to give inaccurate par-
aphrases. They may have taken a pragmatic shortcut that is to say, arrived
at an interpretation without fully analysing the sentence; but it is also possible
that they interpreted the sentence correctly, but, since the interpretation didnt
make much sense, rejected it in favour of something more plausible. However,
the fact that some participants appeared to be unable to spot the discrepancy
between the meanings of the original sentences and their paraphrases of them
when they were specically invited to do so in the second phase of the experi-
ment does suggest that they took the pragmatic shortcut route.
Further support for this interpretation of Fillenbaums results is provided
by Ferreira (2003), who found that people often misinterpreted passive sen-
tences describing unlikely events: for example, when asked to name the
aected participant in The dog was bitten by a man, the participants said man
in over 20 per cent of all trials. However, they practically never made this
DABROWSKA text Make up 26/3/04 4:34 pm Page 24
mistake with active sentences: The man bit the dog was virtually always inter-
preted correctly. So it is not the case that people have a tendency to correct
pragmatically odd sentences: normalisation seems to occur primarily with sen-
tences which are non-canonical, and hence fairly dicult to process.
In my own work (Dabrowska 1997), I have shown that shallow processing
can also occur in the absence of a pragmatic bias towards an incorrect inter-
pretation and is inversely correlated with educational achievement: that is to
say, the more formal schooling a person has had, the more likely they are to
make full use of the grammatical information available. The least educated
participants in the study were usually able to extract locally encoded informa-
tion, but had diculty in processing long-distance dependencies. For example,
when confronted with parasitic gap7 sentences such as It was her boss that
Louise persuaded that Margaret will interview, they were much better at iden-
tifying the interviewer (Margaret) than the interviewee (the boss). They were
also sometimes unable to establish hierarchical relationships between clauses
in multiply embedded sentences. Thus, the sentence Paul noticed that the fact
that the room was tidy surprised Shona was frequently interpreted to mean
Paul noticed that the room was tidy and this surprised Shona in other
words, the respondents appeared to extract simple clauses from the input (Paul
noticed something, the room was tidy, something surprised Shona) and then
attempted to integrate their meanings without making use of the grammati-
cal information available. Similar results have been reported by Chipere (2001,
2003).
Let us conclude the discussion with a particularly tricky example. Wason
and Reich (1979) found that nearly all the people they tested misinterpreted
the sentence No head injury is too trivial to ignore to mean No head injury
should be ignored, however trivial it may be. But according to the rules of
English grammar, the sentence cannot possibly mean this: it can only mean
the exact opposite. Consider the following:
Therefore:
Why is it that people systematically misunderstand the last sentence, but not
the other three? The problematic sentence is quite complex, both syntactically
DABROWSKA text Make up 26/3/04 4:34 pm Page 25
and semantically. The noun phrase no head injury, which functions as the
theme of ignore, does not appear in its usual direct object position (as in We
ignored the head injury) but at the beginning of the sentence, as the subject of
the main verb, which makes it dicult to establish the relationship between
them. The sentence also contains three negatives: the particle no and two log-
ically negative expressions, too trivial (i.e. not serious enough) and ignore
(take no notice). Furthermore, it is very odd from a pragmatic point of view,
since it presupposes that serious matters should be ignored and trivial things
attended to (cf. This is too trivial to ignore), as well as enjoining the addressee
to ignore head injuries, which are widely known to be potentially dangerous.
It is not particularly surprising, then, that many people misinterpret the sen-
tence. What is interesting, however, is that they are so condent about their
erroneous interpretation and nd it extremely dicult to see the correct one,
even when they are told what it is.
3.3 Frequency
Fluent language users are remarkably sensitive to the statistical properties of
their language. This sensitivity manifests itself at all levels of linguistic organ-
isation, from phonemes and combinations of phonemes through to words and
grammatical constructions (Ellis 2000), and appears to be quite detailed.
People seem to know not just the relative frequencies of individual words and
word senses, but also how often they appear in specic constructions: for
example, that some transitive verbs (e.g. accuse) are much more likely to occur
in the passive construction than others (e.g. watch: see Trueswell 1996).
Similarly, when a verb allows more than one type of complement, people
know which kind of complement it typically takes: e.g. while both push and
move can occur in either a transitive or an intransitive construction, push
prefers the former and move the latter (MacDonald 1994); accept takes NP
complements more often than sentential complements, while suggest has the
opposite preference (Garnsey et al. 1997); imply is more likely than hope to
take an overt complementizer when it occurs with a sentential complement
(Trueswell et al. 1993); and so on (see MacDonald 1999 for further examples).
People are also sensitive to frequency information at more abstract levels.
For example, when confronted with an ambiguous expression such as
Someone stabbed the wife of the football star who was outside the house, where
the relative clause (who was outside the house) can modify either the upstairs
or the downstairs NP (the wife or the football star, respectively), English
speakers usually prefer the latter interpretation that is to say, in the absence
of contextual information supporting the other reading, they interpret the
sentence to mean that it was the football star, not the wife, who was outside
the house. However, speakers of many other languages, including Spanish,
normally choose the other interpretation when they hear or read the corre-
sponding sentence in their own language.8 This reects the statistics of the
DABROWSKA text Make up 26/3/04 4:34 pm Page 26
Some linguistic expressions (e.g. irregular forms) are retrieved from the
memory and others (e.g. all novel expressions) must be computed. Still others,
such as frequently occurring word combinations and inected forms of
common words, can be either retrieved or computed. However, there is a
trade-o between storage and computation: a set of general, highly compact
rules does not require a large storage capacity, but poses severe demands on
the processing system; and conversely, the retrieval of prefabricated chunks is
computationally less demanding, but requires a large and ecient memory.
Linguists like general rules and principles. When choosing between compet-
ing accounts of the same phenomena, they have traditionally relied on the prin-
ciple of economy. General solutions are always preferred to more specic ones,
and it is assumed that any regularities that can be captured by means of general
rules and principles should be so captured. It is also held as a matter of course
that anything that is computable need not be, and hence is not, stored.
DABROWSKA text Make up 26/3/04 4:34 pm Page 27
Notes
(as in Yes, she is), the articles, and regular past-tense inections. Still later come
the third-person -s and uncontractible auxiliary forms, and, nally, the
contractible copula and auxiliary (e.g. Shes tired, Hes coming) (see Brown
1973; de Villiers and de Villiers 1973). Auxiliaries and various complex con-
structions such as questions, Stromswold (1995) maintains, are also acquired
in the same order by nearly all children.
The universality and alleged uniformity of language acquisition are some-
times taken as evidence for an innate universal grammar which develops
according to a biologically determined timetable and does not depend on
other cognitive subsystems. As we shall see in this chapter, there are some
serious problems with this argument.
1. Individual dierences
as their receptive vocabularies (i.e. they are able to produce almost all the
words they understand), while others may be able to understand more than
200 words, but still not produce any (Bates, Dale and Thal 1995). Children also
dier with regard to the kinds of words they learn in the initial stages of lexical
development. So-called referential children initially focus primarily on object
labels (that is to say, concrete nouns), while expressive children have more
varied vocabularies with more verbs and adjectives and a signicant propor-
tion of social routines and formulas such as stop it, I want it, dont do it (Nelson
1973, 1981). Finally, there are dierences in the pattern of growth. While many
children do go through the vocabulary spurt that Stromswold alludes to at
some point between 14 and 22 months, a signicant minority about a quarter
show a more gradual growth pattern with no spurt (Goldeld and Reznick
1990).
Grammatical development is also anything but uniform. Children may
begin to combine words as early as 14 months or as late as 25 months, and
show correspondingly large dierences in MLU (mean length of utterance)
and other global measures of grammatical development a year later (Bates,
Dale and Thal 1995). At 42 months, the dierences between the most and least
advanced normal children are equivalent to 306 months (Wells 1985). Some
children learn to inect words before they combine them into larger structures,
while others begin to combine words before they are able to use morphologi-
cal rules productively (Smoczynska 1985: 618; Thal et al. 1996). Some children
are extremely cautious learners, while others are happy to generalise on the
basis of fairly sparse evidence, which results in large dierences in error rates
(Maratsos 2000). Marked individual dierences have also been found in
almost every area of grammatical development where researchers have looked
for them, including word order, negation, case marking, the order of emer-
gence of grammatical morphemes and, yes, in the development of English
auxiliaries (Jones 1996; Richards 1990; Wells 1979), questions (de Villiers and
de Villiers 1985; Gullo 1981; Kuczaj and Maratsos 1983) and multi-clause sen-
tences (Huttenlocher et al. 2002; see also Lieven 1997 for a review).
Children also dier in acquisition styles, or the strategies they use to learn
to speak (Nelson 1981; Peters 1977; Peters and Menn 1993). Analytic or ref-
erential children begin with single words, which they articulate reasonably
clearly and consistently. Holistic or expressive children, on the other hand,
begin with larger units which have characteristic stress and intonation pat-
terns, but which are often mush-mouthed and may consist partly or even
entirely of ller syllables such as [dadada]. Peters (1977) argues that holistic
children attempt to approximate the overall shape of the target utterance
while analytic children isolate and produce single words. These dierent start-
ing points have a profound inuence on the course of language development,
since they determine how the child breaks into grammar. Analytic children
must learn how to put words together to form more complex units. They typ-
ically do this by rst learning how to combine content words, whereby they
DABROWSKA text Make up 26/3/04 4:34 pm Page 32
frequency of such sentences in the speech of their parents and, and, more
importantly, their teachers on the other hand (Huttenlocher et al. 2002). The
latter is theoretically more signicant, since, unlike the former, it cannot be
attributed to genetic factors. Many studies of twins have found that they are
often substantially delayed relative to other children (see Mogford 1993), pre-
sumably because they get less personalised input than singletons. But
perhaps the most convincing evidence for the relevance of input tailored to the
childs needs comes from studies of spoken language development in hearing
children of deaf parents. While such children are rarely completely cut o
from spoken language, they clearly get less of it than children born to hearing
parents. Even more importantly, they often get little personalised input from
adults, which has a profound inuence on their language development.
Consider, for example, Jim, a boy studied by Sachs et al. (1981). Jims
parents had very little oral language and communicated with each other in
sign, but they did not sign to him. He was exposed to English primarily
through television, which he watched frequently, but he also occasionally
played with hearing children. Sachs et al. began to study Jims linguistic devel-
opment when he was aged 3;9, just as he was beginning speech therapy. It was
immediately clear that Jims language was not just delayed, but qualitatively
dierent from that of normally developing children. Although he had a rea-
sonably good vocabulary (only slightly smaller than an average child of his
age), his comprehension was quite poor, and he also had severe articulation
problems. Perhaps as a result of these, he only produced very short utterances:
his mean length of utterance (MLU) at the beginning of the study was 2.2,
which would be appropriate for a child just over two. Unlike ordinary children
learning English, Jim made many word order errors, and often produced
decidedly un-English sentences such as those in (1).
stand, and provide some sort of feedback on the childs production. This
could be in the form of expansion of the childs incomplete utterances, puz-
zlement at incomprehensible utterances, or simply a continuation of the topic
introduced by the child or compliance with a request, both of which would
indicate that the childs communicative intentions have been understood.
Secondly, Jims language shows that restricted input aects morphosyntactic
development as well as lexical development in fact, in his case the former was
clearly much more seriously aected than the latter. It is worth pointing out
that Jims case is not an isolated: Todd and Aitchison (1980) report similar
deviant constructions in the speech of another hearing child of deaf parents.
So both the quality and the sheer quantity of input do matter. On the other
hand, they do not seem to matter as much as one might expect, since children
who get only a tenth of the input addressed to the most priveleged children do
not take ten times longer to achieve the same level of competence. How can
we explain this apparent discrepancy?
It is important to bear in mind that most children get vast amounts of input.
Cameron-Faulkner et al. (in press) estimate that a typical toddler hears
between 5,000 and 7,000 utterances per day. This means that between the ages
of 1, when language acquisition may be said to begin in earnest, and 4, when
most children achieve productivity in the basic structures of their language,
they will have heard something of the order of 6,570,000 utterances, includ-
ing about a million WH questions, a million Y/N questions, 660,000 instances
of the full transitive construction (a subject followed by a verb followed by an
object), about 40,000 relative clauses, and almost 7,000 passives.4 So halving
the amount of input may have little eect on language learning simply because
most children get more input then they actually need: half of 6.6 million is still
a very large number.
The second point to bear in mind is that while reducing the amount of lan-
guage experience may have little eect on the linguistic development of some
children, it does aect others namely, children with low IQs, small short-
term phonological memory capacity, and sensory decits who appear to be
far more vulnerable to the eects of suboptimal input (Snow 1995). Thus,
quality and amount of input may help to explain an apparent paradox. There
are reports in the literature of children who developed language normally in
spite of very low IQ, small short-term phonological memory capacity, or other
impairments, which have led some researchers to conclude that language
development does not depend on such cognitive abilities. On the other hand,
speech therapy clinics are full of disabled children with severe language prob-
lems, which suggests that there are close links between linguistic and non-
linguistic development. We can reconcile these conicting facts by recognising
that non-linguistic cognition does matter but good input can compensate for
cognitive impairments (within limits); and, conversely, good language learn-
ers may be able to compensate for suboptimal input. However, when several
inauspicious circumstances conspire, language will suer.
DABROWSKA text Make up 26/3/04 4:34 pm Page 36
fact that blind children develop language somewhat dierently from seeing
children. Because many aspects of contextual information are not accessible
to them, blind children must make very good use of whatever information they
have got, and cognitive impairment may preclude them from doing so.
Furthermore, blind children have to be more proactive in obtaining informa-
tion about meaning. There are two ways in which learners can determine what
a new linguistic expression means: they can either observe competent speak-
ers use it and note as much as possible about the context, or they can experi-
ment that is to say, produce it and see what happens. Since blind children
cannot observe as much as sighted children, they must experiment more. The
result is that they often produce utterances which they clearly do not under-
stand. This meaningless language or verbalism of the blind is sometimes
regarded as an indication of deviant development, or even cognitive impair-
ment. However, it seems more illuminating to view it as a very good way of
obtaining clues about what linguistic expressions might mean.5
There is considerable evidence that blind children make more use of imita-
tion, repetition and verbal routines than seeing children, and tend to adopt a
holistic learning strategy (Dunlea 1989; Mulford 1988; Prez-Pereira 1994;
Urwin 1984). This suggests that they rely more heavily on distributional infor-
mation, presumably in order to compensate for the lack of visual information
(Prez-Pereira and Castro 1997). This language learning strategy, of course,
makes greater demands on the childs phonological memory. Thus, a short
phonological memory span, or poor auditory processing skills, will aect
blind children more than seeing children, and may be another reason for the
large individual dierences in the linguistic accomplishments of blind chil-
dren.
A particularly striking example of how a blind childs heavy reliance on
verbal routines may result in the production of very unusual structures is given
by Wilson and Peters (1988). Between the ages of 3;2 and 3;6, the child they
studied, Seth, produced a number of questions such as those given in (2) below
(Wilson and Peters 1988: 2513).
(2) a. What are we gonna go at [=to] Auntie and? [answer: Auntie and
Priya]
b. What did I get lost at the, Dad? [answer: at the store]
c. What are you cooking pan? [answer: pancakes]
Such utterances are peculiar by any standards; but they are particularly prob-
lematic for Principles and Parameters theory, since they apparently violate
universal constraints on movement. Sentence (2a) violates the coordinate
structure constraint, which states that items cannot be moved out of a coor-
dinate structure (or a more general constraint, subjacency, which states that
constituents cannot be moved across more than one bounding node: in
this case, NP and S). (2b) and (2c) are violations of the maximal projection
DABROWSKA text Make up 26/3/04 4:34 pm Page 38
It seems that in the anomalous utterances in (2), Seth put such a sentence with
a blank in an interrogative structure.
These anomalous constructions in Seths speech were not, of course, a
direct result of blindness, but rather a consequence of his unusual linguistic
experience. They do, however, oer particularly compelling evidence that
unusual experience can result in unusual language production.
We have seen that there are considerable individual dierences in the way that
children learn to use language. Some children rst produce single words
(mostly nouns) and later learn to combine them to form more complex expres-
sions; others begin with short phrases or sentences and extract words from
them; and still others probably the majority use both strategies. Some chil-
dren are rapid piecemeal learners and use a variety of grammatical construc-
tions in very early stages of development, before they have fully analysed
them; others look for more general patterns and consequently may appear to
be less advanced. Some children jump to conclusions on the basis of relatively
little data a risky strategy which results in high error rates; others are more
cautious and consequently make relatively few errors.
We have also seen that it is not true that language acquisition is insensitive
to environmental and non-linguistic cognitive factors. Environmental depri-
vation, sensory impairments and, as we shall see, cognitive decits all have an
eect on language development. However, unless several dierent factors con-
spire, children are usually able to overcome these adverse circumstances and
develop normal language. As we will see in the next two chapters, children with
very low IQs and children who have suered damage to the areas of the brain
which normally specialise in linguistic processing may also develop relatively
DABROWSKA text Make up 26/3/04 4:34 pm Page 39
Notes
It has been known for well over a century that damage to the language areas
of the brain often has catastrophic eects on an individuals ability to produce
and understand language. Lesions in the third frontal convolution of the left
hemisphere, the so-called Brocas area, often lead to Brocas aphasia, which is
characterised by slow, eortful speech and diculties with function words and
inections (cf. example (1) below). Patients suering from this type of aphasia
are usually able to understand ordinary spoken language, although they may
struggle on formal tests of comprehension where they are deprived of contex-
tual information and must rely on grammatical cues alone. For example, they
may be unable to work out who chased whom given a semantically reversible
passive sentence such as The boy was chased by the girl, but will have no prob-
lems with The soup was eaten by the girl, where they can use a non-syntactic
strategy. Lesions to Wernickes area, i.e. the upper posterior part of the left
temporal lobe, in contrast, tend to lead to severe comprehension diculties in
the presence of uent but often deviant speech (cf. example (2)). Such patients
make frequent lexical errors, substituting semantically related but inappropri-
ate words (e.g. table for chair) or incorrect sounds (e.g. bekan for began), or, in
more severe cases, unidentiable phonological forms such as taenz.
(1) [Retelling the story of the fox and the crow] King . . . Singing . . .
Singing loud . . . Meat. Perfect! (Goodglass 2001: 64)
(2) [Response to a therapists question about why he is in hospital]
Whats wrong with me because I . . . was myself until the taenz took
something about the time between me and my regular time in that
time and they took the time in that time here and thats when the the
time took around here and saw me around in it its started with me no
time and then I bekan work of nothing else thats the way the doctor
nd me that way. (Obler and Gjerlow 1999: 43)
knowledge is actually embodied in the wiring of these areas of the brain: spe-
cically, that Brocas area is the seat of grammatical knowledge while lexical
knowledge is at least partially represented in Wernickes area. As we shall see,
however, such claims are rather problematic for a number of reasons.
The rst problem with the strict localisation view is the fact that the relation-
ship between the lesion site and the nature of the resulting impairment is far
from straightforward. Language impairment sometimes occurs after brain
damage outside the classical language areas (Dronkers et al. 2000), even in
right-handed monolingual patients with no known prior neurological condi-
tion which might have caused a reorganisation of language function in the
brain. If fact, aphasic-like symptoms may occur after damage to subcortical
areas like the basal ganglia and the thalamus (Basso et al. 1987; Brown 1975;
Crosson et al. 1986; Dmonet et al. 1991; Lieberman 2002). This suggests that
many regions of the brain outside the classical language areas are involved in
language processing a supposition conrmed by a growing number of brain
imaging studies (Dmonet et al. 1992, 1993; Mazziotta and Metter 1988;
Posner et al. 1988; Warburton et al. 1996; Wise et al. 1991; Zatorre et al. 1992).
Conversely, damage to the language areas does not always lead to language
impairment, or to the predicted type of impairment. Dronkers et al. (2000)
describe patients with damage to Brocas area who are not agrammatic, and
patients with lesions to Wernickes area who do not have Wernickes aphasia.
Furthermore, persistent full-blown aphasias are rarely found in patients with
damage to Brocas or Wernickes area alone: when the lesion is conned to just
the classical language areas, patients usually recover (Dronkers et al. 2000;
Lieberman 2002).
Recovery after damage to the language areas of the brain is particularly
good in children. As a rule, children recover quite quickly from the eects of
the lesion, and, if it occurs before the onset of language, the expectation is that
they will make a more or less full recovery that is to say, within a few years,
their language abilities will be within the normal range, though usually slightly
below normal controls (Aram 1998; Bates 1999; Bates et al. 1997). Lesions sus-
tained in middle and late childhood typically do leave some decits, although
these are more likely to be apparent in the childrens writing and schoolwork
than in ordinary conversation (Bishop 1993; Martins and Ferro 1992, 1993;
van Hout 1991).
It is also worth noting that the aphasic symptoms observed in children who
have suered brain damage are very dierent from those found in adults.
Wernickes (uent) aphasia is extremely rare in children, even after damage to
Wernickes area, as is the telegraphic speech characteristic of Brocas aphasia.
Child aphasics typically dont talk very much, and when they do, they use
DABROWSKA text Make up 26/3/04 4:34 pm Page 42
simpler structures but they rarely make errors. Furthermore, children are
likely to have a mild aphasia after just about any kind of brain damage, includ-
ing damage to the right hemisphere and the non-linguistic parts of the left
hemisphere (Aram 1998; Gardner 1974). This suggests that the localisation of
linguistic functions is much more widely distributed in children than in adults,
which doesnt t in very well with the idea that the basic blueprint for grammar
is genetically prespecied and embodied in the wiring of Brocas area.
But the most dramatic illustration of the plasticity of linguistic organisa-
tion comes from studies of individuals who lack the left hemisphere alto-
gether. Some cases of catastrophic childhood epilepsy can be cured or at least
alleviated either by removing the entire hemisphere in which epileptic attacks
begin (hemispherectomy) or, more commonly, by stripping it of its cortex
(hemidecortication). Not surprisingly, children who have undergone such
drastic surgery suer from a variety of decits, including linguistic decits.
However, the nature of their impairment is often very dierent from what one
might expect, given what happens in brain-damaged adults.
It is important to bear in mind that the hemispherectomised population is
very heterogeneous: patients dier in the actual condition which lead to the
operation, the age at which it appeared, the age at which they were operated
on, the success of the operation in controlling seizures, and the state of the
remaining hemisphere. All of these factors, and probably many more, have an
eect on how well language develops after surgery. Consequently, the nal
outcome after hemispherectomy is extremely variable. However, given the
special status of the left hemisphere in language processing, we would expect
its removal to have a dramatic and devastating eect on an individuals lan-
guage abilities, while removal of the right hemisphere should normally spare
language (although it might impair the ability to use prosody and to interpret
gurative language, both of which are known to require right-hemisphere
involvement). This expectation, however, is not corroborated by the available
evidence.
A recent survey of 49 post-hemispherectomy patients (Curtiss and de
Bode 1998) found no signicant relationship between side of lesion and lan-
guage outcome. In fact, nearly a third of the right hemispherectomised
patients had no productive language at all, compared to 16.7 per cent of the
left hemispherectomies, which suggests that the right hemisphere plays an
important role in language acquisition. Even more surprisingly, seven of the
thirty left-hemispherectomised patients that is to say, over 23 per cent had
normal or nearly normal language.
Curtiss and de Bode acknowledge that their measure of linguistic ability,
the Spoken Language Rank, is fairly crude, so it is possible that a more sen-
sitive test of morphosyntactic abilities would reveal subtle decits in the left-
hemispherectomised group. Such decits were found by several researchers,
including Dennis and Kohn (1975), Dennis and Whitaker (1976), Dennis
(1980), and Vargha-Khadem and Polkey (1992). Dennis and Kohn (1975), for
DABROWSKA text Make up 26/3/04 4:34 pm Page 43
It is worth noting that many of the ungrammatical sentences used in the study
were locally grammatical, so the relatively high level of performance could not
have been achieved if participants carried out only a supercial analysis. As
the authors point out:
At least the following kinds of syntactic ability are required by our task:
an awareness of strict subcategorisation frames, considerable sensitivity
to the function vocabulary, the ability to compute hierarchical phrase
structure, and the ability to handle discontinuous syntactic dependen-
cies. It appears that the task could not be performed successfully by
using general semantic knowledge or an appreciation of transitional
probabilities and/or prosodic patterns associated with particular lexical
items. In general, any appeal to template matching as an alternative
strategy for performing this task seems to us to require templates which
DABROWSKA text Make up 26/3/04 4:34 pm Page 45
oers strong evidence in favour of the claim that the selective vulnerability of
grammatical morphemes in aphasia may be at least partly attributable to
global diminution in cognitive resources.
Most of the work reviewed so far is based on aphasia in speakers of English
and other typologically similar languages. In recent years, however, it has
become increasingly clear that brain damage may aect speakers of dierent
languages in dierent ways and consequently, that some of the conclusions
about the organisation of language in the brain arrived at on the basis of data
from the familiar European languages may be in need of revision. Two main
ndings are relevant in this connection. First, it appears that users of typolog-
ically very dierent languages may suer aphasia after lesions which do not
cause linguistic problems in speakers of English. The traditional literature on
aphasia, for example, is virtually unanimous in asserting that in the vast
majority of cases, language impairment occurs after damage to the dominant
hemisphere i.e. the left hemisphere in right-handers and the right hemisphere
in most left-handers. Damage to the non-dominant hemisphere may aect a
patients pragmatic skills, but it rarely results in serious decits in the core lin-
guistic areas of lexis and grammar. However, this does not appear to be the
case in Chinese. Hu et al. (1990) reported that in their sample of Chinese right-
handed aphasics, about 30 per cent had suered damage to the right (i.e. non-
dominant) hemisphere. This suggests that Chinese speakers, or at least a large
proportion of Chinese speakers, make much greater use of the right
hemisphere when speaking than do speakers of Indo-European languages.
Researchers working on sign language have also found that linguistic decits
in signers can occur following damage to the right hemisphere and some non-
linguistic regions of the left hemisphere as well as Brocas and Wernickes
areas, suggesting that the areas of the brain involved in processing sign lan-
guage only partly overlap with those used by spoken languages (Corina 1998).
These ndings have been conrmed by imaging studies (Corina 1996; Neville
and Bavelier 2000). Thus, it appears that experience with dierent languages
might cause speakers to recruit additional regions of the brain in addition to
the classical language areas.
Secondly, it has been shown that the traditional aphasias, and Brocas
(non-uent) aphasia in particular, look very dierent in typologically dier-
ent languages. One of the hallmarks of Brocas aphasia in English-speaking
patients, apart from dysuency and other problems with articulation, is
agrammatism, which is often manifested as telegraphic speech: the patient is
able to produce nouns and some other content words, but has severe diculty
with function words and grammatical endings, resulting in speech output rem-
iniscent of the style we nd in telegrams. However, non-uent aphasics in
highly inected languages such as Polish, Serbo-Croatian, German and
Italian, and agglutinative languages such as Turkish, Finnish and Hungarian,
are not agrammatic that is to say, they do not omit inections, though they
often use them inappropriately (Bates et al. 1991; Menn and Obler 1990;
DABROWSKA text Make up 26/3/04 4:34 pm Page 47
Slobin 1991). Slobin (1991), for example, found that inectional morphology
in Turkish Brocas aphasics was remarkably well preserved . . . with hardy a
trace of telegraphic speech (1991: 150). His subjects tended to provide the
right grammatical morphemes where required, although compared to normal
controls, the range of axes in their speech was quite restricted similar to
the range of forms available to Turkish 3-year-olds. Slobin also notes that in
cases of word-nding diculty, patients often used pronouns instead of
nouns, and these were correctly inected. Even the most non-uent patient,
whose speech consisted mostly of single word utterances, provided the appro-
priate case inections on nouns.
Thus, the results from highly inected languages are at odds with those
obtained from English-speaking patients. The agrammatic patients described
by Slobin and Bates et al. had suered similar lesions to agrammatic English-
speaking aphasics, yet they had clearly not lost all grammatical knowledge. It
seems safe to conclude, then, that grammatical knowledge does not reside in
the pattern of connections in Brocas area.
Elizabeth Bates and her collaborators (Bates and Goodman 1997, 1999; Dick
et al. 2001), there are some rather striking parallels between a persons lexical
and grammatical diculties, which can be observed in a wide range of de-
cits, including developmental disorders, aphasia, and the linguistic impair-
ments which accompany Alzheimers dementia and normal ageing. Thus,
individuals who omit function words also tend to omit lexical items rather
than substitute inappropriate ones. This pattern is found in Brocas aphasia as
well as in other non-uent decits such as Down syndrome and some forms
of specic language impairment. In contrast, lexical substitution errors
(semantic paraphasias) and the substitution of inappropriate grammatical
functors (paragrammatism) are the hallmarks of Wernickes aphasia, and are
also quite characteristic of the speech of individuals suering from another
uent disorder, Williamss syndrome. Finally, the excessive use of pronouns
and semantically light forms such as thing, make and do in anomia and
Alzheimers disease (and, to a lesser extent, in normal ageing) is accompanied
by a similar reduction of the grammatical repertoire: individuals belonging to
these groups tend to avoid complex syntactic structures such as passives
(Bates, Harris et al. 1995). The co-occurrence of lexical and grammatical de-
cits in all these groups and the similarity in the symptoms that they demon-
strate across linguistic domains strongly suggest a common underlying
mechanism.
It is undeniable that some regions of the brain are more involved in linguistic,
and specically grammatical, processing than others. However, the strongest
version of the anatomical specialisation hypothesis that grammar resides in
the pattern of connections in Brocas area is clearly false. As we have seen,
there is considerable evidence that individuals who have suered lesions to
Brocas area do not lose their grammatical knowledge, but are simply unable
to access it at will. Furthermore, the most entrenched grammatical patterns,
such as basic word order or case inections in morphologically rich languages,
generally do remain accessible. This suggests that linguistic knowledge is rep-
resented in a redundant manner in various regions of the brain, with the lan-
guage areas acting as a kind of central switchboard. There is also evidence of
close links between grammatical and lexical decits, which in turn suggests
that these two aspects of a speakers linguistic competence are closely inter-
twined.
Another important lesson to be learned from the research on aphasia is that
our capacity to use language is extremely resilient. In immature individuals,
language can survive the loss of the language areas or even of the entire left
hemisphere. In adults, such large-scale reorganisation is not possible, perhaps
because the regions which take over language processing in brain-damaged
DABROWSKA text Make up 26/3/04 4:34 pm Page 49
speech is slow, poorly articulated and agrammatic (they often leave out oblig-
atory function words and inections, at least in English and typologically
similar languages). They are also unable to use grammatical cues in compre-
hension: for example, they have problems with reversible passives such as The
boy is being chased by the girl, where it is impossible to tell from the meaning
of the words alone who does the chasing and who is being pursued. However,
the general cognitive abilities of Brocas aphasics are good: they are able to
solve complex problems (as long as these dont require the use of language),
they behave in socially appropriate ways, and they can make excellent use of
contextual information to infer the meaning of an utterance. In fact, their
pragmatic skills often allow them to compensate very successfully for their
grammatical decit, and for this reason Brocas aphasia was for a long time
regarded as a purely productive decit.
Decits which seem to aect language and language alone have also been
observed in acquisition. While the majority of developmental language disor-
ders can be attributed to hearing loss, general learning diculties or emotional
problems, some children develop language late and with considerable di-
culty in spite of the fact that they appear to be completely normal in every
other respect. Such children, diagnosed as suering from specic language
impairment, or SLI, usually have problems with several aspects of language
use, including articulation, vocabulary and grammar. Their utterances are
shorter and less uent than those of normal children; they frequently omit
function words and axes, and tend to avoid more complex grammatical
structures. They also make more errors of commission than other children,
and perform below age norms on comprehension tests. Most SLI children do
eventually learn to speak normally, although some residual problems often
remain; consequently, they are at risk for reading disorders, and tend to
perform less well than unimpaired children at school.
More surprisingly, language can apparently develop more or less normally
even in the face of severe cognitive impairment. Williamss syndrome (WS) is
a case in point. WS is a rare neurodevelopmental disorder characterised by a
complex of physical abnormalities (including heart and circulatory defects
and a characteristic pixie-like face) and mental retardation. Individuals with
WS typically have performance IQs of about 5060 and nd it dicult to nd
their way around or perform simple everyday tasks such as tying their shoe-
laces. Their linguistic skills, however, appear to be remarkably well preserved.
WS adolescents are uent and articulate quite loquacious, in fact. Their
utterances are generally well-formed and contain a variety of complex gram-
matical structures, including full passives, relative clauses and multiple embed-
ding. Here, for example, is how a 17-year-old WS girl with an IQ of 50
described the experience of having an MRI scan:
(1) There is a huge magnetic machine. It took a picture inside the brain.
You could talk but not move your head because that would ruin the
DABROWSKA text Make up 26/3/04 4:34 pm Page 52
whole thing and they would have to start all over again. After its done
they show you your brain on a computer and they see how large it is.
And the machine on the other side of the room takes pictures from the
computer. They can take pictures instantly. Oh, it was very exciting!
(Bellugi et al. 1993: 183)
In most aphasic patients, some linguistic abilities are preserved (see Chapter
4), so it is inaccurate to say that they lack language; and while they usually
perform well on cognitive tests, their non-linguistic cognition is rarely, if ever,
completely unaected. A number of studies, for example, have reported that
aphasics obtained lower scores than other brain-damaged patients on non-
verbal intelligence tests (see Baldo et al. 2001; Hamsher 1998; Keil and
Kaszniak 2002; Lebrun and Hoops 1974), although it is not clear whether this
is a result of aphasia as such or of constructional apraxia (inability to copy or
construct complex gures), which often accompanies aphasia. Many aphasics
also have problems with relatively simple tasks which do not involve language,
such as using the right colours to ll in outline drawings, associating non-lin-
guistic sounds with pictures, showing what one does with various objects (e.g.
a toothbrush), and sorting. The co-occurrence of the linguistic problems with
such non-linguistic decits does not entail that there is a functional link it is
possible, for example, that these non-linguistic tasks are subserved by adjacent
areas of the brain which are damaged at the same time as the language areas
but it should make us careful in drawing conclusions about the presumed
independence of language from other cognitive processes.
The same point can be made even more forcefully with regard to language-
impaired children. Individuals with SLI have problems with language, but
they are certainly not languageless; and although, by denition, they have
normal or above-normal intelligence, there is ample evidence that various sub-
populations of children diagnosed with SLI suer from associated non-
linguistic decits. The most often reported impairments involve auditory
processing and short-term memory, but a variety of other problems have been
observed as well (Bishop 1992; Curtiss and Tallal 1991; Curtis et al. 1992;
Gathercole and Baddeley 1990; Hill 2001; Johnston 1997; Jones et al. 1994;
Merzenich et al. 1996; Montgomery 1995; Vargha-Khadem et al. 1995).
Again, the co-occurrence of language disorders with such non-linguistic de-
cits does not prove that the former are a consequence of the latter: it is pos-
sible that the children had been misdiagnosed, or that their non-linguistic
problems are a result of the language impairment (cf. Stromswold 1996).
However, the existence of such associated decits suggests that the term spe-
cic language impairment is a misnomer.
The case for language without intelligence is equally problematic. It is true
that the hyperlinguistic individuals mentioned earlier have low intelligence.
The WS subjects described in the literature usually have IQs of about 5060,
and Christophers full-scale IQ is probably between 70 and 80. However, this
does not mean that they are cognitive vegetables. As Tomasello (1995) points
out, IQ is a number derived by dividing a persons mental age by their chron-
ological age (and multiplying the result by 100). Thus, the IQ of 50 obtained
by the girl quoted in the preceding section means that, at 17, her performance
on an intelligence test was equivalent to that of an average 8-and-a-half-year-
old. The studies documenting the good performance of WS subjects on
DABROWSKA text Make up 26/3/04 4:34 pm Page 54
linguistic tasks involved adolescents and young adults with mental ages
between 5 and 9. Since normal children in this age range speak uently and
understand passives, negatives, conditionals etc., we should not be surprised
by the fact that WS adolescents are also able to understand such structures.
The same argument also applies to Christopher, of course and in his case,
one might also add that his verbal IQ scores were within the normal range,
that he performed extremely well on a test of pragmatic inference, and that he
was able to teach himself foreign languages by reading grammar books. Thus,
describing people like Christopher or the WS patients studied by Bellugi and
her colleagues as lacking intelligence is simply preposterous.
Furthermore, it is important to note that none of these hyperlinguistic
individuals actually has normal language. People with WS show considerable
language delay in infancy (Bellugi et al. 1994; Paterson et al. 1999; Thal et al.
1989). They tend to make good progress in middle and late childhood, and by
adolescence, their linguistic skills are comparable to those of normally devel-
oping 67-year-old children (Karmilo-Smith et al. 1997; Volterra et al. 1996)
that is to say, about right for their mental age, although even adolescents
have diculties with complex syntactic structures (Grant et al. 2002). In other
words, the linguistic skills of individuals with WS are good only in compari-
son with their non-verbal abilities. This was demonstrated particularly dra-
matically by Stojanovik et al. (submitted), who, unlike other researchers,
directly compared WS and SLI individuals on the same tests. In line with pre-
vious research, they found dissociations. The SLI participants performed at
age-appropriate levels on standardised tests of non-linguistic cognition, and
below age levels on language tests. The WS participants, on the other hand,
performed well on language tests relative to their non-verbal cognition scores.
Crucially, however, they were no better than the SLI participants on the lan-
guage tasks (in fact, they were slightly worse on some measures). Given that
SLI is supposed to be a selective impairment of language, it makes no sense
to say that language skills in WS are intact.
Similar observations can be made about Christopher. We do not know
much about his early linguistic development (he was already 29 when Smith
and Tsimpli began their investigation), although we do know that he was a late
talker. His command of foreign languages, while impressive, is far from
perfect: in fact, he seems to have internalised very little syntactic knowledge,
although his vocabulary and morphology are very good. He tends to follow
English word order in spontaneous production and translation, and his gram-
maticality judgements are strongly inuenced by English. Smith and Tsimpli
conclude that Christophers syntax is basically English with a range of alter-
native veneers (1995: 122) and observe that it is possible that despite speak-
ing many languages, Christopher really has only one grammar (129).
On the other hand, argue Smith and Tsimpli, Christophers command of
his native English is virtually perfect. He performed at ceiling on TROG (Test
of Reception of Grammar), which shows grammatical knowledge equivalent
DABROWSKA text Make up 26/3/04 4:34 pm Page 55
(2) a. *This is the girl that I saw her with John at the cinema.
b. This is the girl that I saw with John at the cinema.
(3) a. *Mary believes the claim which John is a very intelligent man.
b. Mary believes the claim that John is a very intelligent man.
(4) a. *Which student do you think that could solve the problem?
b. Which student do you think could solve the problem?
individuals have poor language skills, again relative to other cognitive domains.
In both groups, however, the decit is a matter of degree.
We now turn to a more important question: what exactly do these dissoci-
ations show? They do indicate that language is not merely an outgrowth of
general cognition. But then there is no such thing as general cognition: what
we have is a varied repertoire of cognitive skills and abilities. Some of these
(e.g. auditory short-term memory) are clearly implicated in language acquisi-
tion and processing, while others (e.g. the ability to nd ones way around in
an unfamiliar neighbourhood) are not. It is to be expected, then, that some
cognitive decits will have a profound eect on language performance, while
others will be of little or no consequence.
On the other hand, the dissociations do not show that grammar is an innate,
domain-specic and informationally encapsulated module indeed, the very
fact that they are partial entails that there are important links between verbal
and non-verbal cognition. Specically, the fact that both lexical and grammat-
ical development is signicantly delayed in virtually all mentally impaired chil-
dren including those who are described as hyperlinguistic suggests that
certain cognitive infrastructures, to borrow Elizabeth Batess (1994) term,
must be in place before grammatical development can begin. One should
stress, however, that these are fairly modest: equivalent, perhaps, to the mental
abilities of a normal 2-year-old (cf. Bates 1997). Furthermore, the fact that the
language of the mentally handicapped again including those who are hyper-
linguistic is also impaired, albeit sometimes to a lesser degree than other
aspects of their cognitive functioning, suggests that some grammatical abil-
ities require more elaborate cognitive infrastructures.
This brings us to the nal point: it is not accurate to say that language pro-
cessing in hyperlinguistic individuals is simply not quite as good as in
normals: there is evidence that it may also be qualitatively dierent. Several
researchers have noted that the spontaneous speech of people with WS often
includes strange word choices. For example, one of the young people studied
by Bellugi observed that The bees abort the beehive (that is to say, leave it);
another indicated that she would have to evacuate the glass (empty it) (Bellugi
et al. 1994: 32). Individuals with WS also tend to use more low-frequency
words than normals on word uency tasks, which prompted some researchers
to conclude that their mental lexicons are organised in an unusual (Bellugi et
al. 1994) or even deviant (Pinker 1991) fashion.
The language of children and adolescents with WS also shows some gram-
matical peculiarities. In contrast to normal children and adults, for example,
they perform better on high-frequency than on low-frequency regular verbs,
but show no frequency eects on irregulars (Thomas et al. 2001). Karmilo-
Smith and colleagues (1997) found that their French-speaking WS subjects
made signicantly more gender-agreement errors than controls, especially
when the experimental stimuli were nonce words rather than real words.
Particular diculties with agreement were also observed in the speech of an
DABROWSKA text Make up 26/3/04 4:34 pm Page 57
Italian WS girl studied by Capirci et al. (1996). Capirci et al. also note that the
girl made syntactic errors that do not occur in the speech of normal children
at comparable stages of linguistic development. Finally, Karmilo-Smith et
al. (1998), using an online language task, found that their WS subjects showed
normal sensitivity to phrase structure violations and restrictions imposed by
an auxiliary on the form of the following verb, but not to lexical verb subcat-
egorisation frames.
Such ndings suggest that people with WS may acquire and process lan-
guage in a qualitatively dierent way (cf. Bellugi et al. 2000; Karmilo-Smith
1998; Karmilo-Smith et al. 1998). There is some support for this conclusion
from brain imaging studies, which have shown that WS individuals show dis-
tinctive ERP patterns which are not found in any other group (Bellugi et al.
1994; Neville et al. 1993). Furthermore, they show no left-hemisphere domi-
nance for language that is to say, unlike normal adults and older children,
they process language bilaterally.
As we saw earlier, Christophers language is also deviant, though apparently
in dierent ways. The dierence is most readily seen when we compare his
spontaneous speech with that of WS subjects. People with WS speak uently
and generally correctly and use a variety of complex syntactic structures.
Christophers conversation, in contrast, is monosyllabic and non-native-like.
On the other hand, Christopher performs much better than WS subjects on
formal tests of language development and on grammaticality judgement
tasks.
At the moment we know too little about the language abilities of hyperlin-
guistic individuals to draw any denite conclusions. It is possible, however,
that they learn and process language in qualitatively dierent ways from
normal people. If this is the case, their abilities oer further evidence of the
human brains ability to wire itself for language in dierent ways, according
to the mental resources available rather than of the existence of an innate,
informationally encapsulated language module.
DABROWSKA text Make up 26/3/04 4:34 pm Page 58
6 Biological underpinnings
All normal humans, and no other species, have language. Many animals, of
course, have quite elaborate communication systems, but human language is
very dierent from animal languages they are not just simpler, but qualita-
tively dierent. Unlike other species, humans are able to refer to objects in
their environment (as well as to absent or imaginary entities) and combine
words and phrases to form novel messages. Furthermore, human languages,
even if they require some innate knowledge, are largely learned, since children
growing up in dierent linguistic communities end up speaking dierent lan-
guages. Last but not least, human languages, unlike the vocalisations of our
closest ape relatives the chimpanzees are under cortical control, and hence
conscious and deliberate. Chimpanzee vocalisations, in contrast, are bound:
they are triggered by a particular emotional state, are very dicult to suppress
in its presence, and cannot be produced in its absence (Goodall 1986).
Virtually all language researchers agree that language acquisition would be
impossible without some kind of innate structure: a atworm cannot learn
Flemish, or a chicken Chinese, however hard they and their teachers might try.
What is much less clear is what kind of innate structure is necessary for this
task, and the extent to which it is task-specic (devoted exclusively to language
learning). According to the generative tradition, which dominated linguistic
research for much of the latter half of the twentieth century, all humans are
endowed with Universal Grammar, a mental module specifying the univer-
sal properties of human languages and the parameters along which they can
vary, without which language acquisition would be impossible. Universal
Grammar is believed to have evolved after we separated from the apes and to
contain knowledge which is applicable exclusively to language learning in
other words, it is regarded as an adaptation which is both species- and task-
specic.
However, the idea of a language module or organ (cf. Anderson and
Lightfoot 2002; Chomsky 1975, 2000), though it may initially seem attractive,
becomes considerably less appealing when we consider the neuroanatomical
DABROWSKA text Make up 26/3/04 4:34 pm Page 59
Biological underpinnings 59
evidence. One reason for this is that so far, despite intensive research, no one
has been able to nd this module. Anatomically, the human brain is not very
dierent from that of the great apes. It is considerably larger (about three times
the size of that of a chimpanzee), and it diers in the relative proportions in
that humans have more association cortex and a larger cerebellum. However,
there is no evidence that humans evolved new cortical areas or structures not
present in the apes (Elman et al. 1996; Mller 1992, 1996; Preuss 2000). The
so-called language areas (Brocas and Wernickes areas) have homologues in
ape brains (Preuss 2000) and are involved in other cognitive operations as well
(Fiez et al. 1995; Fox et al. 1988; Mazziotta et al. 1982).
One might object that the presence of homologous areas in ape brains does
not entail that the details of the microcircuitry are the same: it is possible that
in the course of human evolution, our brains have been rewired to encode
linguistic knowledge (and presumably knowledge underlying other uniquely
human abilities). However, there are several reasons for doubting that the
genome species the details of cortical connections, as opposed to the general
pattern.
First, there is an arithmetical problem which Mller (1996: 626) has dubbed
the problem of the poverty of the genetic code. The human brain contains
approximately 100 billion neurons. Assuming that each neuron is connected
to 1000 other neurons (a rather conservative estimate), we have 1014 (i.e.,
100,000,000,000,000) synapses. The human genome, however which encodes
information necessary for building the rest of the body as well contains only
about 24,000 genes (How many genes 2003), about 99 per cent of which are
shared with chimpanzees (Ebersberger et al. 2002). The disparity between
these two gures suggests that the details of cortical microcircuitry are prob-
ably not fully specied in the genome, but rather emerge through complex
interactions of genetic and environmental factors (cf. Elman et al. 1996).
Secondly, there is considerable evidence that cortical tissue is quite plastic
in that it can support a variety of dierent representations, depending on the
input it receives (Deacon 1997, 2000; Elman et al. 1996; Johnson 1998). We
have already seen in Chapter 4 that if the regions which are normally respon-
sible for linguistic processing are damaged early in infancy, language usually
develops in other parts of the brain. Similar reorganisations have also been
observed for other cognitive functions in many species. When a particular area
is deprived of its usual input, it may assume another function. For example, if
a monkeys nger is amputated, the area which would normally receive input
from it will be colonised by input from adjacent ngers; or parts of the audi-
tory cortex may take over visual processing if they do not receive input from
the ears. Such reassignment of function is most pronounced if the damage
takes place early in infancy, but also occurs to some degree in mature individ-
uals. Furthermore, when a living brain is surgically rewired so that a partic-
ular region receives input from a dierent source so that the auditory cortex
receives information from the eyes, for example the new recipient region
DABROWSKA text Make up 26/3/04 4:34 pm Page 60
develops some characteristics of the normal recipient region (in our example,
the auditory cortex takes on visual representations). But perhaps the most dra-
matic evidence for plasticity has come from experiments involving transplan-
tation of fetal neural tissue to a dierent area of the brain in another animal
of the same species, or even a dierent species. Amazingly, the transplanted
neurons grow appropriate connections and become functionally integrated
with the host neurons (Deacon 1997, 2000), demonstrating that the genetic
program for growing the brain is not only exible, but also quite similar across
species. For obvious ethical reasons, such experiments have not been per-
formed on humans, but research on brain reorganisation in amputees and con-
genitally deaf or blind individuals is generally consistent with the animal data.
Moreover, there have been several apparently successful attempts to trans-
plant fetal pig neural cells into the brains of human patients suering from
Parkinsons disease in an eort to improve their motor control. A post-
mortem carried out after one such patient died revealed that the pig cells had
survived and had grown appropriate connections in the host brain (Deacon et
al. 1997).
It must be stressed that there are limits to this plasticity (cf. Johnson 1993,
1998). While it is possible for a lesioned brain to reorganise itself, particularly
if the lesion occurred early, this is by no means always the case: the results of
brain damage can be catastrophic. Similarly, while it is possible to transplant
fetal neural tissue to a dierent site in another brain, it is certainly not pos-
sible to transplant anything anywhere. Furthermore, we do not know how well
the transplanted tissue or rewired brains work. Although they do seem to take
on representations appropriate to the input they receive, they may not do so
as eciently as an intact brain would. Thus, the fact that brain tissue is fairly
plastic in the sense that it can take on a variety of representations does not rule
out a certain degree of specialisation: it is perfectly possible that some regions
are better suited for certain tasks than others. Even a fairly modest degree of
plasticity, however, is incompatible with the view that specic regions are pre-
destined to support a particular function because they encode the information
necessary for that function.
Thus, it seems that the genetic wiring diagrams for the brain are fairly
crude: fetal axons know the general area where they should terminate, but
not the specic cell. Patterns of connectivity which can support complex com-
putations such as language processing emerge later, largely as a result of
experience (cf. Elman et al. 1996; Johnson 1998). In what sense, then, can lan-
guage be said to be innate? To answer this question, it will be useful to appeal
to a taxonomy of ways to be innate oered by Elman et al. (1996; see also
Elman 1999). Elman et al. distinguish between three types of innateness:
Biological underpinnings 61
The evidence for cortical plasticity reviewed above argues against innately
specied cortical representations.1 This suggests that the innate biases that
construct a brain capable of acquiring language operate at the other two levels.
Architectural and chronotopic constraints require less genetically specied
information, so the small number of genes in the genome (relative to the
number of synapses in the brain) is not a problem. They operate indirectly, by
enabling certain processes and making certain outcomes highly probable, but
they also allow for variations after injury or when environmental conditions
are dierent. However, architectural and chronotopic constraints are not spe-
cic enough to encode details of linguistic knowledge: instead, they build a
brain that can perform certain types of computations and support certain
types of learning.
It follows that human adaptations to language involve innate information-
processing and learning biases rather than innate knowledge. We now turn to
the question of what these adaptations might be, and how they might have
evolved.
2.1 Preadaptations
Our evolutionary history saw two periods of relatively rapid increase in brain
size. During the rst of these, which occurred between 2 million and 1.5
million years ago, the brains of our ancestors doubled in size from approxi-
mately 450 to 900 cubic centimetres, and during the second, which began
about half a million years ago and lasted until about 300,000 years ago, they
reached their present size of approximately 1350 cc. Both of these periods of
brain expansion appear to have coincided with changes in the upper respira-
tory tract which would have enabled our ancestors to produce a wider range
of sounds (Lock and Peters 1996), so it is likely that they correspond to major
developments in the evolution of language, perhaps the emergence of sym-
bolic communication and combinatorial syntax respectively.
However we interpret the fossil record, it is clear that language is a relatively
DABROWSKA text Make up 26/3/04 4:34 pm Page 62
All of these abilities are particularly well developed in humans, but they are
also found in apes (Lock and Colombo 1996; Premack 1980), so they can be
regarded as preadaptations for language. Since they are clearly useful for lan-
guage, it is reasonable to assume that, once humans acquired a rudimentary
linguistic system, they were selected for, and rened as a result: the vocal tract
became increasingly suitable for speech production, we became better at co-
ordinating motor routines, acquired more sophisticated conceptual abilities
and a better memory (especially for cross-modal associations), and so on.
However, before any of this could happen, our ancestors had to acquire an
ability without which it is impossible to have language as we know it: the
capacity for cultural learning.
Biological underpinnings 63
Biological underpinnings 65
cal sense. The propensity for faithful imitation ensures that all individuals in a
community acquire (more or less) the same signs. Our mind-reading skills, on
the other hand, ensure that we attach roughly the same meanings to these
signs.3 Both of these abilities are a necessary prerequisite for language.
Additional evidence for a possible link between motor planning and lan-
guage comes from comparative studies of humans and other primates. All pri-
mates are able to learn sequences of actions or stimuli, which suggests that it
is an ability that we inherited from our common ancestors. However, a more
detailed analysis of the performance of humans and modern-day non-human
primates reveals that there are some subtle dierences in how they approach
the task: specically, humans plan the whole complex sequence before execut-
ing it, whereas non-human primates appear to program each stage indepen-
dently (although chimpanzees show some ability to plan ahead; see Conway
and Christiansen 2001 for a discussion).
The ability to program and execute a whole sequence of actions as one unit
may have developed in humans as an adaptation to throwing (Calvin 1983,
1993). Accurate throwing is useful for chasing away predators and for hunting.
However, it is quite a demanding task, in that it requires the coordination of
the action of many muscles and very precise timing. Furthermore, because it
is a very brief action, there isnt enough time for the brain to respond to feed-
back from the moving arm; therefore, unlike slower actions such as climbing
a tree to reach a piece of fruit, the whole motor sequence for throwing must
be planned before the launching phase is initiated. As Calvin put it:
your arm is an unguided missile shortly after the throw has begun. You
must plan perfectly as you get set to throw, create a chain of muscle
commands, all ready to be executed in exactly the right order . . . You
need something like a serial buer memory in which to load up all the
muscle commands in the right order and with the right timing relative to
one another. (1993: 234)
Calvin also notes that because neurons are rather inaccurate time-keepers, in
order to achieve the precise timing necessary to hit the target at any distance,
the planning mechanism must average the output of a very large number of
neurons, and hence it is possible that the advantages of accurate throwing
might have driven the selection for a bigger brain in our ancestors.
Whatever its evolutionary origins, the ability to package a series of actions
as a single unit allows more elaborate planning, since the complex unit itself
can be incorporated into a larger action plan. It could also be exapted for lin-
guistic use: a sequence of words functioning as a single unit that is to say, a
phrase can be incorporated into a larger unit, thus giving rise to hierarchi-
cal syntax. It is interesting to note, then, that while the giant apes are able to
learn hierarchically structured action sequences, their ability to do so is quite
limited similar to that of a 2-year-old human child (Conway and
Christiansen 2001).6 This appears to parallel the upper limit of their linguis-
tic abilities. Attempts to teach language-like systems to captive apes have
shown that they can learn simple ordering rules, but they do not seem to be
able to combine phrases hierarchically.
DABROWSKA text Make up 26/3/04 4:34 pm Page 67
Biological underpinnings 67
While it is undeniable that in the course of human evolution our brains and
vocal apparatus adapted to the exigencies of linguistic communication, we
should not lose sight of the fact that languages, too, are products of evolution.
The emergence of language is a process in which individual speakers make
innovations which either do or do not spread through the community. Whether
or not a particular variant will spread depends at least in part on its commu-
nicative tness. Other things being equal, speakers naturally prefer structures
which are easier to produce; and through failure to communicate, they learn to
avoid structures which are dicult to understand.7 This means that when there
is more than one structure expressing a particular function, the structure which
is more dicult to process will be used less frequently. The next generation of
speakers will have the same processing biases, but also fewer opportunities to
learn the more dicult structure. This will result in a further drop in frequency
of the more dicult variant, until it eventually disappears.
In other words, the story of the emergence of language is one of co-evolu-
tion: humans adapted to language, but language (or rather languages) also
adapted to humans. Indeed, there are reasons to think that languages did most
of the evolving. As Deacon (1997) points out, languages need humans more
than humans need languages (if it came to the crunch, we could probably
survive without them; but they need our brains to survive and reproduce) so
the selectional pressures on languages were stronger. Furthermore, cultural
change is several orders of magnitude faster than biological change. Any sub-
stantial restructuring of the brain requires hundreds of thousands or even mil-
lions of years. Grammars, on the other hand, can change radically in a few
centuries, and lexical innovations spread in a matter of years and, in this era
of mass media, sometimes in a matter of months or even weeks.
One particularly good example of language adjusting to humans is the phe-
nomenon of categorical perception of speech sounds. When presented with a
range of stimuli which dier along some continuous parameter, such as voice-
onset time (VOT), people tend to perceive sounds up to a certain point on the
continuum as belonging to one category and sounds that fall beyond that
point as belonging to another category. For example, stops with a VOT of less
than 35 ms are perceived as voiced, while those with a greater VOT are per-
ceived as voiceless. Thus, an articially produced sound that is intermediate
between a prototypical /t/ and a prototypical /d/ is not perceived as something
in between the two phonemes, but as either one or the other, depending on
whether the VOT is less than or more than 35 ms. We know that at least some
such perceptual biases are innate because they are present in very young
infants, even when the ambient language does not make use of a particular
phonemic contrast.
Our tendency to perceive speech sounds categorically is useful from a
communicative point of view, since it allows ecient coding of phonemic
DABROWSKA text Make up 26/3/04 4:34 pm Page 68
Biological underpinnings 69
guages are hard to nd, and they tend to be rather uninteresting: they are
statements like all languages have vowels and consonants, all languages have
nouns (words prototypically referring to things) and relational words, all lan-
guages have recursion. The more meaty universals are either statistical (most
languages have property X) or implicational (if a language has property X,
then it also has property Y). This seems more compatible with the view that
universals are a result of languages converging on certain solutions than with
the view that universals are innate constraints dening a possible human lan-
guage.
The fact that restrictions on movement cannot be learned (if indeed this is a
fact) does not entail that they must therefore be innate: it could also mean that
we need a better linguistic theory one which does not require movement. Of
course, linguistic theory still needs to account for the dependency phenomena
that movement rules were postulated to explain; the point here is that the
alleged unlearnability of a particular grammatical rule or principle cannot be
taken as evidence for the innateness of Universal Grammar unless we have
independent evidence for the psychological reality of that rule or principle.
Biological underpinnings 71
This, Chomsky asserts, occurs in spite of the fact that utterances that would
enable them to distinguish between the two hypotheses are so rare that a
person might go through much or all of his life without ever having been
exposed to the relevant evidence (Chomsky 1980a: 40). However, the asser-
tion is not supported by any data on how frequent or infrequent the relevant
utterances actually are, either here or in the many other subsequent rehash-
ings of the argument. Other linguists have taken the trouble to look at the
corpus data, and found that relevant examples constitute about 1 per cent of
all interrogative utterances in a wide range of corpora (Pullum and Scholz
2002). Now 1 per cent may not seem like a lot but since children are exposed
to vast amounts of input (cf. Chapters 2 and 3), it is 1 per cent of a very large
number: Pullum and Scholz estimate that by age 3, a typical child will have
heard between 7,500 and 22,500 relevant instances.9
The tendency to assert that some aspect of linguistic knowledge could not
possibly be learned from the input, without providing any supporting evi-
dence, is endemic in the nativist literature (see Pullum and Scholz 2002 for
further examples). This seriously undermines the poverty-of-the-stimulus
argument, since there is accumulating evidence that the information available
in the input is richer than previously thought (Redington and Chater 1998).
For example, lexical co-occurrence patterns can be used to classify words into
grammatical categories, and prosodic phrasing provides clues about constit-
uency. Also, in spite of frequent claims to the contrary, it appears that chil-
dren receive plenty of negative evidence, that is to say, evidence that an
utterance they have just produced is ill-formed by adult standards (see
Chouinard and Clark 2003 for an up-to-date discussion). It is true that some
kinds of knowledge (e.g. properties of the various kinds of empty categories)
cannot be acquired from the input available to children. But again, this does
not necessarily mean that the child must have innate knowledge of these: an
alternative conclusion would be that we need to look for ways of accounting
for the same phenomena which do not postulate such invisible and inaudible
constructs.
Biological underpinnings 73
Posing the question in this way is just as unfair to the nativists as Bickertons
original question was to the opposite camp. No sane nativist believes that the
dierence in meaning exemplied in Bickertons sentences is directly encoded
in our genes. But this raises the question of what actually is encoded: does the
human genome contain statements to the eect that a trace must be properly
governed? Does it contain instructions for nding traces? What form do these
take? The point, of course, is that proclaiming that some aspect of our linguis-
tic knowledge is innate does not answer the question of how we come to have
it it merely raises three new ones:
As we will see shortly, all of these questions have yet to receive a satisfactory
answer.
4.2.1 Evolution
For someone who attaches so much importance to innate linguistic knowl-
edge, Chomsky is remarkably reticent about how the capacity for language
evolved. In 1980, he side-steps the issue by stating that:
These skills [i.e., the ability to acquire language] may well have arisen as
a concomitant of structural properties of the brain that developed for
other reasons. Suppose that there was selection for bigger brains, more
cortical surface, hemispheric specialisation for analytic processing, or
many other structural properties that can be imagined. The brain that
evolved might well have all sorts of special properties that are not indi-
vidually selected; there would be no miracle in this, but only the normal
workings of evolution. We have no idea, at present, how physical laws
apply when 1010 neurons are placed in an object the size of a basketball,
under the special conditions that arose during human evolution.
(Chomsky 1980b: 321)
DABROWSKA text Make up 26/3/04 4:34 pm Page 74
Other researchers (e.g. Calvin and Bickerton 2000; Jackendo 2002; Pinker
and Bloom 1990) have been considerably more explicit. However, although the
scenarios they have proposed are not unreasonable, they are still very sketchy,
mutually incompatible, and, worst of all, unveriable. This is unavoidable:
since languages do not fossilise, accounts of language evolution are necessar-
ily little more than scientic just-so stories.
Biological underpinnings 75
Thus, while something about the normal version of human FOXP2 seems to
help an individual to develop language, the gene itself does not contain a blue-
print for grammar. FOXP2 is a transcription factor, or a gene which regulates
the functioning of other genes. It is possible that these other genes are respon-
sible for encoding the principles of Universal Grammar into the developing
brain but if so, they have not been found yet. Another possibility is that FOXP2
exerts its inuence in a much more roundabout way, for example by controlling
the amount of brain tissue available for sequencing ne facial movements.
5. Conclusion
et al. 2002; Lieberman 1991, 1998; Tomasello 1999; Worden 1998), such as
producing and recognising sounds, planning motor sequences, categorisation,
storage and retrieval of information, predicting the actions of other individ-
uals, and cultural learning. Many researchers maintain that we also have some
language-specic adaptations, although there is considerable disagreement
about what these are and how they may have evolved. However, it should be
clear that, for reasons discussed earlier, a good theory of language will keep
such stipulations to a bare minimum. Furthermore, they should be treated as
tentative hypotheses to be conrmed or rejected by biological research rather
than as central tenets of the theory.
Notes
Biological underpinnings 77
Part II
The Building Blocks of Language
DABROWSKA text Make up 26/3/04 4:34 pm Page 80
DABROWSKA text Make up 26/3/04 4:34 pm Page 81
7 Words
The standard argument for the innateness of syntax goes something like this:
children acquire the grammar of their native language remarkably quickly, on
minimal exposure, and with no overt teaching. Moreover, they go through the
same developmental stages and make relatively few errors. Most importantly,
some aspects of language (e.g. knowledge about empty categories, constraints
on movement) are acquired without any experience. Since no known learning
theory could begin to explain how this happens, linguistic knowledge must be
innate.
With this argument in mind, consider the following facts about lexical devel-
opment:
1. In early and middle childhood, children learn about ten new words a day
approximately one in every waking hour (Carey 1978; Miller 1986; Miller
and Gildea 1987).
2. Words are often learned after a few presentations in ambiguous conditions,
and sometimes after only one presentation (Carey 1978; Dickinson 1984;
Heibeck and Markman 1987; Markson and Bloom 1997).
3. Most words are learned without overt teaching, and attempts to teach
vocabulary at school sometimes lead to very strange results.1
4. Lexical development exhibits some striking regularities. For example,
colour terms are acquired at a specic maturational stage; and dimensional
adjectives and other spatial terms are acquired in a fairly uniform order.
5. Most words are acquired without eort and almost without errors, in spite
of the fact that lexical meanings are often very dicult to pin down. (Many
dictionary denitions are grossly inadequate, and even philosophers nd it
dicult to explicate our tacit knowledge of everyday concepts.)
6. Children can acquire words from domains of which they have no experi-
ence. For example, blind children have been shown to acquire visual terms
without much diculty (Landau and Gleitman 1985).
Applying the same logic to these facts, the conclusion seems inescapable:
lexical knowledge, or at least a signicant part of it, must be innate. In fact,
this is precisely the conclusion drawn by Chomsky:
DABROWSKA text Make up 26/3/04 4:34 pm Page 82
Barring miracles, this means that the concepts must be essentially avail-
able prior to experience, in something like their full intricacy. Children
must be basically acquiring labels for concepts they already have.
(Chomsky 1991: 29)
The process of learning just picks them [i.e. innate lexical meanings] out
en bloc, once the phonetic label is assigned to the concept. Association
now becomes a very minor component of the mechanism: What is being
associated is just the sound for that concept and the innately available
concept itself . . . Association is marginal to the process: it boils down
to a heavily constrained pairing of certain labels to certain innate con-
cepts. (Piattelli-Palmarini 1989: 31)
The problem with this conclusion, impeccable as the logic may be, is that it is
patently absurd, since it commits us to the belief that humans are born with
innate concepts such as inning, poppadom, widget, aerobics, Internet, baptise
and overdraft. While such a position is logically consistent, I suspect that
anyone who has ever tried to explain what an inning is to someone with no
experience of baseball or cricket, or what an overdraft is to someone from a
culture that does not use banks, will be rather sceptical.
Furthermore, as Gleitman (1990) convincingly demonstrates, matching the
concept and the phonological label is not as trivial a task as Piatteli-Palmarini
makes it out to be. Determining which aspect of the world a particular part of
a parental utterance refers to is dicult enough even when it refers to some-
thing in the childs immediate environment; but most parental speech is not a
running commentary on situations observed by the child. Utterances such as
Eat your peas, Lets go to the park or It rained last night do not co-occur with
episodes of pea-eating, going to the park, and rainy weather; and episodes of
thinking, knowing, understanding, hoping, wondering, guessing and suppos-
ing all look quite similar. More damaging still, asserting that lexical concepts
are innate hasnt quite eliminated the miraculous: it has merely shifted it to
another domain. We still need to explain how the species acquired these innate
abilities (cf. Chapter 6, section 4.2). A moments reection will reveal that they
could not have come into being through any known evolutionary process
unless thousands of years ago evolution knew that someday we were going
to invent baseball, banks, and tness clubs. Hilary Putnam makes a similar
point:
Words 83
Moreover, they would also have been useful to our prehuman ancestors, so
evolution would have had plenty of time to operate.
Locative expressions are also comparatively easy to study because they have
fairly concrete meanings, so there is a large body of research on their mean-
ings and their acquisition. These can be interpreted in the context of research
on infant spatial cognition and infant perceptual processing mechanisms and
their neurophysiological bases.
Moreover, it is possible that childrens knowledge of spatial terms helps them
to learn other aspects of their language. It is well known that spatial terms are
often the starting points of grammaticalisation processes which eventually
produce tense, aspect and case markers, complementizers, comparatives,
morphemes expressing causal relations, and other grammatical functors (see
Casad 1996a; Comrie 1976; Dabrowska 1996; Heine 1997; Rudzka-Ostyn 1985;
Traugott 1978). This suggests a conceptual anity between spatial terms and
various grammatical notions, and in fact some linguists have argued that gram-
matical notions should be interpreted in localistic terms even when there is no
historical evidence that they derive from spatial concepts (Anderson 1971, 1994;
Hjelmslev 1935; Jackendo 1983; Miller 1972, 1974). If such conceptual an-
ities do exist, children could use their knowledge of spatial concepts to get a
handle on the more abstract notions encoded by various grammatical functors.
The conventional wisdom on the acquisition of locative terms is that it
involves learning how to apply them to . . . prior knowledge about space
(Clark 1973: 62; see also Johnston and Slobin 1979; Levine and Carey 1982;
Slobin 1973) that is to say, to prelinguistic concepts of containment, support,
occlusion etc. (In this context, prelinguistic doesnt necessarily mean innate:
the concepts could emerge out of prelinguistic sensorimotor experience.)
There is good supporting evidence for this view:
1. Locative expressions are acquired very early. The rst terms usually appear
at 1617 months, and they have been reported as early as 1214 months
(Choi and Bowerman 1991: 100); by the time the child goes to school, the
system is rmly in place.
2. Children acquiring dierent languages acquire equivalent words in roughly
the same order. The order of acquisition is determined by non-linguistic
knowledge, though acquisition may be delayed if the linguistic form is di-
cult (Johnston and Slobin 1979).
3. Errors in the use of basic spatial terms are extremely rare in spontaneous
speech (Choi and Bowerman 1991; Sinha et al. 1994; Tomasello 1987).
4. Children generalise newly acquired concepts to new contexts very quickly,
often within days of the rst use (Choi and Bowerman 1991; Tomasello
1987).
In spite of this, I will argue that learning locative terms does not consist simply
in matching linguistic labels to pre-existing concepts. The main reason for this
DABROWSKA text Make up 26/3/04 4:34 pm Page 85
Words 85
Figure 7.1 Is the banana in the bowl? (adapted from Herskovits 1985: 19)
DABROWSKA text Make up 26/3/04 4:34 pm Page 86
inside region: compare There is a banana in the bowl and There is a crack in the
bowl. Or consider the expression The bird is in the tree: in this case, all the
boundaries of the container are virtual. (All three examples are taken from
Herskovits 1985, 1986.)
Making words t the world also involves a fair amount of idealisation, as
objects need to be assimilated to the abstract geometric descriptions embod-
ied in the semantics of locative expressions. Thus a frying pan might be ideal-
ised as a container (cf. the expression in the frying pan); a river as a line (along
the river) and a table as a horizontal surface (on/under the table) (Dabrowska
1993; Herskovits 1985; Talmy 1983).
The crucial point here is that all of these are conventional, and hence lin-
guistic, phenomena. This can be easily demonstrated by showing that dier-
ent languages require dierent construals. In English, a bird is in the tree, but
many other languages require dierent terms (usually the equivalent of on;
notice that even in English, an apple would be on, not in, the tree, in spite of
the fact that the relationship between the apple and the outer boundary of
the tree is exactly the same as that between the bird and the outer boundary
of the tree). In English, the inside of a bowl can be either the region dened
by the wall, the bottom and the virtual lid (of whatever shape) or the bowl
itself, but many languages allow only the former conceptualisation. Hence, if
you want to say There is a crack in the bowl, you either have to use another
preposition (e.g. the equivalent of on) or a non-spatial paraphrase (something
like The bowl is cracked or The bowl has a crack). Furthermore, dierent lan-
guages require dierent idealisations of the same objects. English idealises a
frying pan as a container, so your fried eggs are in the pan; in Polish, they
would be on the frying pan (na patelni). If a ball rolls between the legs of a
table and emerges on the other side, an English speaker will say that it rolled
under the table: English requires that the table be construed as a horizontal
surface. In contrast, in Atsugewi, the ball would roll through the table because
the table is thought of as a three-dimensional object (Talmy 1983). Sometimes
even dierent dialects of the same language require dierent conventional
construals. In Britain, you leave your car in the street. In America you leave it
on the street, but your neighbours children playing near it would be in the
street; and if you decided that the driveway would be a safer place to park,
your car would in the driveway, not on it.
As a nal illustration, let us consider an extended example from Cora, an
Uto-Aztecan language spoken in a mountainous region of Mexico. Cora has
an extremely rich system for encoding spatial relationships which includes
spatial adverbs, particles, and verbal prexes which are used in various com-
binations to convey elaborate descriptions of spatial congurations. Cora
locatives have been extensively studied by Eugene Casad (1977, 1982, 1988),
and the reader is referred to these works for a comprehensive description.
Because the system is so rich, it cannot be described in any detail here; but its
very richness and utter foreignness are an excellent illustration of the central
DABROWSKA text Make up 26/3/04 4:34 pm Page 87
Words 87
theme of this chapter, namely, that humans possess complex abstract systems
which are extremely unlikely to come prewired.
The following discussion is based on data from Casad and Langacker
(1985), which describes the semantics of just two morphemes, u (which can be
glossed as inside) and a (outside). Because of space limitations, I cannot
justify the analysis proposed by the authors. However, it should be stressed
that Casad and Langacker do not ask the reader to take the analysis on faith;
as demonstrated throughout the paper, various quirks and idiosyncrasies of
Cora locationals make perfect sense given the construal described here.
Although a means outside, it can also be used to refer to location inside a
shallow container such as a basin or pan (as opposed to, say, a jar, which
requires u). In other words, as far as the u/a distinction is concerned, shallow
containers do not count as containers: any object located near the outside
surface of the landmark is considered outside the landmark.
As in the case of the English in and out, the container with respect to which
things are located can be virtual. Consider the complex expressions mah
away up there to the side in the face of the slope and muh away up there near
the centre line of the face of the slope. Both of these locationals consists of
three morphemes. The rst morpheme, m, is a deictic prex which can be
glossed as at a medial distance from the speaker (in contrast to other pre-
xes meaning near the speaker and far from the speaker). The nal h means
on the face of the slope and contrasts with suxes meaning at the foot of
the slope and at the top of the slope. Between the two come u and a, which
mean inside and outside but inside and outside what? Casad and
Langacker argue that the landmark region is dened relative to the line of
sight:
MED
VP
Figure 7.2 mah away up there to the side in the face of the slope (adapted from
Casad and Langacker 1985: 262)
Both sentences can be used to describe the same objective situation: namely,
to state that the dogs tail has been lopped o. But what does a short tail have
to do with slopes and containment, and why does the sentence with a mean
the same thing as the sentence with u?
In fact, they dont mean quite the same thing, as they would be appropriate
in dierent contexts. (1a) would be used to describe a situation in which the
dog is viewed from behind, as in Figure 7.3a, while (1b) describes a dog viewed
from the side, as in Figure 7.3b. The motivation for using a and u should now
be apparent. In (1a), the slope facing the observer is the dogs rump, and the
tail is inside the region dened with respect to the observers line of sight. In
(1b), the slope is the dogs side, and the tail sticks out outside the region
dened by the line of sight.2 Notice that this usage requires a (metaphorical)
projection of a term normally used to refer to parts of a mountain onto a body
part, motivated by the topographical similarity of a slope and an animals
rump and conventionalised in the grammar.3 While the Cora u may be said to
expresses containment, containment is very much in the eye of the beholder.
DABROWSKA text Make up 26/3/04 4:34 pm Page 89
Words 89
1.2 Polysemy
Most locative expressions are massively polysemous. For example, according
to Brugman (1988; see also Lako 1987) the English preposition over has
nearly 100 distinct senses, of which 24 designate various spatial relationships.
A few of these are exemplied in (2) below.
Figure 7.4 Some in, on, and above relationships: key in a door, coin in a box,
crack in a bowl, bird in a tree, apple on a tree, picture on a wall, plaster (Band-Aid)
on a leg, fly on a window, lamp above a table, lamp on a table
DABROWSKA text Make up 26/3/04 4:34 pm Page 91
Words 91
codes them by means of three prepositions, in, on and above. Some other lan-
guages (e.g. Greek and Spanish) conate and and code
separately. Kikuyu also has only a two-way distinction, but it con-
ates the prototypical on (as in The lamp is on the table) and above, and
expresses everything else with a generalised locative morpheme. Italian,
Portuguese, Swedish, Indonesian, Mandarin and Finnish all make a three-way
distinction, but each carves up the semantic space in a slightly dierent way.
Dutch and German divide the set of relationships into four categories, and
Japanese into eight.
It is clear, then, that having innate categories corresponding to specic loca-
tive lexemes would be of little use to learners, since the detective work involved
in determining which categories out of the thousands of possible ones were
actually employed in their language would not be much easier than creating
them on the basis of the evidence.
On the other hand, innate concepts need not coincide with specic lexemes.
One could claim, for example, that what is innate is category prototypes
(archetypal congurations around which future categories will grow) or some
more basic dimensions along which lexemes dier (semantic features). I will
discuss the latter possibility later on. The former may initially seem quite
attractive, since there is considerably less variation between languages in the
congurations that they choose for category centres. For example, although
the English preposition on, the Polish na and the French sur all designate
dierent categories, the prototypical sense (a trajector object located on top
of, and supported by, a horizontal landmark) is the same in all three languages.
However, a moments reection will reveal that this theory suers from exactly
the same problems as the more radical view espoused by Chomsky, Fodor and
Piattelli-Palmarini. As the discussion below will make clear, the number of
distinct concepts that would have to be encoded in the genome would still be
intractably large; and the learner would still need to determine category boun-
daries.
1. Verbs which code path of motion, e.g. hairu go in, deru go out, ireru put
in, dasu take out, agaru go up, oriru go down. Functionally, this is the
DABROWSKA text Make up 26/3/04 4:34 pm Page 92
This system makes it possible to code aspects of the situation in several parts
of the sentence (often redundantly):
The rst sentence literally means something like The book, having entered the
box, is (there): the verb hairu conates the idea of motion and the path of
motion (from outside inwards). The semantic contribution of the particle ni
is entirely redundant: it does not add any information to that expressed by
haitte imasu. The same information can be expressed using (4), which is even
more redundant, since the locative information is distributed across three
morphemes: the particle ni at, the relational noun naka inside and the verb
hairu enter.
Other languages have locative systems which dier even more from those
found in the familiar European languages. To appreciate the cross-linguistic
diversity, in this section we will look at a few rather exotic systems: the Mixtec
system of body-part terms, Tzeltal dispositional adjectives, and absolute
systems in several languages.
Words 93
Body-part terms are used to refer not only to the relevant part of the land-
mark object, but also to the neighbouring region in space: at mountains head
could mean over the mountain as well as on top of the mountain. Some of
the terms in the system designate human body parts, some animal body parts,
and some can refer to either, which results in rather strange (to us) patterns of
polysemy. For example, the term ci`i` belly can be used to refer to the middle
of the front surface of a vertically oriented object (the region where the belly
would be in an upright human) or the underside of a horizontally oriented
object (where the belly would be in a four-legged animal), or the vicinity of
either of these. Thus, with a horizontal landmark such as a table, ci`i` can be
translated as under (9).
It should be obvious from this brief description that English and Mixtec
carve up spatial scenes in very dierent ways, and, consequently, the informa-
tion conveyed by a Mixtec sentence and its English translation is not strictly
equivalent. English is quite explicit about the position of the trajector with
respect to the landmark (contact/no contact) but does not specify the shape of
the landmark, while the opposite is true in Mixtec.
On second thoughts, the use of body-part terms to express locative mean-
ings is not as exotic as it might at rst seem. The American English expression
in back of is also a metaphorical extension of a body-part term, and the same
metaphor is (just about) accessible, though much less transparent, in the prep-
osition behind. The word front in the complex preposition in front of comes
from Latin frons forehead, though in this case the link is purely etymological
and unknown to most speakers. In fact, body-part terms are common sources
DABROWSKA text Make up 26/3/04 4:34 pm Page 94
of locative expressions (see Svorou 1994 and the extensive literature cited
there). What is special about Mixtec is that the use of body-part terms is much
more systematic and more transparent synchronically.
However, the extension of body-part terminology to the locative domain is
conventional: that is to say, although there are broad similarities between lan-
guages in the way they eect the mapping, the details are language-specic and
must be learned. For example, the words for eye, face, forehead, head, breast
and chest are all common sources for the spatial meaning of in front of
(Svorou 1994); but dierent languages make dierent choices from this list. In
this connection, it is also instructive to compare Mixtec and Cora: while
Mixtec projects body-part terms onto inanimate objects such as mountains,5
Cora sometimes projects hilly terms onto parts of the body.
Words 95
object does not have an inherent front, the position of the trajector is
described relative to the location of an observer, usually the speaker. Thus,
depending on where the observer is, the same situation can be truthfully
described by saying The mouse is in front of/behind/to the left of/to the right of
the sh (cf. Figure 7.8).
English also uses an absolute system of geographical terms (north, south,
east and west) to describe macrolocations, or locations of large gures such as
towns and villages with respect to large reference objects. When I say
Doncaster is east of Sheeld, I am using the absolute system.
Some languages use an absolute system to describe microlocations as well.
In such languages, the conguration represented in Figure 7.8 will be
DABROWSKA text Make up 26/3/04 4:34 pm Page 97
Words 97
described by saying something like The mouse is east of the sh. Note that
because the coordinates of the system are xed rather then being projected
from the speaker, the mouse is always to the east of the sh, regardless of the
position of the speaker/observer.
Absolute systems sometimes use the four compass points, and sometimes
other cardinal directions dened by the direction of the prevailing winds
(Guugu Yimidhirr) or some feature of local topology such as a large river
(Yupno) or the sea (many Oceanian languages). In Tzeltal, the relevant topo-
logical feature is the general inclination of the land, which slopes down
towards the north, so anything south of the reference object is said to be
uphill and anything towards the north is downhill. (This is just the overall
direction of the inclination: there are many local irregularities, but the general
direction of the slope xes the system of coordinates. See Brown and Levinson
1993 for further details.) These two opposing directions dene the strong axis
of the system. The system also has a weak axis perpendicular to the strong
axis which lacks opposing poles. Together, the two axes dene a three-way dis-
DABROWSKA text Make up 26/3/04 4:34 pm Page 99
Words 99
tinction: up v. down v. across the slope. In Figure 7.9, the saucer is uphill, the
spoon is downhill and the two coasters are both across the slope from the
glass.
It should be clear from the above discussion that absolute systems are not
only quite dierent from relative systems; they are also dierent from each
other in that they use dierent reference points and dierent numbers of car-
dinal directions.6 While east and west always refer to the same direction,
inland and seawards do not: if you live on an island, inland means east
on the western coast and west on the eastern coast.
down
up
Figure 7.9 The Tzeltal geocentric system of projective relationships
DABROWSKA text Make up 26/3/04 4:34 pm Page 100
For example, Jackendo (1983) notes that the English preposition under can
designate a place (The mouse is under the table), a path (The mouse ran under
the table The mouse traversed the region under the table) or the endpoint of
a path (The mouse ran under the table The mouse ran to the region under the
table), and proposes the following conceptual structures to capture the dis-
tinctions:
The expression up the hill (as in The house is up the hill) receives the following
analysis:
Words 101
2.1 Warlpiri
We will begin by looking at the acquisition of locative terms by Warlpiri chil-
dren. Warlpiri is interesting from our point of view because it does not have
terms corresponding to in, on and under. It does, of course, have terms which
can be used to talk about scenes which in English would be described using
these prepositions, but these are extended senses of expressions whose
primary meanings are rather dierent. The terms are kanunju-mparra down-
along or, in certain contexts, under; kaninja-rni down-hither or, with the
right kind of landmark, in; and kankarla-rra up-thither or on. All three
expressions are bimorphemic and consist of a locative stem and a directional
ax. It is not very clear what the exact semantic contribution of each mor-
pheme is the traditional glosses are not very helpful in this regard but it is
evident that Warlpiri spatial categories are dierent from those found in
English and other familiar languages.
Bavin (1990) conducted a simple experiment testing the comprehension of
locative terms by Warlpiri children aged from 3;11 to 8. The experimental task
involved placing an object as instructed by the experimenter in, on or under
a landmark object. The landmarks used were a toy yunta (a shelter consisting
of a at top with four poles for support), a toy yujuku (a shelter with sides,
open at the two ends), and a toy pick-up truck.
The results were quite intriguing. Kankarla-rra up-thither was fairly con-
sistently interpreted as on by all the children, but responses to instructions
with the other two terms were varied, and depended on properties of the ref-
erence object. Kanunju-mparra down-along was interpreted as under when
the reference object was a yunta; but when the landmark was a truck, 39 per
cent interpreted it to mean in the open back of the truck. Kaninja-rni down-
hither was interpreted as in when the landmark was a yujuku, but when the
landmark was a truck, the children gave inconsistent responses: some placed
the object in the cabin, some on the tray in the back and some under the truck.
In other words, there is no evidence in Bavins data that Warlpiri children are
sieving through the language input looking for labels for their prelinguistic
concepts , and . In fact there is no evidence that they make these
distinctions at all.7
2.2 Tzotzil
Tzotzil is a Mayan language, closely related to Tzeltal. Like Tzeltal, it uses geo-
centric coordinates dened by the general incline of the land, but in the region
DABROWSKA text Make up 26/3/04 4:34 pm Page 102
where Tzotzil is spoken the land slopes westwards, so uphill means east, not
south.
De Len (1994) found that Tzotzil caregivers use geocentric terms
(uphill/east and downhill/west) with children as young as 2, and children are
expected to understand the terms. (One mother in the study rebuked her
3-year-old son for not knowing where east was!) Children know the basic prin-
ciples of the system by around 4;6 (i.e. before European children reliably dis-
tinguish left and right), though they generally use local landmarks rather than
geocentric terms: that is, instead of saying The basket is uphill/east, they
might say something like The basket is towards the market town. They start
producing the terms for sunset and sunrise around 9, which suggests that
around this age the child shifts the local centre of the orientation grid to the
absolute co-ordinate provided by the sun (de Len 1994: 876).
Like Mixtec, Tzotzil also uses body-part terms to describe more immediate
locations. Preliminary work carried out by de Len (1991, cited in Levinson
1994) suggests that this system is acquired even earlier: Tzotzil children begin
to use body-part locatives well before the age of 3.
2.3 Japanese
In Japanese, we saw earlier, locative meanings are distributed across three form
classes: motion verbs, relational nouns and particles, with the rst of these
doing most of the work. Sinha et al. (1994) report on a longitudinal study of
one Japanese child, Adam, observed by the team from age 1;3 to 2;11.
Adams rst locative terms were motion verbs such as dasu take out, toru
take o, hameru t in etc. The rst recorded usage of a locative verb
occurred at 1;3.28; and they are recorded in every session from 1;6 onwards.
Locative nouns (soto outside, naka inside, ue top etc.) emerged shortly
afterwards, around the time of his second birthday. The last to appear were
the particles ni, de at, in, kara from, and e to. All four particles rst
emerged in collocation with koko this place, and their use remained highly
formulaic throughout the period of study. This pattern of acquisition makes
sense, since it is motion verbs and relational nouns that convey specic infor-
mation but it is rather dicult to explain if you believe that children are
simply looking for labels for pre-existing locative concepts corresponding to
in, on etc. If Adam were simply looking for labels for in and on, he should
have learned the locative particles much earlier, since they occurred fairly fre-
quently in the input (in fact, the frequency of locative particles in the speech
of Adams mother was four times that of locative relational nouns, which he
acquired earlier).
This acquisition pattern, incidentally, also argues against the idea that chil-
dren construct meanings out of a universal set of primitives. If this were the
case, more specic concepts should be more dicult to acquire than concepts
which directly correspond to the putative primitives. The data collected by
DABROWSKA text Make up 26/3/04 4:34 pm Page 103
Words 103
Sinha and his colleagues suggests the opposite: the verb hameru t in, which
co-lexicalises the action of inserting and the destination of the patient, is
acquired earlier than the general locative particle ni at, in; toru take o,
which co-lexicalises causative movement and direction, comes long before
kara from, and so on.
2.4 Korean
Another series of studies conducted by Melissa Bowerman and Soonja Choi
(Bowerman 1996; Choi and Bowerman 1991; Choi et al. 1999) compared the
acquisition of morphemes specifying the path of a moving object by English-
and Korean-speaking children. The relevant parts of the locative systems of
the two languages are quite dierent. English path morphemes are preposi-
tions, while in Korean the shape of the path is co-lexicalised with other notions
in the verb. Furthermore, English path morphemes are indierent as to
whether the movement is spontaneous or caused (the same preposition is used
with both transitive and intransitive verbs: go in, put in), while in Korean the
corresponding meanings are expressed by dierent verbs. Finally, Korean path-
incorporating verbs make distinctions which cross-cut those made by English
prepositions: e.g. kkita t a three-dimensional object into or onto another
object can be translated as put in (for example, when describing the action of
putting glasses inside a case), put on (when a cap is put on a pen), or put together
(when two Lego pieces are tted together). Conversely, the meanings expressed
by put in and put on are coded by a number of dierent verbs in Korean: kkita
when two three-dimensional objects t closely, nehta when the t is loose, pwu-
thita when two at surfaces are joined, kkotta when an elongated object is put
into/onto a base, and so on. In addition, there is another set of verbs for des-
ignating the action of putting clothing on various parts of the body.
Choi and Bowerman report that the rst locative expressions usually appear
between 14 and 16 months, and that children use them productively by 1620
months. The childrens use of locative expressions is generally correct, and it
follows language-specic patterns from the very beginning. English-speaking
children acquire their rst path morphemes (up, down, in, out, on, o ) very
early (usually between 16 and 17 months, and sometimes as early as 1214
months) and almost immediately generalise them to spontaneous and caused
motion and to changes of body posture. For example, they might say in when
putting something into a box or when they themselves are climbing into a box;
up when they want to be picked up, when going up stairs, and when standing
up. Korean children, in contrast, always use dierent verbs to refer to caused
motion and to spontaneous motion. Thus, they will say kkita when tting an
object into a hole, but they never overextend this term to describe a situation
when they themselves creep into a narrow space; and they never overgeneral-
ise ancta sit down and anta pick up and carry to situations involving upward
or downward motion.
DABROWSKA text Make up 26/3/04 4:34 pm Page 104
Words 105
these to me. We can put the things my uncle gave me away now.). This was fol-
lowed by a testing phase, which occurred immediately, after a weeks delay, or
a month later. In the testing phase, participants were shown an array of objects
and were asked to pick out the koba or the object given to the experimenter
by her uncle.
Markson and Bloom found that adults performed signicantly better than
children on the word-learning task when they were tested immediately after
training, but the dierences between age groups disappeared after a week.
About 70 per cent of the participants could still point to the correct object
when tested a month after training. This replicates the results obtained in
earlier fast-mapping studies. The new nding was that participants were
equally good at remembering which object was given to the experimenter by
her uncle; in fact, when tested immediately after the training phase, children
performed signicantly better on this task than on the word-learning task.8
Fast mapping, then, is not restricted to word learning.
On the other hand, Markson and Bloom also found that people are not
equally good at remembering any arbitrary fact about an object. In another
experiment, participants watched as the investigator placed a sticker on one of
the unfamiliar objects and heard a similar commentary (Watch where this
goes. This goes here [placing sticker on one of the objects]. Thats where this
goes.). In this version of the experiment, recall was signicantly poorer. It is
not clear whether this was due to the fact that the information was presented
visually or whether the location of a sticker was deemed less relevant by the
participants than the identity of an object. Whatever the reason, their ndings
suggest that fast mapping, though not unique to language, does not apply to
all memorisation tasks. Children are good learners, but not equally good at
learning anything. Something about the human cognitive apparatus makes
certain associations easier to acquire than others.
The general conclusion, then, is that children are good learners, but they are
not unbiased learners. In fact, many would argue that they are good learners
because they are biased towards certain solutions. It is important to realise that
rejecting the notion that lexical concepts are built in does not entail rejecting
the existence of innate structure of any kind. Clearly, children must have some
innate predispositions which make language learning possible. Learning is
impossible without experience, but an organism needs a great deal of innate
structure before it can have any experience, and even more to be able to extract
patterns from it. There is considerable empirical evidence for various con-
straints on lexical development in general (Dockrell and Campbell 1986;
Markman 1987, 1990; Smith et al. 1992; Tager-Flusberg 1986) and for prelin-
guistic biases on the perception of space which shape the acquisition of
DABROWSKA text Make up 26/3/04 4:34 pm Page 106
locative terms (E. V. Clark 1973a, 1976, 1980; H. Clark 1973; Huttenlocher et
al. 1983). The question, then, is not whether children are endowed with innate
structure, but rather what form their innate predispositions take, and it is to
this question that we now turn.
Words 107
Most cognitive scientists would argue that it is not. The following quotation
from Fodor et al. (1980) sums up the prevailing views on the matter quite well:
Its thus important to emphasize that the Empiricist version of TSP [The
Standard Picture, i.e. semantic decomposition theory] is not plausible. If
there are few convincing examples of denitions, there are literally no
convincing examples of denitions which take prima facie non-
sensory/motor terms into a sensory/motor vocabulary. There is, indeed,
no reason to believe even that the denitions we have examples of gen-
erally exhibit an epistemically interesting direction of analysis. On the
contrary, the right hand (deniens) side of standard examples of de-
nitions does not tend, by and large, to be more sensory/motor than the
left hand (deniendum) side. So, for example, the conventional wisdom
has it that grandmother means mother of father or mother; ketch
means two-masted sailing vessel such that the aft (or mizzen) mast is
stepped forward of the helm; bachelor means unmarried man; kill
means cause to die, etc. Its hard to discern a tendency toward the
sensory/motor in these examples, and these examples are quite typical.
Surely, no one could seriously argue that words like mother, father,
mast, vessel, cause, die, married and man, all of which occur in the
deniens, are somehow closer to transducer language, than, say, grand-
mother, ketch, kill or bachelor are. (Fodor et al. 1980: 268)
Fodor et al. are right, of course, when they argue that, for the overwhelming
majority of lexical concepts, verbal denitions are no nearer to sensorimotor
experience than the terms dened. However, there is no reason to think that
the semantic representations of lexical concepts are verbal denitions. Let us
take one of their own examples: the meaning of ketch. While it is true that
terms such as mast or vessel are no more perceptual than the term dened, it
does not follow that our mental representation of the concept ketch does not
include a perceptual component. In fact, I would argue that a visual image is
a central part of the meaning of the term. One acquires the concept that goes
with the phonological label ketch by seeing exemplars and extracting their
shared properties. A verbal denition is helpful to the extent that it draws
attention to the relevant characteristics, but it is neither necessary (since it is
possible to acquire the concept without ever hearing the denition) nor su-
cient (a person who had been presented with a denition but never seen the
real thing would have an underspecied representation of the concept).
One might argue that knowing the meaning of ketch involves much more
than knowing what it looks like: equally relevant are the knowledge that it is
a type of ship (and hence that it is used for travelling over water), an under-
standing of how it moves, and so on. While these observations are undeniably
true, they would only constitute a valid objection if this kind of knowledge
could not be represented using symbols derived from perceptual experience.
DABROWSKA text Make up 26/3/04 4:34 pm Page 109
Words 109
event that it describes and compares it with his own perception of the event; if
the two structures map onto each other, the sentence can be described as true.
While performing the comparison, the conceptualiser develops perceptual
symbols for the simulation of the event, the process of comparing the simula-
tion to the perceived scene, and the outcome of the comparison (a successful
mapping). If the process is repeated on a number of occasions, a simulator will
develop for the concept : that is to say, the conceptualiser will have
learned to simulate the experience of successfully mapping a simulation onto
a perceived situation (Barsalou 1999).
This proposal, though admittedly sketchy and speculative, shows that it is
possible to represent abstract concepts using experientially grounded symbols,
at least in principle, and hence the pessimism expressed by Fodor and others
is unfounded. A dierent, but not necessarily incompatible approach to
abstract concepts is to be found in the work of many cognitive linguists, who
have argued that these concepts are understood at least partly in metaphori-
cal terms, that it to say, by importing structure from more basic domains. For
example, it has been suggested that is understood partially through the
metaphor : hence, when people
get angry, we say that they get hot under the collar, that their blood boils, that
they get all steamed up, and eventually, they may explode (Kvecses 1986;
Lako 1987; Lako and Johnson 1980; Matsuki 1995). Other linguists have
proposed that mental processes such as understanding and reasoning are
understood in terms of more accessible concepts such as and
(see Jkel 1995; Lako and Johnson 1980).
It seems, then, that the idea that meanings are built up from perceptual
primitives does oer a feasible solution to the problem of how we come to have
concepts. Clearly, many problems remain to be solved and many important
details still need to be elaborated. However, the considerable progress that has
already been made suggests that this is a very promising avenue of research. It
should also be emphasised that, other things being equal, there are several
reasons for preferring theories which use perceptual primitives over those
which appeal to amodal systems. First, using perceptual primitives solves the
symbol-grounding problem. Theories which postulate amodal symbols must
provide an account of how these are mapped onto perceptual states and enti-
ties in the world.11 This problem never arises in a system which uses percep-
tual symbols, since these are built bottom-up from perceptual states. Secondly,
using perceptual primitives dispenses with the need to postulate biologically
implausible innate structures with no obvious evolutionary antecedents.
Perceptual systems have been around for a long time and the neurological
structures that they use provide an obvious source of building blocks for
higher-order concepts. Last but not least, perceptual primitives, while con-
straining learning in highly specic ways, oer enough exibility to support
the wide variety of lexical concepts that we encounter in the worlds languages.
All this said, the proof of the pudding is in the eating, and a tting way to
DABROWSKA text Make up 26/3/04 4:34 pm Page 111
Words 111
conclude this chapter is to ask whether a system equipped only with percep-
tual structures and a general learning mechanism could acquire the locative
terms of the various languages that have been discussed earlier. The ultimate
test of the proposal would be to design a computer model which had the basic
perceptual structures built in and see how well it would perform on spatial cat-
egorisation tasks. As it happens, such a model has been built, thus oering us
an opportunity to evaluate the proposal.
the landmark is vertical and sk animal back when it is horizontal, regard-
less of whether there is contact or not. German requires its speakers to pay
attention to both features: auf is used when the trajector is in contact with a
horizontal surface, an when it is in contact with a vertical surface, and ber
when there is no contact. Finally, Japanese has only one term, ue ni, which is
insensitive to either distinction. The network, like a human child, learned
whichever system it was exposed to. The built-in perceptual structures oered
enough constraints to make learning possible and enough exibility to allow
learning systems based on dierent principles.
It should be emphasised that Regiers network is not a realistic model of
human spatial cognition. The structures built into the model are only loosely
motivated by known neural structures, the relationship being one of analogy
rather than exact correspondence. The network did not have to solve all the
complexities discussed earlier in this chapter: its input was highly idealised
(the objects in the movies were simple, two-dimensional gures labelled tra-
jector and landmark) and it did not have to deal with the additional compli-
cations posed by polysemy or virtual boundaries. Furthermore, the network
lacked the prelinguistic knowledge that we know children bring to the acqui-
sition task, and it had no social experience (it did not interact with other
agents). However, the fact that it was able to extract the relevant features of
spatial congurations and succeeded in generalising them to novel instances
shows that it is possible to build semantic representations from perceptual
primitives, and thus oers indirect support for the conjecture that this is what
children do as well.
5. Conclusion
A common argument for the innateness of syntax is that children could not
learn certain aspects of the grammar of their language through any known
learning process, and therefore at least some syntactic knowledge must be
innate. This is an example of what Dawkins (1986) calls an argument from
incredulity: I dont see how children could possibly learn syntax, and therefore
they cannot learn it it must be innate.
Explaining language acquisition by appealing to innate knowledge does
not answer the question of how we came to have language; it just shifts it from
an ontogenetic to a phylogenetic level. Saying that the species acquired lan-
guage through some unknown evolutionary process is not an advance on
saying that children acquire language through some unknown learning
process. The unlearnability argument was generally accepted partly because
the generative tradition in psycholinguistics has tended to overestimate chil-
drens grammatical accomplishments (that is to say, impute to them knowl-
edge that they do not have: cf. Dabrowska 1997) and underestimate their
lexical accomplishments (what is actually involved in learning a word) thus
DABROWSKA text Make up 26/3/04 4:34 pm Page 113
Words 113
Notes
1. For example, Miller and Gildea (1987) found that children participating
in vocabulary-building programmes at school produced sentences such as
I was meticulous about falling o the cli and Our family erodes a lot. They
were puzzled by these until they consulted the dictionary that the children
were using and learned that meticulous means very careful and erode, to
eat out.
2. Presumably, one can use both forms when describing a currently invisible
dog: in this case, the speaker would assume the vantage point of a virtual
observer. Virtual observers also gure in the semantics of English expres-
sions. Imagine you are describing a family photograph depicting Grandpa
Watkin standing beside Granny Watkin. You can say either Grandpa
Watkin is to the left of Granny Watkin or Grandpa is on Grannys right,
depending on whether you are assuming the point of view of the person
looking at the photograph or of the people in it.
3. This is not an isolated example of such projection: other hilly topo-
graphic adverbs also have extended senses relating to body parts (Casad,
personal communication).
4. The following discussion is an adaptation and extension of a similar dis-
cussion in Bowerman (1993, 1996). I would like to acknowledge my debt
to her.
DABROWSKA text Make up 26/3/04 4:34 pm Page 114
Words 115
As we saw in the last chapter, learning the basic lexical units of a language is
a daunting task. But to know a language, one must also be able to produce and
understand larger expressions: morphologically complex words, phrases and
sentences. Such complex structures can be assembled by combining smaller
units according to the rules of grammar.
Intuitively, grammatical rules and words seem to be very dierent kinds of
things. Words are formmeaning pairings, learned by associating stretches of
sound with meanings. Rules, on the other hand, are procedures for manipu-
lating words or rather, classes of words. A simple, and rather hackneyed,
example is the rule for forming the past tense of regular English verbs, which
adds the -ed sux to the verb stem. Linguists usually represent this procedure
with a formula such as this:
Simple though it is, the past-tense rule is abstract, in that it can apply to any
(regular) verb; consequently, unless its application is pre-empted by the exis-
tence of a stored irregular form, speakers can use it to produce the past tense
of any verb in their vocabulary, and of any new verb that they might encoun-
ter.
It is rules which allow us to go beyond the expressions we have actually
heard and produce novel forms. Thus, to become competent speakers of their
language, children must acquire some mental analogue of rules. And in fact,
there is plenty of evidence that they do. Sticking to the English past-tense rule,
a classic study conducted by Jean Berko in 1958, and numerous replications
by other researchers since then, have demonstrated that children are able to
inect novel verbs in experimental conditions: for example, when presented
with a nonsense verb like zib they are able to form the past tense form zibbed.
Outside the laboratory, nearly all children occasionally produce regularisation
errors such as breaked and catched, sometimes after a period of correct pro-
DABROWSKA text Make up 26/3/04 4:34 pm Page 117
duction. Until the mid-1980s, nearly everyone agreed that these well-known
facts provided direct evidence that children learn symbolic rules of the kind
postulated by linguists.
evidence that they are equipped with a powerful learning mechanism dedi-
cated specically to acquiring language. Rumelhart and McClellands network
used a relatively simple, general-purpose learning mechanism to extract pat-
terns from the input, and produced the same kind of behaviour simply by
forming low-level associations between representations, without any explicit
rules. This raised the possibility attractive for some, unpalatable for others
that the rule-like behaviour seen in humans might also be a product of such a
general-purpose, low-tech mechanism.
As critics were quick to point out, there were many problems with this early
model (Lachter and Bever 1988; Pinker 1991; Pinker and Prince 1988). The
network could learn rules that are not attested in any human language (e.g.
form the past tense by mirror-reversing the order of segments), but it could
not learn some fairly common patterns (such as reduplication of the stem). It
sometimes produced blends which sound extremely strange to human ears
(e.g. membled for mailed, conafted for conicted). Because it only represented
phonological properties, it could not handle homophones with dierent past
tense forms such as lie to recline (past tense lay) and lie tell lies (past tense
lied). It was strongly dependent on phonological similarity, and hence unable
to inect novel regulars which did not resemble regulars in the training set.
Most damning of all, critics argued, Rumelhart and McClelland got their
model to generalise the regular pattern by manipulating the input. The
network was initially trained on just ten verbs, eight of which were irregular.
When these had been learned, Rumelhart and McClelland increased the train-
ing set to 420 verbs, of which the overwhelming majority were regular. It was
this vocabulary discontinuity or, to be more precise, the fact that it was sud-
denly swamped by regular mappings that caused the network to generalise.
Some of these problems were addressed in later models (e.g. Daugherty and
Seidenberg 1992; Hare and Elman 1995; Hare et al. 1995; MacWhinney and
Leinbach 1991; Marchman 1993; Plunkett and Marchman 1991, 1993). The
later models incorporated one or more layers of hidden units between the
input and the output units, allowing the network to form internal representa-
tions, and hence making it less dependent on surface similarity. They also
adopted more realistic training regimes, and some represented aspects of
semantic structure, which enabled the network to distinguish homophones.
However, the critics also struck at an idea that lies at the very heart of the
connectionist approach to modelling morphological rule learning namely,
the claim that regular and irregular inections are handled by a single mech-
anism by drawing attention to signicant psychological and neuropsycho-
logical dierences between regular and irregular forms (see Pinker 1998, 1999;
Pinker and Prince 1992; Prasada and Pinker 1993). As pointed out earlier, chil-
DABROWSKA text Make up 26/3/04 4:34 pm Page 119
We have seen that there is good evidence that regular inections are dierent
from irregulars in several important ways although it is worth noting at this
juncture that most of the evidence relates to just one morphological sub-
system, namely, the English past tense. On the other hand, even die-hard
critics of connectionism like Pinker and Marcus acknowledge that articial
neural networks can model some aspects of human linguistic knowledge and
are able to generalise previously learned patterns to novel input. These critics
DABROWSKA text Make up 26/3/04 4:34 pm Page 120
Table 8.1 Circumstances in which memory is not accessed and the default
inflection is applied (adapted from Marcus et al. 1995)
Circumstance Examples
of lie to tell untruths is lied, not lay, in spite of the fact that its homophone
lie to recline is irregular. Hence, Yesterday I lay on the sofa and lied to my boss.
The regular inection is also used with various non-canonical roots, includ-
ing names, truncations, acronyms, and phonological sequences which repre-
sent noises or which stand for the word itself rather than its referent that is
to say, when the word is mentioned rather than used. (The sentence I checked
the article for sexist writing and found three mans in the rst paragraph means
that the speaker found three occurrences of the word man, not three adult
male human beings pressed between pages 1 and 2: the form man refers to the
word, not its referent.)
Finally, the regular inection is used with several categories of derived
words. Typically, derived words inherit the grammatical properties of their
head, so if the latter is irregular, the derived word, too, is irregular. Thus, the
past tense of overdo is overdid because the past tense of its parent, the verb
do, is did; and the plural of stepchild is stepchildren because the plural form of
child is children. Some derived words, however, cannot inherit their grammat-
ical properties from their parent because they have the wrong kind of parent.
For example, a verb derived from a noun or an adjective cannot inherit its
past-tense form from the original word because nouns and adjectives do not
have past-tense forms. When this happens, the derived word is inected regu-
larly: the past-tense form of the verb to spit, meaning to put on a spit, is
spitted (not spat); and the past tense of to right return to an upright position
is righted (not rote or wrote). This principle also applies to nouns derived from
other categories: we speak of ins and outs and ifs and buts.
Some words have more convoluted and more interesting genealogies.
Sometimes a verb is derived from a noun which itself was derived from a verb
(in other words, the verb has a noun as parent and a verb as grandparent), or
a noun might be derived from a verb which was derived from a noun. What is
interesting about such cases is that the intervening category (the middle gen-
eration) makes it impossible for the new word to inherit the inectional prop-
erties of its grandparent. For example, the verb cost calculate the costs, as in
She costed the proposal, is derived from the noun cost the amount of money
needed for a particular purpose, which in turn comes from the irregular verb
cost meaning be worth a particular amount of money. However, cost calcu-
late the costs cannot inherit the irregular past-tense form from cost be worth
a particular amount because it was derived via the noun, and nouns are not
marked for tense; consequently, the information about the irregularity of the
original verb is not accessible to the processing mechanism, and the default
inection must apply. The same phenomenon can be observed in words
derived via a name. The common noun Mickey Mouse (meaning simpleton)
is derived from the name of the Disney character, and since names dont have
plural forms, it cannot inherit the irregular plural of its grandparent, the
common noun mouse. Hence, the plural of Mickey Mouse is Mickey Mouses,
not Mickey Mice.
DABROWSKA text Make up 26/3/04 4:34 pm Page 123
frequency and applicability. In both the plural and the past tense in English,
the regular and the irregular inections rely on dierent morphological pro-
cesses. Regular past-tense and plural forms are formed by suxation, which
results in inected forms which are easily segmentable into a morpheme
which signals a type of activity and a morpheme that signals tense (e.g. play
+ ed) or a morpheme that designates a type of thing and a morpheme that
indicates plurality (cat + s). Most irregular inections, in contrast, involve
stem changes: ringrang, catchcaught, manmen, mousemice. The stem-
change mechanism is less transparent, and hence probably more dicult to
generalise. Furthermore, because there are several dierent patterns of stem
changes competing with each other, learners are forced to attend to the indi-
vidual properties of each word, rather than to similarities between words, and
this is clearly not conducive to unconstrained generalisation.
To summarise: it is undeniable that regular inections in English are very
dierent from the irregulars. Unlike the irregular patterns, they are not
restricted to a narrowly dened group of stems, but apply elsewhere, when
no other pattern can be used; and there are sharp contrasts in the way that
regular and irregular inections are acquired, processed and aected in lan-
guage impairment. What is not clear is whether these dierences are a result
of the fact that regular and irregular inections rely on dierent mental mech-
anisms, or whether they are attributable to the idiosyncratic properties of the
English system of marking past tense and plurality specically, to the eect
of dierences in frequency, applicability, morphological process, or some
combination of these factors. Because of these confounds, the English past-
tense and plural inections are simply not very good testing grounds for the-
ories dealing with the mental status of rules.
But the dual-mechanism theory, of course, is not just about English inec-
tions: it is about language in general. The circumstances in which the default
inection applies are supposed to fall out automatically from the organisation
of the human capacity for language; hence, the same set of circumstances
should be associated with the regular inection in all languages. What we must
do, then, is to test this prediction against data from other languages. Ideally,
we would want to nd an inectional system in which there are at least two
truly productive endings and in which the regular inection (i.e. the inection
that speakers use in the circumstances enumerated in Table 8.1) applies to a
minority of types. The existence of such minority default systems would
show conclusively that regularity does not depend on high type frequency, and
hence provide strong evidence in favour of the dual-mechanism theory.
Do we know of any such systems? Marcus et al. (1995) argue at length that
in the German plural and past-participle inections the regular patterns
applies to a minority of forms, and oer the Arabic plural as a third example.
Unfortunately, the last two of these putative cases are problematic for two
reasons. First, it is not clear that the regular patterns in these subsystems do
in fact apply to a minority of types (see Boudelaa and Gaskell 2002; Bybee
DABROWSKA text Make up 26/3/04 4:34 pm Page 125
1995), and in any case, they certainly apply to more types than any one irreg-
ular pattern, of which there are a number in both languages. Secondly, like the
English past-tense and plural inections, these systems use dierent morpho-
logical mechanisms for regulars and irregulars. The past participle of regular
(or weak) verbs in German is formed by adding a sux (-t) and in most cases
also a prex (ge-), while most irregular (or strong) past participles, like irreg-
ular past-tense and past-participle forms in English, require stem changes
(and a dierent sux, -en, as well). In the case of the Arabic plural, the regular
mechanism (the so-called sound plural) simply adds an ax to a stem, while
the irregular mechanism (the broken plural) involves modifying the stem in
various ways, sometimes quite dramatically. The German plural, on the other
hand, appears to be a perfect test case for the dual-mechanism theory, and it
is to this inection that we now turn.
German has ve plural suxes (-e, -en, -er, -s and zero), three of which (-e, -er
and zero) are sometimes accompanied by umlauting of the stem. Clahsen
(1999), Clahsen et al. (1992), Marcus et al. (1995), Pinker (1999) and others
have pointed out that one of the suxes, -s, has a special status. Like the
regular plural ax in English, it is used in the special circumstances enumer-
ated in Table 8.1: with novel words, names, borrowings, onomatopoeic words,
acronyms, etc. This, they argue, shows that -s is the regular or default ending
for the German plural.
What makes the -s ending particularly interesting is the fact that it applies
to a small number of nouns only about 4 per cent (Marcus et al. 1995: 227)
which makes it a minority default par excellence. Furthermore, most other
plurals are also formed by suxation, thus eliminating the morphological
process confound; and at least two other endings are productive that is to
say, they apply to fairly large and open-ended classes of nouns.
Proponents of the dual-mechanism theory have proposed that German-
speaking adults and children learning German know that -s is the default
ending and therefore apply it whenever a plural form cannot be retrieved from
memory. They support this claim with three types of evidence. First, children
learning German often overgeneralise -s; and although they also overgeneral-
ise the other endings, regularisation errors (i.e. overgeneralisations of the
regular ending) are an order of magnitude more frequent than irregularisa-
tions (Clahsen et al. 1992: 247). Secondly, adult speakers tend to use -s with
novel words, particularly unusual-sounding novel words (Marcus et al. 1995).
Finally, some researchers have argued that -s plurals are processed in a qual-
itatively dierent way (see Clahsen 1999 and the work cited there).
Hence, the German plural has come to be regarded as the denitive evidence
for the dual-mechanism theory, the exception that proves the rule and also
DABROWSKA text Make up 26/3/04 4:34 pm Page 126
-en (11.7 per cent). In other words, -s overgeneralisations are more frequent
than we would predict on the basis of the frequency of the ax alone, showing
that frequency is not the only factor determining overgeneralisation rates; they
are not, however, more frequent in any absolute sense.
In fact, overgeneralisation rates for -en, the most frequent ax, are very high
whichever method we use to calculate them: in Clahsens own corpus, the rela-
tive overgeneralisation rate for -en is 22.5 per cent almost the same as that for
-s. Clahsen et al. do not regard this as a problem; in fact, they manage to turn
it into an argument in favour of the dual-mechanism theory, arguing that chil-
dren overgeneralise -en because they think it is the default ending. As evidence
for this, apart from the high overgeneralisation rates, they point out that
regular plurals do not occur inside compounds,3 and the children in their
sample who had high overgeneralisation rates for -en tended to leave it out in
compounds, even when the adult form contained a plural marker. For example,
instead of saying Dose-n-ner can-s-opener, Bauer-n-hof farmer-s-yard,
Kche-n-fenster kitchen-s-window, their subjects said Dose-ner, Bauer-hof
and Kche-fenster. However, there are some serious problems with this argu-
ment. First, the children did not invariably leave -en out in compounds: in fact,
they supplied it most of the time. Secondly, the -en in compounds does not nec-
essarily have a plural meaning: note, for example, that Bauernhof is prototypi-
cally a yard belonging to a single farmer, not several farmers; likewise,
Kchenfenster is a window in a kitchen, not kitchens. This suggests that the -en
in compounds is not a plural marker, but either a historical residue or a linking
morph inserted for phonological reasons (cf. Wegener 1999). Last but not least,
the nding that children are particularly prone to overgeneralise the plural
marker which they leave out in compounds has not been conrmed by other
researchers (see Gawlitzek-Maiwald 1994).
The claim that German adults treat -s as the default is equally problematic.
It is true that speakers tend to use -s more often with nonce words than with
real words. However, they do not apply it indiscriminately to all, or even most,
novel forms. In Kpckes (1988) experiment, for example, -s plurals accounted
for only 17 per cent of the responses. Furthermore, speakers used dierent
axes with dierent nonce words, depending on the words gender and
phonological make-up: for example, speakers preferred -en with feminine
nouns ending in -ung and -schaft and masculine nouns ending in schwa, -e
with masculine nouns ending in -ling, -s with nouns ending in a full vowel, and
so on. By and large, the likelihood of a particular plural marker appearing
after a particular nonce word depended on how frequently that marker
occurred with real words similar to the nonce word (i.e. real words with the
same endings, the same number of syllables, and so on). Other researchers (e.g.
Gawlitzek-Maiwald 1994) have obtained similar results. Thus, nonce-word
experiments show that German speakers are aware that each of the ve plural
axes has a dierent domain of application, and that they apply the axes
accordingly.
DABROWSKA text Make up 26/3/04 4:34 pm Page 128
Another source of evidence for the special status of -s comes from two
psycholinguistic experiments tapping speakers performance online, i.e. while
they are actually processing plural forms. The rst of these, by Clahsen et al.
(1997), found that people processed high-frequency irregular nouns faster than
low-frequency irregulars while no such dierences were found with regular
nouns. The second, by Sonnenstuhl et al. (1999), used the cross-modal priming
technique to investigate the eect of singular versus plural primes on lexical
decision tasks involving singular nouns. In experiments using this technique,
people are asked to decide as quickly as possible whether a visually presented
target (in this case, a singular noun) is a word. It is well known that reaction
times are shorter when participants hear a related word (called a prime) imme-
diately before seeing the target, and the experimenter can vary the kind of
prime and measure the eect on reaction times to the target word. Sonnenstuhl
et al. used two kinds of prime: the noun stem itself and the plural form of the
noun. Their results showed that a plural form with -s was as good a prime for
the singular as the singular form itself (reaction times to a singular form like
Karton box were the same when it was preceded by Karton or by Kartons); but
for -er plurals, the singular was a better prime than the plural (participants
reacted faster when Kind child was preceded by Kind than when it was pre-
ceded by its plural Kinder). The authors interpretation of this result is that reg-
ulars, but not irregulars, are morphologically decomposed during processing:
in other words, when processing regular nouns, speakers obligatorily access the
stem form and hence react as if they had actually heard the stem.
Several researchers have pointed out various problems with these experi-
ments, some methodological and some pertaining to the interpretation of the
results (Drews 1999; Sereno et al. 1999). Perhaps the most serious problem is
that both studies compared -s with -er, the one plural ax which everyone
agrees is not productive. Of all the German plural axes, -er is the most
restricted in applicability (it is used only with strong masculine and neuter
nouns) and the least predictable in distribution, so each -er plural must be
stored in the lexicon. It is the only ax which was almost never used by sub-
jects in Kpckes (1988) nonce-word production task, and in an acceptability
judgement study conducted by Marcus et al. (1995), most nonce words with
-er were judged to be the least acceptable of all axed plurals. Finally, -er is
very rarely overgeneralised by children (Behrens submitted; Ewers 1999;
Gawlitzek-Maiwald 1994; Kpcke 1998). Thus, the dierences between -s and
-er could be due to the special characteristics of -er rather than the properties
of the default plural. To test the predictions of the dual-mechanism theory, it
would have been much more appropriate to compare -s with the more produc-
tive axes (-en and -e).
Yet another problem with treating the German -s as a default analogous to
the English -s plural is that, unlike its English counterpart, the German ending
does not invariably apply in all the circumstances in Table 8.1. One systematic
exception, acknowledged by Marcus et al. (1995), is bahuvrihi compounds,
DABROWSKA text Make up 26/3/04 4:34 pm Page 129
which take whichever plural ending is required by the rightmost element: for
example, the plural form of Gromaul big mouth, braggart is not *Gromauls
but Gromuler because the plural of Maul mouth is Muler. Another
obvious exception is rare words: since -s applies to only 4 per cent of German
nouns, the vast majority of all nouns, rare or frequent, take other plural axes.
In the remaining circumstances, -s is the usual ending, but it is not the only
option, a fact which raises serious doubts about the viability of the dual-mech-
anism explanation. Consider plural marking on borrowings, for example.
Marcus et al. (1995), citing Kpckes (1988) data on recent borrowings, point
out that about 50 per cent of them take the -s plural. They regard this as satis-
factory evidence that -s is used with borrowings. However, as a matter of arith-
metical necessity, if 50 per cent of borrowings take -s, the other 50 per cent do
not. With such a high proportion of exceptions, plural marking on borrow-
ings should be something for Marcus and his co-authors to worry about, not
cite as support for their theory. Note, too, that a large proportion of recent
borrowings in German are either English words or words that entered German
via English. Since many German speakers, particularly of the younger gener-
ation, have a good working knowledge of English, it is not unreasonable to
assume that at least some of the new words were borrowed with their English
plural. Languages do sometimes borrow foreign plurals along with the singu-
lar (cf. English criteria, alumni, seraphim), and this is more likely when the
sux also exists in the borrowing language.
Other exceptions include nominalised verbs, which inect like ordinary
nouns (Essen- meals, from essen to eat; Schau-en shows, from schauen to
look; Kuf-e purchases, from kaufen to buy), and some nominalised VPs
(e.g. Tunichtgut-e neer-do-wells, Habenichts-e have-nots). It is not true,
therefore, that nouns derived from other categories always take the regular
plural. Nor is -s invariably used with names and abbreviations. The latter can
take -en when the nal vowel is stressed: for example, the plural of LPG, from
Landwirtschaftliche Produktionsgenossenschaft farming cooperative, is either
LPGs or LPGen (Wegener 1994: 269). And although -s is strongly favoured with
names, the preference is much stronger with surnames than with place names
(which often take - rather than -s), product names (which can take -, and
sometimes also -en or -e: see Wegener 1999), or rst names (so we have Emil-e
and Beate-n alongside Emil-s and Beate-s). Finally, -en and -e plurals are
applied in default circumstances when the noun stem ends in -s (Indefrey 1999).
To conclude: it is undeniable that -s has a somewhat special status among the
plural axes in German. The acquisition and nonce-word data shows that it is
generalised more readily than one would expect, considering its low frequency,
and it is the usual plural ending in most of the circumstances listed in Table 8.1.
Both of these ndings call for an explanation.4 However, the fact that other
endings are also found in default contexts raises serious doubts about the expla-
nation put forward by proponents of the dual-mechanism theory; and there is
little evidence that German speakers, either child or adult, treat it as a default.
DABROWSKA text Make up 26/3/04 4:34 pm Page 130
The German plural, then, is certainly not the denitive proof of the dual-
mechanism theory that it is sometimes claimed to be. To evaluate the claims
made by the theory, we need to examine still more data. In the following sec-
tions, we will consider two further inectional subsystems the Polish geni-
tive and dative whose properties, as we shall see, make them particularly
interesting test cases.
(2) a. tenor tenor (voice), gen. tenoru tenor tenor (singer), gen. tenora
b. tl-ok crowd, gen. tl-oku tl-ok piston, gen. tl-oka
(3) a. motor motorcycle, gen. motoru traktor tractor, gen. traktora
b. smak taste, gen. smaku ssak mammal, gen. ssaka
DABROWSKA text Make up 26/3/04 4:34 pm Page 132
The problem is that the mere existence of such pairs does not help us to deter-
mine which ending is regular, since both homophony and rhyme are symmet-
rical relations: for every -a noun which has a homophone that takes -u, there
is an -u noun which has a homophone that takes -a; likewise, for every -a noun
that rhymes with an -u noun there is an -u noun that rhymes with an -a noun.
The remaining nine criteria, however, can be applied to the Polish genitive, and
it is to these that we now turn. (Novel words and overgeneralisations in child
language will be dealt with in the following subsections.)
Low-frequency words are equally likely to take either ending. Of the 100 least
frequent masculine nouns from a frequency list compiled by Kurcz et al.
(1990), 48 take -a and 46 take -u.
Borrowings usually take -a if animate and mostly -u if inanimate. Some,
however, are not inected at all, usually because they are in some way non-
canonical and hence cannot be assimilated to any of the established templates.
For example, guru and urdu do not sound like Polish nouns, which never end
in -u in the nominative. Kamikadze, attach, and dingo are phonologically
similar to Polish neuters, but they are assigned the masculine gender because
of their meaning (kamikadze and attach prototypically refer to males, and
dingo inherits its gender from its superordinate pies dog). Because of this, the
neuter pattern is not applicable and such nouns remain uninected. The word
boa is even more interesting. It inherits its gender from its superordinate waz
snake, which is masculine. Although words ending in -a in the nominative are
usually feminine, there is a sizeable group of masculine nouns with this ending
(e.g. poeta poet, artysta artist, logopeda speech therapist); these, in spite of
their gender, are inected like feminines. However, such nouns invariably refer
to human beings, and because of this semantic discrepancy, the pattern cannot
be applied to boa, an animal.
The fact that some words, particularly non-canonical-sounding words, are
left uninected is dicult to accommodate in the dual-mechanism theory,
according to which the default inection must apply if there are no similar
examples stored in memory which the speaker could analogise from. One
could argue, of course, that the default ending in this case is zero. The problem
with this is that such non-canonical words are found in all three genders, and
they are left uninected in all cases, not just in the genitive. Following the same
logic, we would have to conclude that zero is the default ending for all cases
and genders. This would amount to saying that in a richly inected language
like Polish, all the real work is done by associative memory, while the rule
system is on hand to add zero that is to say, do nothing when memory fails.
This is clearly not a very attractive proposal.
Unusual-sounding words: as explained earlier, words which cannot be
accommodated in any established template are simply left uninected.
Words which are mentioned rather than used can occur in the citation form
(4), in the form in which they appeared in the text quoted (5), or in the form
required by the syntactic context of the sentence in which they are mentioned
DABROWSKA text Make up 26/3/04 4:34 pm Page 133
(6). In the latter case, the usual ending is used: -a nouns such as autor author
take -a and -u nouns like rasizm racism take -u.
(4) Jak sie pisze autor?
How do you spell autor (author:NOM)?
(5) Skresl wszystko od autorowi do rasizm.
Cross out everything from autorowi (author:DAT) to
rasizm (racism:NOM).
(6) Nie moge znalezc tego drugiego autora/rasizmu.
I cant nd the other (occurrence of the word) autor
(author:GEN)/rasizm(racism:GEN).
Truncations usually take the same endings as the words from which they are
derived either -a or -u, depending on the base noun. Thus, dyr (from dyrek-
tor director) and merc (from mercedes) both take -a, like their full forms;
while hasz, like its base haszysz hashish, takes -u. When the truncated stem
is not a noun, the choice of ending is determined by semantic factors. For
example, sam supermarket is derived from (sklep) samoobsl-ugowy self-
service (shop). Samoobsl-ugowy is an adjective and hence takes the adjectival
ending -ego in the genitive, which sam cannot inherit because it is not an adjec-
tive and does not sound like one, so in this case, the usual semantic criteria
apply: sam, like most nouns designating large objects most naturally concep-
tualised as locations, takes -u.
Acronyms are often left uninected; but when they do take an ending, it is
nearly always -u. By this criterion, then, the default ending is -u. However, the
overwhelming majority of acronyms designate organisations and institutions,
i.e. abstract entities, so the choice of ending can be explained by appealing to
semantic criteria.
Words derived from a dierent category: when the noun is derived from a
dierent category (e.g. a verb or an adjective), the choice of ending depends
primarily on the morphological process used in the derivation. Most mascu-
line nouns derived from verbs designate either the agent of the action (e.g.
pisarz writer) or the instrument used to perform the action (zszywacz
stapler), while those derived from adjectives tend to designate individuals
possessing a particular characteristic (e.g. dziwak an eccentric from dziwny
strange, eccentric). These nouns invariably take -a. As mentioned earlier (see
note 7), in such cases semantic and morphological factors work together: der-
ivational suxes come to be associated with a particular inectional ending
because of the meaning of the words formed with the sux; and the same
inectional ending is used even when the derived word doesnt have the right
semantic properties. On the other hand, nouns derived using the three
u-friendly axes, -unek, -ot and -izm, take -u. This is also compatible with their
semantics, as nearly all nouns with these axes are abstract.
Nouns can also be derived from words belonging to other categories by
DABROWSKA text Make up 26/3/04 4:34 pm Page 134
removing the verbal or adjectival sux from the base. Most words derived in
this way take -u: bieg a run from biegac to run, wykl-ad lecture from
wykl-adac to lecture, braz the colour brown from brazowy brown (adj.) etc.
Again, this is largely attributable to semantics, since such words tend to desig-
nate abstract concepts. However, the use of -u with back-formations is not uni-
versal: -a is used when the resulting noun refers to an animate being (e.g. psuj
a clumsy person who is always breaking things from psuc to break, brutal a
brutal person from the adjective brutalny brutal) and very occasionally with
inanimates as well, especially when the derived noun refers to a concrete object
(gryz a bite from gryzc to bite, skret roll-up from skrecac to twist).
The last possibility for deriving new nouns from other categories involves
conversion. This option is available only with adjectives such as uczony
learned or sl-uzacy serving which can also be used as nouns meaning
scholar and servant. Such nouns invariably keep their adjectival inection.
Words derived via a dierent category can take both endings, and the choice
depends on the morphological processes used in the derivation. If the derived
noun is formed with an a-friendly ax like -acz, -nik, or -arz, it takes the ending
preferred by the ax. For example, the noun utleniacz oxidant is derived from
the verb utleniac to oxidise, which in turn was derived from the noun tlen
oxygen, and it takes -a in spite of the fact that tlen takes -u. On the other hand,
nouns derived from denominal verbs by back-formation require -u. Thus,
dodruku is the genitive form of the noun dodruk an additional print run,
derived from the verb dodrukowac to print additional copies, which in turn
comes from drukowac to print, which is derived from the noun druk print.
Bahuvrihi compounds referring to inanimate objects typically take -u,
although there are many exceptions; animates, on the other hand, always take
-a. Here are some examples:
(7) -u inanimates:
piorunochron lightning rod (lit. lightning-protect), gen.
piorunochronu rwnolegl-obok parallelogram (lit. parallel-side),
gen. rwnolegl-oboku
(8) -a inanimates:
katomierz protractor (lit. angle-measure), gen. katomierza
trjkat triangle (lit. three-angle), gen. trjkata
(9) -a animates:
garkotl-uk a clumsy cook (lit. pot-break), gen. garkotl-uka
stawong arthropod (lit. joint-leg), gen. stawonoga
Particularly relevant for our purposes are compounds which have a masculine
noun as the rightmost element but require a dierent ending from the noun.
As the following examples illustrate, neither ending has a special status: a com-
pound containing an -a noun may require -u and vice versa.
DABROWSKA text Make up 26/3/04 4:34 pm Page 135
(12) Ale malutki gon! Nigdy nie widzial-em takiego malutkiego ______.
What a tiny gon! I have never seen such a tiny ______.
The remaining words were introduced as place names, which show no prefer-
ence for either ending, thus providing a base-line condition.
Of the 24 masculine items, four ended in consonants or consonant clusters
which, in the real lexicon, are strongly associated with -a, and four in conso-
nants or consonant clusters strongly associated with -u. The remaining words
had a variety of osets typical for masculine nouns.
DABROWSKA text Make up 26/3/04 4:34 pm Page 137
to use -a when the noun was presented as an object and -u when it was pre-
sented as a substance. This pattern was found in 57 respondents; the remain-
ing three gave equal numbers of -a responses in both conditions. Nouns
presented as place names took both endings with equal likelihood, and thus
pattern between the two extremes.
It is clear, then, that regularity in the technical sense of the dual-mechanism
theory is not a good predictor of productivity. Polish speakers freely general-
ise the irregular masculine inections to novel nouns but do not consistently
DABROWSKA text Make up 26/3/04 4:34 pm Page 138
supply the neuter ending, in spite of the fact that it is regular. Furthermore,
speakers are sensitive to both the semantic and the phonological factors aect-
ing the distribution of masculine endings, conrming that the latter are irreg-
ular in the technical sense of the dual-mechanism theory.
100
90
80
70
% correct
60
50
40 Masculine
30 Femine
20 Neuter
10
0
2;6 3;6 4;6
Age
form (i.e. the nominative) in contexts requiring the genitive, but such errors
gradually disappeared. As can be seen from Figure 8.2, overgeneralisation
errors were also relatively infrequent, so from about age 2, the childrens use
of the genitive is overwhelmingly correct.
Crucially, the pattern of development of the regular and irregular inec-
tions is very similar. Since the masculine and feminine inections apply to
classes of about the same size (44 per cent and 41 per cent respectively of the
childrens noun vocabulary), the simplest method of measuring growth is to
compare the number of noun types used with a particular inection. Figure
8.3 shows the cumulative number of masculine and feminine nouns used with
a genitive ending during the period from emergence (rst use) until acquisi-
tion (the point at which the child reliably supplies the correct ending) in the
four children. It is clear from the gure that both inections develop at the
same rate and show similar patterns of growth: progress tends to be slower at
the beginning, then speeds up, but neither inection shows a sudden take-o.
Marking rates on masculine and feminine nouns also show similar patterns of
growth (see Dabrowska 2004: study 2).
We now turn to the childrens overgeneralisation errors. According to the
dual-mechanism theory, the vast majority of all errors should be regularisa-
tions: that is to say, children should overgeneralise the feminine -i/-y and pos-
sibly -a (which is the regular ending for neuter nouns as well as an irregular
one for masculines). As shown in Table 8.5, -a was indeed overgeneralised
more often than -u, but this seems to be better explained as a frequency eect:
-a was also 34 times more frequent in the childrens correct productions, and
the relative overgeneralisation rate, which, as explained earlier, factors out the
DABROWSKA text Make up
26/3/04
4:35 pm
Page 140
Figure 8.2 Correct uses and overgeneralizations (OG) of genitive endings in the four children.
Note: The ages on the y-axis represent the endpoint of the relevant period. There is no data for Kasia between the ages of 3;1 and 3;11, so
the horizontal distances on the y-axis are not proportionate to the time between the data points.
DABROWSKA text Make up
26/3/04
4:35 pm
Page 141
Figure 8.3 Cumulative number of masculine and feminine noun types used in the genitive case
DABROWSKA text Make up 26/3/04 4:35 pm Page 142
Basia 1,387 1,118 1,422 2 1 0 0.4 0.1 0.0 0.5 0.8 0.0
Inka 1,050 1,424 1,187 9 2 3 0.6 0.1 0.2 0.9 0.5 0.4
Jas 1,137 1,429 1,264 39 10 1 2.3 0.4 0.1 3.3 2.3 0.1
Kasia 1,465 1,119 1,638 17 15 2 2.2 1.3 0.3 3.5 11.2 0.3
Whole corpus 3,039 1,090 3,511 67 28 6 1.4 0.4 0.1 2.2 2.5 0.2
Notes:
a
AOG rate (absolute overgeneralisation rate) = overgeneralisation errors as a
percentage of the number of noun tokens requiring other axes
b
ROG rate (relative overgeneralisation rate) = overgeneralisation errors as a
percentage of the total number of tokens with a given ax
eects of frequency, is very similar for both endings. Crucially, the one ending
which is uncontroversially regular, namely the feminine -i/-y, was overgener-
alised least often.
The children rarely used any endings in gender-inappropriate contexts: that
is to say, they tended to use masculine endings with masculine nouns (though
not necessarily the right masculine ending), the feminine ending with feminine
nouns, and the neuter ending with neuter nouns. As noted earlier, the gender
of most Polish nouns can be reliably predicted from the phonological form of
the nominative, so the infrequent occurrence of gender errors suggests that
children are sensitive to such cues from a very early age. The picture that
emerges, then, is that where there are consistent patterns, children learn them
with ease, and overgeneralisation levels are very low; but where the distribu-
tion of endings is less predictable, error rates are higher. It should be stressed,
however, that these are relative dierences: the overall overgeneralisation rates
are very low, ranging from 0.5 per cent in Basia to 3.3 per cent in Kasia.
Moreover, they remain consistently low throughout the entire period from
emergence until age 5;0 (cf. Figure 8.2).
Such low error rates in a system as complex and irregular as the Polish gen-
itive suggest that childrens performance relies to a considerable extent on the
use of stored forms. A closer analysis of the childrens early usage conrms
this suspicion. As pointed out earlier, there was a period of about ve months
after emergence during which children did not always supply the genitive when
required: some nouns which should carry the genitive inection were left in the
citation form (the nominative). What is interesting is that during this period
of inconsistent usage, genitive and nominative forms were used with dier-
ent populations of lexemes that is to say, at any one point in time, each child
DABROWSKA text Make up 26/3/04 4:35 pm Page 143
consistently provided the correct genitive form of some nouns and left others
in the citation form. Of course, over time the number of noun types which
were consistently inected in genitive contexts increased, while the number of
noun types which were left uninected got smaller and smaller, so if we con-
sider the entire period, we will nd that some nouns were used in both forms
in genitive contexts. However, in such cases, all the uses of the citation form
generally preceded all the genitive uses. We can see this from Table 8.6, which
summarises case marking on all the nouns in each childs transcripts which
occurred at least twice in genitive contexts (and hence could have been used in
dierent forms) during the period of inconsistent use. In Basias transcripts,
for example, there are 22 noun lexemes which occurred in genitive contexts
during this period at least twice. For eight of the 22 lexemes, there are no
attested uses of the nominative form in genitive contexts: that is to say, as far
Basia 8 7 5 2 22
Inka 32 2 2 1 37
Jas 19 2 4 1 26
Kasia 25 0 8 2 35
as we know, these nouns were used correctly from the very beginning. Another
seven lexemes were never inected during this period: all the attested forms are
nominative. Finally, seven lexemes occurred in both forms. For ve of the
seven, all the recorded nominative forms precede all the recorded instances of
the genitive. For example, there are 21 recorded occurrences of the lexeme tata
dad in contexts which require the genitive. Fifteen of these are in the nomi-
native case, and they all occurred before the genitive form appeared. The
remaining two lexemes were used inconsistently: that is to say, a nominative
form occurred at least once in genitive contexts after Basia had produced a
genitive form of the same noun. There were only six examples of such devel-
opmental U-turns in the entire corpus, and four of these were one-o slips (i.e.
the child used the nominative after a genitive only once). Thus, during the
period when the genitive was not consistently supplied, the childrens perfor-
mance on specic nouns was remarkably consistent, supporting the conjecture
that they were acquiring the genitive inection on an item-by-item basis.
On the other hand, there is also evidence that the children acquired some
rule or schema which allowed them to produce inected forms which they had
not heard before. We have already seen that, from about age 2, they marked
the genitive virtually 100 per cent of the time, which suggests that they were
able to add the inection to all the nouns in their vocabulary. Furthermore,
although overgeneralisation errors were rare, their very existence provides
DABROWSKA text Make up 26/3/04 4:35 pm Page 144
The research on the Polish genitive inection reported in the preceding section
brought one unexpected nding: that Polish speakers, adults as well as chil-
dren, are not fully productive with the neuter ending. This is puzzling, since
the genitive neuter inection is one of the simplest and most regular parts of
the case-marking system. Why then do speakers nd this relatively straightfor-
ward inection so troublesome? The answer, it seems, is to be found in the
peculiar structure of the neuter class.
The majority of neuter nouns are morphologically complex, and most
contain one of a few highly productive derivational morphemes. Over a half
of all neuter nouns are action nominalisations formed by adding the sux -nie
to a verb stem; and another 30 per cent are formed with one of three other
highly productive suxes, -cie, -stwo and -ko. As a result, most neuter nouns
are clustered in several small but very crowded regions of phonological space
dened by these axes, and there are relatively few nouns outside these clus-
ters. The masculine and feminine classes, of course, also contain large
numbers of morphologically complex nouns; however, none of the clusters is
so large, relatively speaking; and there are many nouns distributed fairly
evenly throughout the space between the clusters.
If, as suggested above, learners initially extract fairly specic generalisations
over clusters of nouns rather than fully abstract rules, then it is possible that
in the case of neuter nouns, they might never progress beyond these local gen-
eralisations, perhaps because there are not enough nouns outside the clusters
to force them to generalise further. This would result in a situation where
mature speakers would lack a general rule which applies to all neuter nouns,
but would know how to produce the dative forms of words ending in -nie,
-stwo, -ko etc. (or, if they pay attention to meaning as well as form, action
nominalisations formed with the -nie sux, abstract nouns formed with -stwo,
-ko diminutives, and so on). On the other hand, speakers would extract more
general rules in addition to such local regularities for masculine and feminine
nouns, which are more evenly distributed in phonological space.
DABROWSKA text Make up 26/3/04 4:35 pm Page 145
The results for the simplex words were similar to those in the genitive study:
participants performed close to ceiling on masculine and feminine nouns, but
gave the expected response to neuter nonce nouns less than 50 per cent of the
time (see Table 8.7). Most errors on simplex neuter nouns involved use of the
masculine or feminine endings (which were overgeneralised with approxi-
mately the same frequency); but in just under 10 per cent of the trials, partic-
ipants failed to inect the nonce word at all, and simply used the nominative
form in a grammatical context requiring the dative. This distribution of errors
was due to all participants responding inconsistently, rather than some partic-
ipants consistently using the same incorrect response with all or most of the
neuter nouns: in other words, the same respondents used the correct neuter
ending with some nouns, the masculine and feminine endings with others, and
left some nouns uninected.
The gures in Table 8.7 also show that participants supplied the expected
ending with morphologically complex neuter nonce words considerably more
often than with simplex neuters; in fact, performance on neuters with the
nominalising sux -nie was even better than on masculines and feminines.
This indicates that speakers do indeed rely on low-level patterns. On the other
hand, participants were able to supply the expected ending with morphologi-
cally simple neuter nonce words in almost half of the trials, which suggests
that they also have a more general rule. This higher-level generalisation,
however, is less well entrenched than the more specic patterns, and too weak
to be reliably used in production.
It is worth noting that performance on -nie nominalisations was much
better than on diminutives. This is almost certainly a type frequency eect: as
explained earlier, there are about ve times as many -nie nominalisations as
there are diminutives with -ko in the real lexicon. Thus, type frequency does
matter, although it does not seem to matter as much as phonological coher-
ence: the class of neuter nouns is necessarily larger than the class of neuter
nouns ending in -nie, yet speakers apparently have a well-entrenched rule for
the latter but not the former. Notice, too, that although the masculine ending
applies to more than twice as many nouns as the feminine -e, performance on
DABROWSKA text Make up 26/3/04 4:35 pm Page 147
masculines was actually slightly worse than on feminines. This suggests that
the eects of type frequency are non-linear: robust generalisations require a
relatively large number of exemplars; but beyond a certain threshold, further
exemplars do not aect productivity.
8. Interim conclusions
Earlier in this chapter, we saw that research on the English past-tense and
plural marking systems appears to support the dual-mechanism theory. In
both of these systems, a single, phonologically unrestricted ax applies in a
variety of circumstances in which memory cannot be accessed, and there is
convincing psycholinguistic and neurolinguistic evidence that regular and
irregular inections are processed dierently by the healthy brain and dier-
ently aected in neurological disorders. However, it is not clear whether these
dierences reect the properties of two dierent mental mechanisms, or
whether they are simply an artefact of the idiosyncratic properties of the
English system of marking tense and plurality. We also saw that in the German
plural, the distinction between regular and irregular endings is much less clear-
cut. One of the endings, -s, does have a special status, but it is not always used
in default circumstances, and, more importantly, speakers do not treat it as a
default: German-speaking children overgeneralise -en or -e more frequently
than -s, and both children and adults prefer these two irregular endings with
nonce words.
The Polish results cast further doubt on the idea that there is a clear-cut dis-
tinction between regular and irregular inections. We have seen that irregular
inections can appear in default circumstances, and that at least some regular
inections (e.g. the dative feminine endings) are sensitive to phonological
properties. Furthermore, regularity is a poor predictor of productivity: some
irregular inections (e.g. the two genitive masculine endings) are freely gener-
alised to new words, while some regular ones (e.g. the genitive neuter and
dative neuter endings) are not.
The dative experiment also revealed very strong similarity eects on neuter
nouns, in spite of the fact that the dative neuter inection is regular. This is
important in the context of the debate between the connectionists and propo-
nents of the dual-mechanism theory, since it undermines one of the strongest
arguments against connectionist approaches to language. As noted by Pinker
(1999) and Marcus (1998), connectionist models are, by their very nature,
highly reliant on local similarities, and consequently have problems generalis-
ing to novel patterns which do not resemble items they have encountered in
the training set. Research on the Polish case system suggests that this is also a
feature of linguistic generalisation in humans, so this property of connection-
ist models is an argument in their favour, not against them.
Polish speakers restricted productivity with neuter inections also has
DABROWSKA text Make up 26/3/04 4:35 pm Page 148
It is true that the semantic features of the verb are not relevant in English.
But the pattern-extraction mechanism cannot be blind to semantics because
some inectional endings in some languages are associated with classes of
stems sharing certain semantic properties. The choice of the genitive inection
of masculine nouns in Polish, for example, to the extent that it is predictable
at all, depends primarily on semantic factors. Thus, the learning system cannot
simply ignore semantics: it must be able to identify which (if any) semantic fea-
tures are relevant to a particular inection.
But will giving the network access to semantic representations do the trick?
According to Pinker and Prince (1988) and Kim et al. (1994), the answer to
this question is an emphatic no. They point out that the choice of inection
depends not on the presence or absence of a particular semantic property, but
on the grammatical structure of the word. As explained earlier, denominal
verbs, deverbal nouns, bahuvrihi compounds, nouns derived from names and
so on take regular inections in English, even when they are homophonous
with irregulars. Thus, the past tense of spit put on a spit is spitted, not spat,
and the plural forms of sabre-tooth and Mickey Mouse are sabre-tooths and
Mickey Mouses, respectively, not sabre-teeth and Mickey Mice. The reason for
this, according to the dual-mechanism theory, is that these words are exocen-
tric (headless), and so information about irregular inections cannot be
passed on to the derived word. This generalisation, however, cannot be cap-
tured using semantic features, since headlessness is not a semantic property:
no aspect of meaning is shared by spit put on a spit, sabre-tooth, Mickey
Mouse and all other exocentric words. Endocentric compounds, on the other
hand, always take the same inections as their heads, regardless of their
meaning. For example, chessman and strawman pluralise like man, even
though they do not refer to adult male human beings; and become, overcome,
come about, come down with, come on and so on share the same irregularity as
DABROWSKA text Make up 26/3/04 4:35 pm Page 152
come (their past-tense forms are became, overcame, came down with, came on)
in spite of the fact that they designate very dierent types of events.
Semantic extensions also tend to take the same inection as the basic sense
on which they are based, however tenuous the link between the senses. This is
particularly dramatic in the case of verbs like come, go, have and get, which,
as Pinker and Prince put it, are magnicently polysemous . . . yet they march
in lockstep through the same non-regular paradigms in central and extended
senses regardless of how strained or opaque the metaphor (1988: 11213).
Consider the verb get. The Collins COBUILD English Language Dictionary
lists 43 senses for the verb itself and 94 additional senses for various verbpar-
ticle combinations such as get o, get out and get up. All these senses have the
same past-tense form, got, because they all share the same head.
Kim et al. (1991, 1994) provide experimental evidence suggesting that
English speakers, including young children, are sensitive to the subtle dier-
ence between exocentric and endocentric words. In one experiment, they pre-
sented 35-year-olds with novel verbs homophonous with existing irregular
verbs. Some of the novel verbs were denominal (for example, the experimenter
announced that he was going to y this board, and proceeded to put ies all
over the board), and some were semantic extensions (the experimenter said
that Mickey was going to y down the road and then had him drive very fast
down the road). Subsequently, the children were asked to complete a sentence
describing what had just happened (Mickey just . . .). The general nding was
that children were more likely to use the regular inection with denominal
verbs than with semantic extensions in other words, they were more likely to
say You ied the board than Mickey just ied down the road. However, the
results were very noisy. The children gave the expected regular response to
denominal verbs only 64 per cent of the time, and the expected irregular with
semantic extensions only 23 per cent of the time. They also inappropriately
regularised 47 per cent of the semantic extensions (e.g. Mickey just ied down
the road), and used the irregular form with 6 per cent of the denominal verbs
(e.g. You just ew the board, meaning You just put ies all over the board).
In another experiment, Kim et al. presented nouns with irregular roots as
names (e.g. Mr Tooth), as exocentric compounds (e.g. snaggletooth, a stued
toy with large teeth) or as ordinary nouns (purple tooth), and elicited plural
forms from slightly older children (7- and 8-year-olds). The children used the
expected irregular plural with 94 per cent of the ordinary nouns. Rather unex-
pectedly, however, they also used it with 70 per cent of the names and exocen-
tric compounds. The dierence in frequency of irregular responses is
signicant, showing that the children were sensitive to the distinction between
ordinary nouns and the same roots used as names or in exocentric com-
pounds; but the high proportion of irregular responses in the latter case is
problematic for the dual-mechanism theory.
Kim and colleagues argue that these results show that speakers base their
decision whether to use the regular or the irregular inection on purely gram-
DABROWSKA text Make up 26/3/04 4:35 pm Page 153
object, while ring to put a ring around designates a kind of action). Because
of this, the new form tends to be perceived as a dierent word and cannot
inherit the grammatical properties of the base; hence, speakers apply whatever
inection would normally apply to the new word.
The above proposal is quite similar to that put forward by proponents of
the dual-mechanism theory. The main dierence is that it explains grammati-
cal behaviour without appealing to an autonomous level of grammatical rep-
resentation. It also has an additional advantage, in that it can explain why
semantic extensions sometimes lose the grammatical properties of the moti-
vating sense and take a dierent inection, and why some irregular-sounding
words derived from another category are inected irregularly.
Situations of the rst kind arise when the extended sense is so dierent from
the motivating sense that the two are no longer perceived as the same word. It
was noted earlier (cf. example (2)) that the genitive form of tenor in Polish is
tenoru when it refers to a singing voice and tenora when it refers to a singer
with that voice. The contrast is quite systematic: words designating other
singing voices also take -u, while words designating singers take -a. In English,
the plural of louse, when used to refer to a group of particularly unpleasant
people, is louses, not lice; and the past tense of hang to execute by hanging is
hanged, not hung. Occasionally, when the link between the basic sense and the
extended sense is not very salient, speakers may be unsure whether or not they
are the same word. When the word mouse refers to a small grey rodent, its
plural form is mice. Many English speakers, however, hesitate to use the irreg-
ular plural when talking about more than one computer mouse, although they
tend to prefer it to mouses. It seems that the reason why they are so squeam-
ish about mice is that the two concepts are perceived as very distant. The com-
puter device and the rodent do share certain properties they are both small
and grey but they are very dierent kinds of things, so many speakers judge
such supercial similarities to be irrelevant. Because the link between the two
concepts is so tenuous, speakers are not sure whether they should be regarded
as the same word, and, consequently, they are not sure which plural form they
should use when talking about several pointing devices.12
Evidence that words derived from another category are not invariably given
regular inections can be found in a series of experiments conducted by
Ramscar (2002). In one experiment, participants were introduced to the nonce
noun frink, which was dened as the Muscovite equivalent of the Spanish
tapas; it is served in bars, and usually comprises chilled vodka and some salted
or pickled sh. They were then presented with a homophonous verb meaning
to consume a frink, and were asked to supply the past-tense form. A substan-
tial majority (73 per cent) produced an irregular form such as frank or frunk.
In another experiment, Ramscar had speakers rate the acceptability of past-
tense forms of real verbs which were either denominal (e.g. sink meaning to
hide in a sink) or semantic extensions of an irregular verb (e.g. the metaphor-
ical sense of sink, as in My hopes sank). Two other groups of participants rated
DABROWSKA text Make up 26/3/04 4:35 pm Page 155
the same verbs for denominal status and semantic similarity to the prototyp-
ical sense of an existing irregular. Ramscar found that semantic similarity and
grammatical status were highly correlated, and that speakers intuitions about
semantic similarity to the basic sense were a better predictor of the acceptabil-
ity of irregular past-tense forms than their intuitions about denominal
status.13
Thus the data discussed by Pinker, Marcus, Kim and their colleagues does
not necessarily imply that an abstract autonomous level of grammatical struc-
ture is necessary to explain speakers abilities. On the other hand, the facts
need to be accommodated in some way, and it is clear that they cannot be
accommodated in any existing connectionist network. Existing connectionist
models of the past tense do not represent words; and although they are good
at categorising, it is doubtful that they could learn to make the kind of cate-
gorisation judgements required in this task. To deal with the more complex
problems discussed in this section, more sophisticated models will be required
which have access to various types of linguistic and extralinguistic knowledge.
This brings us to the most serious problem yet. All existing connectionist
models are quite restricted in application. The models discussed in this chapter
produce past-tense forms of English verbs. Regiers network, briey described
in Chapter 7, learned to output terms corresponding to various locative rela-
tionships. Other networks had been trained, with varying degrees of success,
to identify word boundaries, assign words to grammatical categories, or
predict the next word in a sentence (Elman 1990, 1993; Gasser and Smith
1998). Thus, each of these models learned to carry out a specic and fairly
simple operation on a particular kind of input. It is not clear, however,
whether such models will scale up that is to say, whether they will be able
to perform more complex linguistic tasks.
More complex types of processing would obviously require more complex
networks. But a more complex network isnt simply a network with more units.
Merely adding units to existing networks is unlikely to solve the problem, for
several reasons. First, as new nodes are added, the number of connections
between them grows exponentially. This means that there are more and more
weights to be set, and as a result, adjusting the weights soon becomes compu-
tationally intractable. Therefore, larger networks will need to show more selec-
tive patterns of interconnectivity, more like the patterns found in the brain,
where there are assemblies of units richly connected with each other, but with
fairly sparse connections with units belonging to other assemblies.
More importantly, the various subtasks involved in language processing
inecting words, identifying word boundaries, accessing phonological forms,
parsing rely on dierent kinds of knowledge, and hence, in network terms,
require dierent connection strengths. A network trained to produce past-
tense forms cannot identify word boundaries, and vice versa. Training it to
perform the second task would involve altering connection weights, which
would interfere with performance on the rst task. Thus, if a network is to be
DABROWSKA text Make up 26/3/04 4:35 pm Page 156
able to perform a number of dierent tasks, various subparts will have to spe-
cialise. Complex problems may also require decomposition into smaller tasks,
and these, too, would have to be assigned to dierent subnetworks.
It follows that a more realistic model of human linguistic abilities will
require a certain degree of functional specialisation or modularisation.14
Unfortunately, it is not clear how such functional specialisation could be
achieved in networks. Presumably some architectural features will have to be
built in (as in Regiers model), but many leading gures in the eld believe that
functional specialisation could arise as a result of learning in other words,
that future models will be able to self-organise (see Elman et al. 1996).
All these problems notwithstanding, it is undeniable that connectionist
models have stimulated research and helped us to develop new ways of think-
ing about language development and language processing. It is important to
realise, however, that connectionism is not a theory of language processing or
development: it is an approach to modelling these and other phenomena.
Connectionist networks can be used to model learning in a variety of frame-
works, empiricist, nativist or constructivist. Specic models, of course, do
embody assumptions about how language is represented in the brain; however,
the theories embodied in existing networks are very simple, almost trivially
simple. The networks developed so far are useful research tools, but they are
grossly inadequate as models of human linguistic abilities.
This does not mean that more realistic models are impossible in principle.
Connectionism is a relatively young endeavour, and much of the emphasis so
far has been on the mechanics of modelling developing suitable algorithms,
architectures, training regimes, and methods of analysing network behaviour.
Future models, it is to be hoped, will implement more sophisticated theories
of language and language acquisition.
Notes
14. Note that the term modularisation is used here to refer to functional spe-
cialisation. The resulting modules, or subnetworks, are not necessarily
anatomically distinct. Moreover, they receive input from other subnet-
works, so they are not informationally encapsulated.
DABROWSKA text Make up 26/3/04 4:35 pm Page 159
9 Syntactic constructions
We saw in the previous chapter that there is no evidence for a strong dissocia-
tion between grammar and the lexicon in morphology: in principle, the same
mental mechanism can account for speakers ability to supply both regular
and irregular inections. It was also suggested that speakers extract patterns
at various degrees of abstraction and that associative memory plays a promi-
nent role in this process.
In this chapter, I will argue that the same may also be true of syntax. This
is seen most clearly in acquisition, so the main body of the chapter will be
devoted to discussing developmental data. We will begin with a summary of
studies showing that there is a very close relationship between lexical and
grammatical development, at least in the early stages of language acquisi-
tion. We will then look at research which might explain this relationship, and
examine converging evidence from a variety of sources suggesting that the
starting point of syntactic development is the acquisition of a repertoire of
rote-learned formulas, or multi-word phrases paired with meaning. Finally,
we will look in considerable detail at the development of a particular con-
struction (or rather family of constructions) English interrogatives in
order to see how more creative patterns develop from such formulas, and
conclude with some observations about lexically specic units in adult lan-
guage.
A series of large-scale studies conducted by Elizabeth Bates and her team have
revealed the existence of close ties between expressive vocabulary size and
grammatical development. Bates et al. (1988), in a longitudinal study of 27
children who were followed from the age of 10 months to 28 months, showed
that there was a very high positive correlation (+.83) between expressive
vocabulary size at 20 months and MLU (mean length of utterance) at 28
months. Indeed, the correlation coecient is as high as the correlations
between MLU measures obtained from dierent samples taken from the same
DABROWSKA text Make up 26/3/04 4:35 pm Page 160
On the other hand, they are also grammatical constructions: they have inter-
nal structure, and they may be partly underspecied and hence allow some
variation. We saw in Chapter 2 that the use of prefabs is one of the processing
shortcuts that help adults produce and understand complex utterances in real
time. I will now review evidence that children learn such chunks, and that they
play an important role in acquisition.
this is?, Where Daddy is?) and then questions with the full form of the copula
in the correct position (What is this?, Where is Daddy?) and eventually ques-
tions with both the contracted and the uncontracted form. Richards (1979)
and MacWhinney (1985) note that while 3-year-old children produce correctly
ordered strings of adjectives before nouns, performance often dips around the
age of 5 before improving again in the school years. The early correct perfor-
mance is attributable to the use of memorised strings such as great big and nice
little.
(1) a. I dont wanna dont throw. (Iwamura 1979, cited in Peters 1983)
b. Can I have break it? [wants to break up a tinker-toy contraction]
(Braine 1976)
c. I dont know wheres Emma gone. (Clark 1974)
d. I want I eat apple. (Clark 1974)
e. Wheres the boy brought the milk? [looking for the milk the boy
had brought] (Clark 1974)
In (1a), the child simply incorporated the previous speakers utterance (dont
throw) into the frame I dont wanna X. The next two utterances are constructed
by combining two phrase-sized units from the childs own repertoire. In (1b),
the general-purpose request formula Can I have X? is combined with what is
probably an unanalysed chunk (at the stage when the utterance was produced,
the child used the verb break only in the expression break it). In (1c), the noun
Emma is used to ll the slot in a well-established formula (Wheres X gone?),
and the resulting expression is combined with another formula I dont know X.
Utterances (1d) and (1e) were probably assembled by embedding a simple con-
structed utterance (I eat apple, the boy brought the milk) inside a formulaic
frame (I want X, Wheres X?).
In other cases, an utterance may be grammatical and its formulaic status is
only revealed when it is used in inappropriate circumstances. One of the rst
plural forms used by Clarks son Adam occurred in the sentence Take the cups
away, which he heard in the day nursery. He subsequently used it at home even
when there was only one cup. Her other son Ivan produced the utterance It
will hurt me after a door had slammed in front of him (Clark 1977). Clearly,
he was not using will to signal the future, but simply recycling an expression
he had heard earlier in a similar context. What is noteworthy about this par-
DABROWSKA text Make up 26/3/04 4:35 pm Page 163
ticular example is that the recycling did involve a certain amount of modica-
tion of the original utterance: he changed the pronoun to reect the change of
speaker.
Such errors have been reported quite frequently in the literature (Chiat 1986;
Clark 1974; Dale and Crain-Thoreson 1993; MacWhinney 1985; Petitto 1987;
Snow 1981), and there is no shortage of theories attempting to explain them.
One popular explanation is that reversals are a result of young childrens
inability to assume the point of view of another person; others have attributed
such errors to the unusual nature of deictic pronouns (unlike most other lin-
guistic expressions, they refer to roles, not categories or individuals), charac-
teristics of the input (in speech addressed to the child, you always refers to the
child and me to the caretaker, so unless the child is exposed to language
addressed to third parties, she has no evidence that pronouns refer to roles
rather than individuals), and to processing complexity (pronoun reversal is a
performance phenomenon, particularly likely when processing demands are
high).
All four factors probably play a role, and their relative importance may vary
over time and from child to child. However, I would like to suggest that the
single most important cause of pronoun reversals is the childs use of precon-
structed phrases.
Several lines of evidence point in this direction. First, cases of consistent
reversals are rare. Most children who reverse pronouns produce correct forms
most of the time, and comprehension tends to be good even in children who
make errors in production (Chiat 1982). This argues against the rst three
hypotheses. Secondly, reversals are particularly frequent in imitations. All the
utterances in (2) above are (deferred) imitations of parental utterances. (2a)
and (2b) are reduced renditions of the parents Ill carry you and sit on my knee;
(2c) is an exact repetition of an utterance used on a number of similar occa-
sions by the childs father; and (2d), an exact imitation of an utterance the
DABROWSKA text Make up 26/3/04 4:35 pm Page 164
child heard in a very similar context the previous day (he was standing at the
top of the stairs in the morning in pyjamas with fabric feet; he reproduced
the utterance the following morning while walking down the stairs, wearing
the same pyjamas). In a study of deictic pronoun use in linguistically preco-
cious children, Dale and Crain-Thoreson (1993) found that 52 per cent of all
reversals occurred in imitations, in spite of the fact that imitations made up
only 5 per cent of the childrens utterances. Many of the remaining utterances
with reversals could well be delayed imitations which werent recognised as
such because the original parental utterance was not captured on tape. For
example, it is likely that the children whose utterances are reproduced in (3)
below heard a parent say something like Do you want mummy to help you
now?/Should mummy help you now? and Do you want me to help you with this?
Dale and Crain-Thoreson argue that reversals result from processing overload
(the child fails to perform deictic shift in cognitively demanding situations),
and note that they are particularly likely to occur in syntactically complex sen-
tences. But if cognitive overload were the problem, reversals should be less fre-
quent in imitations, which require less processing capacity than productive
usage. A more plausible explanation, therefore, is that reversals occur when
children recycle an adult utterance, or parts of it. This often enables them to
produce more complex utterances, but may lead to reversal errors when the
adult utterance contains pronouns.1
It is worth noting that pronoun reversals are especially common in the
speech of two clinical groups blind and autistic children. Autistic children
are well-known for their use of unanalysed formulas, and so this aspect of
their speech supports the claim that pronoun errors are a tell-tale sign of holis-
tic processing. In the case of blind children, inappropriate use of pronouns
may be a consequence of a delayed development of a sense of self (see
Fraiberg 1977) or of limited access to contextual information.2 On the other
hand, there is also evidence that blind children rely on formulaic phrases to a
greater extent than seeing children (Dunlea 1989; Mulford 1988; Prez-Pereira
1994; Urwin 1984; see also Chapter 3), so this could also contribute to the
higher incidence of pronoun errors.
would appear in adult speech (in English, before open-class words). The fol-
lowing utterances produced by Seth, a child studied by Ann Peters (Peters
1995; Peters and Menn 1993), are a good example (the upper-case F in the
glosses marks the position of the ller):
Seth began using ller syllables while still in the one-word stage. During the
period from 19 months to 21 months, they comprised about 30 per cent of the
morphs in his speech (Peters and Menn 1993: 748). It is often dicult to iden-
tify a specic target for these llers, but their presence lent the entire utterance
a more adult-like prosodic contour. Gradually, the llers gained more pho-
netic substance and became more and more dierentiated and eventually they
evolved into various adult forms.
Peterss interpretation of the data was that Seth was unable to reproduce the
adult utterance in full phonetic detail and attempted to approximate it at the
prosodic level:
act as placeholders for grammatical morphemes which will emerge later, and
the frame itself embodies the childs emergent knowledge about the distribu-
tion of these morphemes.
How typical (or untypical) was Seth? Is there any evidence that other chil-
dren use this strategy as well? The use of ller syllables has been reported by
quite a few researchers working on the acquisition of a variety of languages
(Peters 2001); but it is likely that most cases go unnoticed, partly because llers
are most characteristic of holistic children (and holistic children tend to be
underrepresented in language-acquisition research), and partly because
researchers often ignore syllables which have no discernible function. For
example, Braine (1976: 43) notes that one of his subjects inserted short
unstressed syllables (usually [] or [d]) before words, both utterance initially
and medially. These syllables didnt seem to make any semantic contribution
to the utterance and occurred before verbs as well as before nouns, so they
could not be determiners. Because of this, Braine decided to ignore them.
Presumably others have done likewise. Since many researchers working on
grammatical development use orthographic transcriptions, they may not even
be aware of the presence of llers in the speech of the children they study
unless they have transcribed it themselves.4
etc. On the other hand, the object of get-it occurred both before and after the
verb: block get-it, bottle get-it, get-it hat, get-it birds.
The existence of such lexically specic patterns in the speech of young chil-
dren is well documented. What is less well known is that they persist for a long
time after the child has advanced beyond the two-word stage, and seem to play
an important role in the transition to multi-word speech. Thus, when Travis
began to produce utterances containing three or more words, they were nearly
always constructed by adding a single element to an existing formula or
embedding one formula inside another. For example, adding an actor nominal
to her ride X formula allowed her to produce utterances such as Big Bird ride
horsie; embedding Daddys X in see X resulted in novel expressions such as see
Daddys car; big X plus X stuck yielded big rock stuck; and so on.
Several facts about Traviss usage at this time suggest that her grammatical
knowledge remained verb-specic at least until age 2;0. She tended to use
dierent combinations of arguments with dierent verbs. Many verbs were
combined with one argument only: some only with an agent, others only with
a patient or only with a location. This cannot be explained away as a process-
ing limitation, since other verbs occurred in frames with two or more argu-
ments: for example, she consistently expressed all three arguments with the
verb gave (but, interestingly, not with give, suggesting that the past- and
present-tense forms were independent constructions). What is more, she expli-
citly marked locative relations in utterances with some verbs but not others:
for example, she would say Travis sit down chair (without a locative preposi-
tion) but Dana push me real high in a bagswing. Her use of verbal inections
was also restricted to specic verbs. She used 24 verbs in the past-tense form
and 23 in the progressive, but only four verbs were used with both endings. Her
total verb vocabulary at the time was over 150, so over two-thirds of her verbs
were never inected (Tomasello 1992: 254). Likewise, explicit agreement
marking was restricted to the copula and a few other verbs.
Tomasellos diary method allowed him to chart the emergence and gradual
development of his daughters formulas in ne detail, but it is dicult to draw
general conclusions from a single case study. How typical was Travis? Are
other children equally restrained in their generalisations, or do they form
abstract rules right from the beginning? Fortunately, we do have similar data
from other children which allow us to answer these questions.
Lieven et al. (1997) studied the distributional regularities in the speech of
11 children. They recorded all multi-word utterances until the child reached
25 word-based patterns. At this point, the childrens ages ranged from 1;7 to
2;9 and their MLUs from 1.41 to 3.75 (mean 2.41). The researchers then
divided all multi-word utterances into three categories: constructed (lexical
formulas which had occurred at least twice before), intermediate (the second
occurrence of a particular positional pattern, or the rst if both words had
been attested in single-word utterances), and frozen phrases (utterances which
contained at least one word which did not occur in any other expression). They
DABROWSKA text Make up 26/3/04 4:35 pm Page 168
found that lexical formulas accounted for 60.3 per cent of all utterances (range
51.371.9 per cent). A further 31.2 per cent were frozen phrases. Of the
remaining 8.4 per cent, about half showed no positional consistency, and the
other half did (and hence could be instantiations of emerging lexical frames).
The more complex sentences that children produce also tend to be quite for-
mulaic. A detailed analysis of utterances with nite complement clauses pro-
duced by children aged from 2 to 5 conducted by Diessel and Tomasello (2001)
revealed that the main clauses in such sentences are usually very short.
Typically, they consist only of a pronoun (I or sometimes you) and a present-
tense verb, with no auxiliaries or adverbials; and even the subject is sometimes
omitted. Diessel and Tomasello point out that the proposition in such utter-
ances is typically expressed by the complement clause, while the main clause
functions as an epistemic marker (I think, I bet), a deontic modality marker (I
hope, I wish) or an attention getter (see, look), so the structures are, in eect,
monoclausal, and can be paraphrased with a simple sentence:
This analysis is supported by the fact that what appears to be the subordinate
clause is virtually never introduced by a complementizer, and that the main
clause sometimes comes after the complement clause (as in Its in here, I
think).
The children in the study also produced some sentences which do appear to
involve true subordination. These sentences were much less formulaic: the
main clauses contained a variety of subjects and verb forms as well as modi-
ers and expressed an independent proposition, and the subordinate clauses
were sometimes introduced by complementizers. It is noteworthy, however,
that the vast majority of them had one of just three verbs say, tell, pretend
as the main predicate. This, argue Diessel and Tomasello, suggests that they
were licensed by fairly specic schemas like Im just pretending S, NP SAYS,
NP TELL NP S (where SAY and TELL stand for various forms of these two
verbs) rather than a fully general subordination rule, since if children had a
general rule, they would apply it to sentences containing other verbs as well.
Lexically specic units, then, are a ubiquitous feature of early production,
which strongly suggests that young childrens knowledge may be described in
lexically specic terms. However, it is also possible that children do, in fact,
have fully general, adult-like rules but for some reason do not use them in pro-
duction. In order to rule out this alternative possibility, Tomasello and col-
leagues have conducted a series of experiments in which they investigated
childrens ability to use unfamiliar words in novel ways. They found that even
children as young as 1;6 to 1;11 freely assimilate novel nouns to a variety of
constructions. Thus, if they are taught the nonce word wug referring to an
unfamiliar animal, they spontaneously use it in combination with items which
DABROWSKA text Make up 26/3/04 4:35 pm Page 169
are already part of their repertoire, producing utterances such as wheres the
wug?, I want my wug, wug did it and so on (Tomasello et al. 1997). However,
childrens performance in experiments involving novel verbs is dramatically
dierent: the under-3s tend to use novel verbs only in the construction in
which they heard them. For example, Brooks and Tomasello (1999) taught
children aged about 2;11 two novel verbs designating highly transitive actions.
One of the verbs was presented in the transitive construction (Big Bird tammed
the car) and the other in a passive construction (The car got meeked by Big
Bird). They then attempted to elicit transitive sentences by asking what the
agent was doing. They found that only 25 per cent of the children who heard
passive models were able to use the novel verb in an active transitive sentence
describing the same action with a dierent set of participants. However, when
tested on the verb they heard in the transitive construction, all of the children
succeeded in this task.
In other experiments, Tomasellos team attempted to elicit a full transitive
with a novel verb modelled in intransitive (The lion is gorping), imperative
(Gorp, lion!) and presentative constructions (Look, this is called gorping). The
results are very consistent: children aged 2;0 and below virtually never produce
full transitives with a novel verb modelled in another construction; the pro-
portion of children who succeed in such tasks increases steadily with age until
it reaches 80 per cent by about 4;0; after 4, progress levels o slightly, but by
about 8 years of age, virtually all children are able to use a novel verb in the
transitive (see Tomasello 2000). Berman (1993) obtained comparable results
with Hebrew-speaking children. In her experiment, the novel verb was mod-
elled in the intransitive construction and elicited in the transitive. In Hebrew,
this transformation is more complex, as it requires morphological changes to
the verb as well as a rearrangement of the arguments, and so, not surprisingly,
the task turned out to be even more dicult for children: only 9 per cent of
the youngest participants (2;9) were successful; the proportion rose to 38 per
cent by 3;9 and to 69 per cent by 8;0.
While most of the work with novel verbs focused on the transitive construc-
tion, the same methodology has also been used to study childrens productiv-
ity with other structures, with similar results. Brooks and Tomasello (1999),
for example, also attempted to elicit passives, and found that 85 per cent of the
children in the youngest group they tested (age 2;11) were able to produce a
full passive with a novel verb if the verb had been modelled in the passive (with
dierent NPs expressing the agent and patient roles). However, if they had
only heard the verb in active sentences, none of the children produced a full
passive, and only 12 per cent were able to produce a truncated passive (a
passive clause without the by-phrase).
The Brooks and Tomasello study also provides evidence bearing on another
acquisition issue. English-speaking children begin to produce full passives
fairly late typically between 4 and 5 years of age. One explanation for their
late appearance, often put forward by researchers working in the generative
DABROWSKA text Make up 26/3/04 4:35 pm Page 170
tradition (e.g. Borer and Wexler 1987), is that passives are dicult because
they involve movement, which is a late-maturing syntactic ability. An alterna-
tive account, preferred in usage-based approaches, attributes childrens di-
culties to the fact that full passives are very rare in the input: Gordon and
Chafetz (1990), for example, found only four tokens of this construction in
over 86,000 child-directed utterances. The fact that most of the under-3s in the
Brooks and Tomasello study were able to use full passives productively after
brief but intensive training (two 30-minute sessions containing about 100
models) suggests that the latter explanation is correct: that is to say, that fre-
quency in the input is the determining factor.
Further support for the idea that young childrens knowledge about word
order is lexically specic rather than general comes from comprehension
studies involving nonce words. In one such study (Akhtar and Tomasello
1997), children were taught a new activity involving two characters and an
apparatus; then they were given two new characters and the apparatus and
asked to make Cookie Monster dack Big Bird. The main nding was that while
older children (mean age 3;8) did very well, only three out of ten 2;9-year-olds
performed above chance.
But perhaps the most striking evidence for lexical specicity comes from
two studies in which novel words were presented in constructions with non-
canonical (i.e. non-English) word order. In the rst of these, conducted by
Nameera Akhtar (1999), children were taught novel verbs designating highly
transitive actions in syntactic frames with SOV or VSO word order (e.g. Elmo
the car dacking, Gopping Elmo the car) and then encouraged to describe scenes
involving dierent participants using the newly learned verbs. The question
was whether they would correct to the canonical SVO pattern (thus demon-
strating verb-general knowledge of word order) or use the unusual word order
they heard in the model sentences. Akhtar found that the oldest children
(mean age 4;4) virtually always used the SVO pattern, no matter what word
order was used in the model sentences, while the younger children (ages 2;8
and 3;6) corrected to the canonical word order or used the unusual construc-
tions they had been to exposed to during the experiment equally frequently.
However, all age groups consistently corrected to the canonical word order
when a familiar verb was presented in an SOV or VSO pattern (for example,
they would never say He the car pushing). Thus, the younger childrens use of
the unusual word order cannot be explained by appealing to irrelevant factors
such as reluctance to correct an adult. Another interesting nding was that in
nearly all the utterances with non-canonical orders, the agent and patient
arguments were expressed by lexical NPs rather than pronouns, whereas half
of the utterances that were corrected contained pronouns: that is to say, the
younger children would happily say Ernie the spoon dacking but not He it
dacking. This suggests that the corrected utterances may also have been pro-
duced using lexically specic schemas based on pronouns (such as Hes
VERBing it).
DABROWSKA text Make up 26/3/04 4:35 pm Page 171
1. Novel SV: novel verb in the canonical order, e.g. The cow bang;
2. Novel VS: novel verb in a non-canonical order, e.g. Bang the cow
(meaning The cow is wobbling out);
3. Familiar VS: familiar verb in non-canonical order, e.g. Jumping the cow
(The cow is jumping).
The results were broadly compatible with those of the previous study: the 3-
year-olds consistently used SV order in all three conditions, while the 2-year-
olds generally corrected to the canonical word order with familiar verbs (72
per cent of the time), but not with novel verbs (only 21 per cent of the time),
showing that they lacked a general SV schema they could rely on in produc-
tion. As in the Akhtar study, the non-canonical word order was virtually never
used with pronouns (i.e. children did not produce utterances such as Baed
she), while the majority of the utterances with corrected word order contained
pronouns; the 2-year-olds, in fact, were almost ve times more likely to correct
to the canonical word order when the subject was a pronoun. Also, when using
the non-canonical VS order the children never added the missing auxiliary, but
they did tend to add it when using the SV order. Interestingly, they were more
likely to add it after a pronominal subject than after a lexical subject (70 per
cent v. 26 per cent of the time in the 2-year-old group). These asymmetries
suggest that they have learned multi-word units such as Hes VERBing and
Shes VERBing.
There is, however, an interesting twist to these weird word order studies. In
both experiments, children were considerably more likely to reproduce the word
order they heard with a particular nonce word if it was compatible with the
canonical word order in English. Akhtars children produced approximately
four times as many sentences with novel verbs in the control SVO condition as
in the experimental conditions. Abbot-Smith et al. also recorded about four
times as many productive uses of the nonce verb in the Novel SV condition as
in the Novel VS condition. One explanation for this pattern of results is that the
children were actively avoiding the strange word order used by the experimenter,
which would suggest that they had some nascent verb-general knowledge, even
if it was too weak to be reliably used in production. Another possibility is that
knowing a large number of verbs associated with a particular construction
makes it easier to acquire new verbs occurring in the same construction. This
would entail that children make connections between their lexically specic pat-
terns from the very beginning, although they may need time and/or a critical
mass of exemplars before they extract a general rule.
Further evidence about the nature of young childrens syntactic representa-
tions comes from studies which used the preferential looking paradigm. This
DABROWSKA text Make up 26/3/04 4:35 pm Page 172
method involves presenting the child simultaneously with two videos depict-
ing two dierent scenes (e.g. a causative action such as a duck lifting a rabbits
leg and a non-causative action such as a duck and a rabbit each covering its
own eyes) and a test sentence which describes one of them (e.g. The duck is
gorping the bunny). The experimenter then measures how long the children
look at each screen. If they spend more time looking at the matching video, it
is concluded that they are able to extract some information from the syntactic
form of the construction. The greatest advantage of this method is that it can
be used with very young children (1314 months); the main disadvantage is
that it is not clear what kind of information children extract from the linguis-
tic stimuli (cf. Naigles 1990; Tomasello and Abbot-Smith 2002).
Using the preferential looking procedure, Hirsh-Pasek and Golinko (1996:
experiments 5 and 6) found that the oldest children in their study (mean age
289 months) looked longer at the video showing a causative event when pre-
sented with a novel verb in a transitive sentence, but performed at chance when
presented with intransitive sentence; the middle group (24 months) showed a
preference for the matching video with familiar verbs only, thus demonstrating
verb-specic knowledge; and the youngest group (19 months) showed no evi-
dence of understanding the sentential context with either familiar or novel
verbs. Kidd et al. (2001), who tested slightly older children (mean age 30
months), found that they performed above chance on both transitive and
intransitive sentences with nonce words. Another study (Naigles 1990) suggests
that the relevant knowledge may appear somewhat earlier, around 25 months.
Thus, the preferential looking ndings are broadly compatible with the pro-
duction studies described earlier, in that they suggest that lexically specic
knowledge precedes verb-general knowledge; they do, however, suggest that
verb-general knowledge emerges considerably earlier. It is important, however,
not to overinterpret these ndings. The fact that the children were inuenced
by the linguistic form of the sentences accompanying the video presentation
shows that they were picking up on some property or properties of the sen-
tences and linking them to the display but it is by no means clear what exactly
they were attending to. Specically, there is no strong evidence that they were
using word order to assign agent and patient roles: it is possible that they asso-
ciated the experimental sentences with fairly holistic representations of the cor-
responding visual displays. For instance, they could have linked transitive
utterances such as The duck is gorping the bunny to scenes involving two par-
ticipants doing dierent things, and intransitive sentences such as The duck and
the bunny are gorping to scenes in which the two participants do the same thing;
note that in the latter case, the relevant cue might not be word order but the
word and.
Further evidence suggesting that young childrens constructions are lexi-
cally specic while older children have more abstract representations comes
from an experiment by Savage et al. (submitted). In this study, children were
shown pairs of pictures. The experimenter described the rst picture using a
DABROWSKA text Make up 26/3/04 4:35 pm Page 173
particular construction (for example, a passive such as The car got smashed by
the tree), and then the child was asked to describe the second picture, which
showed dierent characters performing dierent actions. Children aged 4 and
above showed evidence of priming (they were more likely to produce a passive
sentence when describing the second picture if the had just heard this struc-
ture), but younger children did not. However, there was priming even in the
younger children in another condition, when there was lexical overlap between
the two sentences that is to say, when the experimenters sentence contained
lexical material that the child could use in her own sentence (e.g. He got
VERBed by it).
Thus, experimental research conrms that early grammatical knowledge is
lexically specic. It also shows that the emergence of general word-order
schemas is a very gradual process. The preferential looking studies suggest
that some form of verb-general knowledge of the transitive construction
begins to develop about age 2;6, but it takes another year or so before chil-
drens representations are strong enough to support production. Verb-general
knowledge about less frequent structures such as the passives takes even
longer to develop.
Later research has tended to corroborate Browns suggestion that the acqui-
sition of grammatical inections is a process of piecemeal addition rather than
sudden revelation. Bloom et al. (1980) reported that the irregular-past,
present-participle and third-person-singular inections tend to be used with
dierent populations of verbs: -ing predominantly with durative non-comple-
tive verbs such as eat, play and write; the irregular past with non-durative com-
pletive verbs such as nd, get and come; and -s predominantly with go and t
(as in This goes/ts in here). More recently, similar ndings have been reported
by Tomasello (1992: see above), Pizzuto and Caselli (1992), Caselli et al. (1993)
and Rubino and Pine (1998).
Rubino and Pines work is particularly interesting since it shows that verb-
specic learning can occur even at a time when the child already has general
rules. They studied the acquisition of subjectverb agreement by a 3-year-old
child learning Brazilian Portuguese. By this stage, the child appeared to have
mastered the agreement system of his language, since the overwhelming
majority of his verb forms had the correct inections: the error rate was only
3 per cent. However, Rubino and Pine were able to show that this low overall
rate was due to very good performance on a few very frequent categories (spe-
cically, the singular inections) and masked much higher error rates in other
parts of the system. Plural forms accounted for only 3.4 per cent of the verb
forms used by the child (which suggests that he avoided using the plural), and
had much higher error rates: 23 per cent for the rst person plural a gente and
43 per cent for the third person plural. (Second person plural forms did not
occur at all in their corpus of almost 1,500 utterances.) Moreover, seven out
of the twelve utterances exhibiting correct 3PL agreement contained one of
two very frequent verb forms. Thus, there is little evidence for the mastery of
plural inections: the child appeared to be learning the system piecemeal.
But perhaps the most convincing evidence for such mosaic acquisition
comes from studies of the development of rules for auxiliary verb placement
DABROWSKA text Make up 26/3/04 4:35 pm Page 175
was that although all of the children were familiar with some of the uses of all
modals and semi-modals, knowledge about the constructions in which a par-
ticular word could occur was not generalised from one member of the class to
another. The children generally performed well on high-frequency modals
(will, would, can, could, should) but error rates on may, might, shall and must
were very high: for example, they were 2560 per cent correct on negatives with
may, depending on age group, and 020 per cent correct on questions with the
same verb. When faced with the task of transforming You may go home into a
question, most of the children (20) simply substituted a more familiar modal
with a similar meaning that is to say, they produced well-formed questions
with will, would, can or could. Six children produced the target structure (May
you go home?); one simply repeated the stimulus sentence using a rising into-
nation. Most revealing, however, were the ungrammatical sentences produced
by some of the children. Two children responded with Do you may go? and 11
added another modal at the beginning of the sentence, producing Can you may
go? Could you may go?, Will you may go?, or Would you may go? Dierences
between the modals showed up even on the imitation task. For example, 83 per
cent of the 5-year-olds imitated He should have nished his breakfast and We
could have gone to the show, but only 8 per cent were able to reproduce faith-
fully an analogous sentence with the auxiliary can (Sally can have chased the
dog). Even more patchy was the childrens command of tag questions. Every
child produced at least one acceptable tag, but error rates were very high, and
no child produced tags with may, might, shall or ought to. Moreover, some
added the same tag to nearly all sentences, which also suggests that they had
not fully mastered the structure.
Modals in WH questions show a similar pattern of development. An elic-
ited imitation experiment conducted by Kuczaj and Brannick (1979) showed
that children learn to place modals correctly rst after one or two WH words
(usually what and where), then after another word, and so on: in other words,
they learn the rule piecemeal. Similar results have been reported by Labov and
Labov (1978) and Ingram and Tyack (1979).
However, Kuczaj and Brannick also found that the children did eventually
generalise, or at least some did. Fifteen of their subjects (out of 60) overgen-
eralised the rule to questions with how come, which resulted in ungrammati-
cal utterances such as *How come will the wolf howl at the moon? However, six
of the children who overgeneralised failed to use modals at all with how long;
and ten correctly repeated questions with how come + past verb (How come the
turtle fell on the ground?) in other words, the overgeneralised rule was not
fully general!
There is also evidence suggesting that the knowledge that children acquire
is not only lexically specic but also construction-specic: that is to say, chil-
dren learn about specic words in specic constructions. Rowland and Pines
(2000) detailed analysis of WH questions produced by a single child revealed
that he tended to prepose some auxiliaries after some WH words and use
DABROWSKA text Make up 26/3/04 4:35 pm Page 177
declarative word order with other WH+AUX combinations, and that the com-
binations that were preposed were more frequent in the input than the non-
preposed ones. Fletcher et al. (2003), using an elicited production task, showed
that 3-year-olds performed very well on WH questions with the copula (which
are very frequent in the input), but were unable to produce structurally iden-
tical questions with the auxiliary should. However, when the specic structure
(i.e. WH questions with should) was modelled for the children in the course of
the experiment, performance improved dramatically. In another experiment,
the children were trained with inverted Y/N questions with is, can or should
and tested on WH questions with the same auxiliary. In this case, the training
did not improve performance: in other words, the children were not able to
generalise knowledge about auxiliary placement in Y/N questions to WH
questions.
Another telling example of piecemeal learning is provided by Todds (1982)
study of the acquisition of tag questions by his daughter Audrey. Audrey rst
began to use tags at 3;2, but except for two frozen forms which occurred very
early on, all her tags were armative. This resulted in ungrammatical utter-
ances when the tagged sentence was also armative: I look funny, do I?, Were
going home, are we? (both said with a rising intonation).5 She began to use neg-
ative tags productively at 4;2, but only with two verb forms, is and will. At 4;11,
she added wasnt NP? to her repertoire of negative tags. Finally, about 5;3,
negation spread to all auxiliaries used in tags and was overgeneralised to neg-
ative sentences, where she had previously supplied the correct form (*There
isnt very much people here, isnt there?).
Taken together, the studies reviewed here provide very strong evidence that
the development of auxiliary placement rules in English is a step-by-step,
bottom-up process involving the gradual accumulation of specic bits of
knowledge which are only later integrated into a system. Children do not
acquire an abstract rule of subjectauxiliary inversion: they learn about spe-
cic forms in specic constructions. They may subsequently develop broader
rules by generalising over these low-level patterns, but the early rules are highly
specic.
This conclusion is corroborated by a more comprehensive study of the
English auxiliary system by Richards (1990). Richardss monograph surveys
the development of auxiliary verbs in seven children over a period of some
eight months. Using recordings of spontaneous speech as well as an elicited
imitation task, he charts the emergence of all allomorphs of the core auxiliary
verbs in all syntactic contexts in which the auxiliary occurs: declaratives, neg-
atives, Y/N questions, WH questions, and emphatic and elliptical uses, includ-
ing tags. The study revealed wide-ranging individual dierences and a great
deal of piecemeal learning:
Across the board developments did not occur, though there were
periods of rapid development for some children. Instead, progress was
DABROWSKA text Make up 26/3/04 4:35 pm Page 178
We saw in the preceding section that childrens early use is dominated by xed
phrases and low-scope patterns. There is considerable disagreement, however,
about the relationship between these early formulas and the more general pat-
terns which emerge later. Some researchers (e.g. Bates et al. 1988; Brown and
Hanlon 1970; Krashen and Scarcella 1978) have argued that formulas are a
developmental dead-end, or at best a side street, a kind of linguistic crutch
which allows children to get by in the early stages of development, but which
is eventually superseded by real syntax with standard combinatorial rules.
Others (e.g. Lieven et al. 1992; Pine and Lieven 1993; Tomasello 2000, 2003)
have suggested that rote-learned phrases play a critical role in the development
of productive use. On this account, the transition to adult syntax is quite
gradual and does not involve qualitative shifts.
In this section, I will address this issue by examining in some detail the
acquisition of a particular construction, or rather family of constructions,
namely, English questions. Question constructions can provide more conclu-
sive evidence about the role of lexically specic learning in development than
argument structure constructions. It makes sense for children to be conserva-
tive when acquiring the latter, since dierent verbs have dierent subcategor-
isation requirements. Questions, on the other hand, share some very general
structural properties (often captured by means of two abstract rules: WH
fronting and subjectauxiliary inversion), so if children acquire fully general
knowledge from the start, this is where we would expect to nd evidence of it.
The acquisition of question constructions is also interesting from a theoreti-
cal point of view, since the two movement rules have played a very important
role in the development of the generative linguistics.
I have already noted that children sometimes acquire interrogative formu-
las such as Whats that? and Wheres X? in early stages of development, long
before there is any evidence of true syntax. We will now examine childrens
questions more systematically in order to determine how prevalent such for-
mulaic uses are, how long they persist, and whether there is any evidence of
developmental discontinuities in the transition to more productive usage
which might signal the acquisition of rules.
olds recorded for ve hours a week over a six-week period. This recording
regime produced samples which are ve to ten times denser than those used in
most studies and comprise 7 per cent to 10 per cent of each childs linguistic
experience during this period. With such relatively large samples, it is possible
to identify recurrent word combinations which are likely to be retrieved from
memory as ready-made units.
Lieven was interested in two types of preconstructed units, or formulas: lex-
ically xed phrases (e.g. Whats that?) and frames with a slot which can be
lled by a variety of expressions (e.g. Wheres X?). For the purposes of the
study, Lieven assumed that an utterance was formulaic if the corpus contained
at least three tokens of the same combination of words produced on three
dierent occasions. A frequency of three or more might not seem like very
much; however, extrapolating from the sample, we can estimate that the child
is likely to have produced 3040 tokens of the utterance in question during the
six-week period covered by the study, so the assumption does not seem unrea-
sonable.6
Lievens ndings are presented in Table 9.1. As we can see, the children had
from 6 to 19 dierent interrogative formulas, and these accounted for between
78 per cent and 98 per cent of their syntactic questions (mean 92 per cent).
The table also contains the results of an analysis of a comparable corpus of
data from a Polish-speaking child, Marysia (Dabrowska and Lieven 2003).
The gures are very similar, thus providing some cross-linguistic validation for
Lievens ndings.
Note: The figures for the English-speaking children (Brian, Chris, Annie and Delia)
are based on Lieven (2003). The figures for the Polish-speaking child, Marysia, are
based on Dabrowska and Lieven (2003).
The data for the four English-speaking children is arranged from least
advanced to most advanced, as measured by MLU. This reveals an interesting
pattern: the longer the MLU, the lower the proportion of xed formulas and
the higher the proportion of frames with slots. The most advanced child,
Delia, also has a relatively high proportion of non-formulaic utterances. This
DABROWSKA text Make up 26/3/04 4:35 pm Page 180
could mean that Delia had acquired some more general knowledge about
interrogative utterances. However, Lieven notes that all but two of Delias
non-formulaic questions could be traced back to utterances in the input in a
relatively straightforward way: that is to say, they were either exact repetitions
of things the mother had repeatedly said before, or were based on frames
which could have been derived from the input (for example, the utterance
Whats Mama making? may be based on a Whats Mama V-ing? frame, derived
from parental utterances such as Whats Mama doing?, Whats Mama
cooking? etc.).
Note: The figures in the column headings indicate the time span during which the
first 10 formulas emerged and the childs MLU at the beginning and the end of this
period. If there was no variation in a particular slot during this early period, the only
attested filler is given in parentheses after the formula. The very first recording in the
Shem corpus already contains 10 formulas, suggesting that they developed earlier.
(e.g. whats this/that?, wheres NP?) show up in all three. Secondly, there are
marked dierences in the rate of development. Naomi acquired her tenth
question formula at age 1;11, while Sarah needed another nine months to
reach the same milestone. Finally, and most importantly, formulaic utterances
also feature very prominently in later development: even at age 3 and 4, a sub-
stantial majority of the childrens utterances can still be accounted for in terms
of a relatively small number of lexically specic units.
Thus the longitudinal data appears to support the scenario proposed earlier,
namely, that development proceeds from xed phrases through formulaic
frames to more abstract patterns. Furthermore, there are no abrupt changes
in behaviour which might signal the acquisition of a general rule. What we see
instead is a slow but steady increase in the proportion of creative utterances,
suggesting a series of small changes. To understand what such small changes
might involve, it will be helpful to examine the development of a few formu-
las in more detail.
The corpus contains seven tokens of this formula in total. Two months later,
Sarah began using other verbs in this construction:
Meanwhile, four other formulas with who appeared in Sarahs speech: who
this?, who is it?, whos that? (all attested for the rst time between 2;8 and 2;11;
note that dierent pronouns were used with dierent forms of the copula),
and, a little later, who(s) goin(g to) VP?, rst attested at age 3;6. Finally,
almost a year and a half after the emergence of the earliest formula, we get
subject questions with who in sentences containing non-pronominal objects
and non-transitive verbs, e.g.:
Thus, development seems to have begun with two xed formulas, who did
it/that? and who broke it/that? The formulaic frame who TRANS.V it/that?
which emerged somewhat later seems to be a generalisation over the two lexi-
cally xed phrases (and possibly others which have not been captured in the
sample). The fully general template for producing questions with who in the
subject slot (who VP?) appeared considerably later, after Sarah had mastered
a number of more specic who-subject formulas.
After this, there are many questions with the names of various people and toys
in the subject slot. A week later, Naomi began substituting other verbs for
doing:
At this point, she could be said to have acquired a template with two slots
(Whats NP TRANS.V-ing?). Finally, just before her third birthday, she began
to use the uncontracted form of the auxiliary, which suggests that she had ana-
lysed whats into its component morphemes:
These early questions were quite stereotypical: the auxiliary was always placed
at the beginning of the sentence, and the subject, if present, was always the
rst person pronoun I; when the subject was left out, it is clear from the
context that the agent of the action was Naomi herself. This suggests that at
this stage, she did not have a general auxiliary placement rule, but merely a
formula for requesting permission.
With time Naomis usage gradually became more exible. First she acquired
a variant of the permission-seeking formula with the auxiliary could instead
of can. Like the canonical form of the formula, this variant was produced
sometimes with and sometimes without an explicit subject.
DABROWSKA text Make up 26/3/04 4:35 pm Page 185
Then, at 2;0.28 seven weeks after the rst recorded use of the formula we
nd the rst question with a subject other than I.
So, as in the previous two examples, early usage was highly stereotypical and
gradually became more exible as new slots opened up inside the formula.
dropped. Then, from about 2;6 onwards, it started rising again until, towards
the end of the period covered by this study, it reached levels similar to those
found in the input.8
The overall inversion rates in Figure 9.2 mask some interesting low-level
patterns suggesting that the children acquired knowledge about specic com-
binations of words rather than general rules. To show such eects, one would
ideally want to compare the proportions of inverted and uninverted questions
for each auxiliary + subject combination at dierent stages of development.
However, such an analysis would require a corpus containing a much larger
sample of each childs speech for each developmental period. Fortunately,
some patterns are discernible even if we collapse gures across auxiliaries or
subjects; and it is also possible to conduct a more ne-grained analysis of
questions with the most frequent auxiliary, namely do.
Table 9.5 shows the proportions of inverted yes/no questions with the most
frequent auxiliaries in the three children, and, for comparison, in one of the
mothers.9 To avoid collapsing across very long periods, and also to give some
sense of development, the Sarah corpus has been split into two parts: Sarah1
covers the period from 2;3 to 3;7, and Sarah2, from 3;8 to 5;1.
Auxiliary N % N % N % N % N %
Note: Percentages are provided only when there were at least eight questions with a
given auxiliary (inverted and uninverted).
a
Present-tense forms only
It is clear from the gures in the table that the children preposed some aux-
iliaries considerably more often than others. All three used the inverted form
by far the most frequently with can. This is also true of the mother, so the chil-
drens behaviour may be simply a direct reection of the input: they heard
both inverted and uninverted questions with the other auxiliaries, so they pro-
duced both types; with can, they virtually never heard uninverted questions,
so they rarely produced them themselves. The remaining dierences are partly
attributable to order of acquisition (since the data was collected over a period
DABROWSKA text Make up 26/3/04 4:35 pm Page 189
of time, if a child acquired the inverted pattern for a particular auxiliary fairly
late, the overall inversion rate for that auxiliary will be low) and partly to the
fact that dierent auxiliaries follow dierent developmental trajectories (for
example, although questions with do and did appeared in Sarahs speech at
about the same time, she used the latter much more consistently than the
former).
Such dierences between inversion rates for individual auxiliaries can be
accommodated in a rules-based approach: even if children acquire a general
rule, they still need to learn about the lexical properties of each auxiliary, and
it is perfectly possible that they acquire such knowledge at dierent times for
dierent auxiliaries. Much more problematic for such approaches is the fact
that there are also substantial dierences in the frequency of inversion with
dierent subjects.
Consider rst questions with do. Do is used when there is no other auxiliary
in the sentence, so the uninverted (i.e. intonation-only) analogues of such
questions normally lack an auxiliary (You want some?), except when they are
emphatic (You do want some?). Table 9.6 shows frequency of inversion with
1sg 0 0 0 0 21 44 0
2sg 29 52 53 54 92 35 153 61
3sg 7 23 1 5 45 52 43 67
Plural 0 1 7 11 48 11 55
36 38 55 38 169 40 207 62
Note: Percentages are provided only when there were at least eight questions with a
given subject (inverted and uninverted).
rst, second and third person singular and with plural subjects in the three
children and, for comparison, in one of the mothers.10 (The gures for plural
subjects have been collapsed to allow meaningful comparisons, as such ques-
tions are relatively rare.) As we can see, the mothers inversion rates are very
similar for all person/number combinations. The children, on the other hand,
put do before some subjects considerably more often than before others, which
is precisely what one would expect if they initially acquired knowledge about
particular combinations of subjects and auxiliaries.11
Moreover, there are dierences between the children in this regard. Naomi
and Shem invert by far the most frequently with second person subjects in
fact, Shem almost never uses do with any subject other than you. The most
DABROWSKA text Make up 26/3/04 4:35 pm Page 190
plausible explanation for this early prociency with second person subjects is
frequency in the input: children hear a great many questions of the form do
you VP?, so they have plenty of opportunity to pick up the pattern. Questions
with third person subjects are also quite frequent, but considerably more
varied (does it VP?, does he VP?, does she VP?, as well as some with lexical sub-
jects), so there are fewer opportunities to learn any particular pattern. Do
questions with rst person and plural subjects, on the other hand, are rare in
the input; and it is also these combinations that Naomi and Shem have most
diculty with.
Sarah, however, shows a very dierent pattern: she is least likely to invert
with second person subjects. It seems that several factors have conspired to
produce this eect. First, as is evident from Figures 9.1 and 9.2, Sarahs early
development was quite slow compared to the other children. She appeared to
have particular problems with the auxiliary do (cf. Table 9.5), possibly because
it was often contracted to [d] or [d ] in the input (as in dyou want an apple?),
and hence not very salient acoustically. Consequently, by the time she learned
her rst do-frames, she already had several well-entrenched intonation-only
formulas such as you want NP? and you know what? which often pre-empted
the emerging templates with auxiliaries. We see this very clearly when we
compare the frequencies of lexical verbs in inverted and uninverted questions.
The latter are dominated by high-frequency verbs: 52 per cent have either want
or know as the main verb, and 73 per cent contain one of just six verbs; the
corresponding gures for inverted questions are 29 per cent and 56 per cent
respectively. Thus Sarahs development, although it followed a dierent course
from that of the other two children, also shows evidence of lexically specic
learning.
Similar dierences in the proportion of inverted and uninverted questions,
depending on the subject, are found with other auxiliaries (cf. Table 9.7); and
Auxiliary N % N % N % N % N %
1sg 28 85 31 74 25 81 129 87 20 65
2sg 32 67 20 53 23 61 177 80 268 79
3sg 69 45 33 29 22 43 298 75 176 76
Plural 7 54 4 3 22 50 20 69
Total 136 55 88 44 73 59 626 77 484 77
Note: Percentages are provided only when there were at least eight questions with a
given subject (inverted and uninverted).
DABROWSKA text Make up 26/3/04 4:35 pm Page 191
again, the dierences are small and not statistically signicant in the mother,
and highly signicant in the children.12 It is interesting to note that the rela-
tive proportions of inverted and uninverted questions with dierent subjects
are dierent from with do: all three children invert most often in questions
with I, and least often with third person singular and plural subjects.
The pattern of inversion/uninversion with the pronoun I is particularly
interesting, and merits a more detailed discussion. Questions with rst person
subjects are relatively rare in the input. Only 3 per cent of the Y/N questions
addressed to Naomi had rst person subjects, and almost half of these had
the form can I VP? Children, on the other hand, ask such questions quite
often, thus oering the analyst an excellent opportunity to determine to what
extent they have generalised the patterns in the speech addressed to them. If
inversion rates with I are as high as with other subjects, this would suggest that
children have fairly general rules; and conversely, low inversion rates would
entail failure to generalise. As we saw earlier, children rarely produce syntac-
tic questions with rst person subjects and the auxiliary do; but inversion rates
for I with other auxiliaries are quite high (cf. Table 9.7). However, a closer
analysis reveals some striking patterns. The high inversion rates that we see in
Table 9.7 are due to the fact that the vast majority of the childrens questions
with rst person subjects are can I VP? questions the only frequent rst
person pattern in the input which all three children consistently invert (see
Table 9.8). If we exclude questions with can, inversion rates with rst person
subjects will be very low, except for the second half of the Sarah corpus, which
covers data for a relatively late period (3;85;1). Moreover, if we scrutinise the
constructions in Table 9.8 more closely, we will see that there is very little
overlap in the inverted and uninverted combinations of rst person subject
and auxiliary: that is to say, for any particular combination, each child either
consistently inverts (e.g. can I . . . ?) or consistently uses declarative word order
(Im . . . ?). (The only exception to this pattern is, again, the second part of the
Sarah corpus: but it is not surprising to nd evidence of more general knowl-
edge in a child of this age.) Thus, the pattern of inversion/uninversion with I
provides particularly clear evidence for piecemeal learning: children acquire
knowledge about particular subjectauxiliary combinations rather than
general rules which apply across the board.
3.4.2 Contraction
In addition to their canonical or full forms, the English copula and some aux-
iliaries have contracted forms (e.g. s, re) which cliticise to the preceding word.
The contracted forms predominate in informal adult speech, and are also
quite frequent in childrens earliest questions. Many children, however, go
through a stage when they prefer the full forms, sometimes after a period of
fairly consistent use of contracted forms. This marks a departure from the
adult pattern which is usually interpreted as indicating that the child has seg-
mented previously unanalysed forms (e.g. whats, wheres) into their component
DABROWSKA text Make up 26/3/04 4:35 pm Page 192
Table 9.8 Inverted and uninverted questions with first person singular subjects
Corpus Inverted questions Tokens Uninverted questions Tokens
Naomi can I . . . ? 25
could I . . . ? 3
I/me V-PRES . . . ? 8
Im . . . ? (copula) 1
Im . . . ? (auxiliary) 1
I cant . . . ? 1
Shem can I . . . ? 27
could I . . . ? 2
is I . . . ? 1
should I . . . ? 1
I V-PAST . . . ? 1
I V-PRES . . . ? 12
Im . . . ? (auxiliary) 3
Ill . . . ? 1
I cant . . . ? 1
I was . . . ? (copula) 1
I was . . . ? (auxiliary) 1
Sarah1 will I . . . ? 2
shall I . . . ? 1
can I . . . ? 18 I can . . . ? 1
did I . . . ? 4 I V-PAST . . . ? 2
do I . . . ? 1 I V-PRES . . . ? 33
I cant . . . ? 1
I didnt . . . ? 1
I be ..? (copula) 1
Sarah2 could I . . . ? 11
shall I . . . ? 1
am I . . . ? (copula) 7
was I . . . ? (copula) 1
can I . . . ? 77 I can . . . ? 2
should I . . . ? 5 I should . . . ? 1
did I . . . ? 21 I V-PAST . . . ? 6
am I . . . ? (auxiliary) 5 Im . . . ? (auxiliary) 4
do I . . . ? 21 I V-PRES . . . ? 27
dont I . . . ? 1 I dont . . . ? 1
I was . . . ? (auxiliary) 1
I cant . . . ? 1
I didnt . . . ? 1
I will . . . ? 1
Naomis didnt I . . . ? 3
mother shall I . . . ? 1
did I . . . ? 1
would I . . . ? 1
can I . . . ? 9 I can . . . ? 1
am I . . . ? (copula) 1 I am . . . ? (copula) 3
am I . . . ? (auxiliary) 1 I am . . . ? (auxiliary) 1
should I . . . ? 3 I should . . . ? 3
I shouldnt . . . ? 1
I cant . . . ? 2
DABROWSKA text Make up 26/3/04 4:35 pm Page 193
morphemes (what + is, where + is). What is particularly interesting from our
point of view is that this development appears to occur at dierent times in
dierent constructions.
We can see this when we compare contraction rates in dierent contexts at
dierent points in time. Ideally, one would want to follow changes in contrac-
tion rates in each question construction (e.g. what object questions with the
auxiliary is, what subject questions with the copula, etc.) but this would
require much larger corpora. With the data available, we can only follow the
combination of the most frequent WH words with the most frequent con-
tracted form, s. However, even when we collapse across dierent construction
types in this way, some patterns are clearly discernible. Figure 9.3 shows the
percentage of contracted forms after what and where in all three children and
after who in Naomi and Sarah (there is not enough data on who in Shem). The
ages on the X axis represent the midpoints of each period; and the nal point
represents the proportion of contracted forms in the input (i.e. the hypothet-
ical end result of development).
As we can see from the gure, adults used contracted forms equally fre-
quently after all question words, which suggests that any dierences in the chil-
drens levels of provision of such forms are not attributable simply to
phonological factors. There was some variation in the parents overall levels
of contraction, which presumably reect dialect dierences; however, within
the input to individual children, the levels of provision with dierent WH
words were very similar. Also, all three children, as well as their parents, con-
tracted after the pronoun that at the same levels (93 per cent or more of the
opportunities) throughout the entire period of observation. (This is not shown
in the gure to avoid cluttering it up.) On the other hand, the children used
contracted forms with dierent frequencies after dierent WH words, and,
what is more, the use of such forms developed along dierent trajectories in
dierent constructions.
The childrens questions with what follow a very similar developmental
pattern. All three initially preferred the contracted form in this context, and
supplied it about as frequently as their parents. This was followed by a period
of relatively frequent use of the full form, and later still, one presumes, by a
return to adult-like levels of provision, although we do not see this in the data
(except in Shem, where there is a slight rise at the last data point). Note,
however, that while the overall pattern of development is similar in all three
children, there are considerable dierences in the timing: the trough in the fre-
quency of contracted forms occurred about 2;9 in Shem and about 4;9 (or
perhaps even later, since this is the last data point) in Sarah.
Contraction after where, on the other hand, follows a dierent course of
development in each child. Sarah supplied contracted forms at about the
same rate throughout the entire period of observation. In Shem, we see a slow
but steady decrease in the frequency of such forms which continued even
when the frequency of contracted forms after what began to rise. Finally, the
DABROWSKA text Make up 26/3/04 4:35 pm Page 194
Naomi data shows a small decrease about 2;6, followed by a rise; note that
the frequency of contracted forms after what fell steeply during the same
period.
Contraction after who also follows a strikingly dierent pattern in the two
children for whom enough data is available. There are few questions with the
contracted form in Sarahs earlier transcripts; but after 3;6, her use of this
DABROWSKA text Make up 26/3/04 4:35 pm Page 195
3.4.3 Errors
Further evidence for piecemeal learning is to be found in the childrens errors.
Errors specically, errors of commission are of course prime evidence for
productive language use, since they involve the production of structures that,
by denition, do not occur in the input. However, this productivity is for the
most part very restricted. Most errors of commission involve agreement or
auxiliary placement. The former kind arise most often when a plural NP is
inserted into a frame containing a singular verb (e.g. wheres X? + deir feets =
wheres deir feets?). Auxiliary placement errors occur when a WH word is com-
bined with a clause containing an auxiliary (e.g. where X? + hes taking it? =
where hes taking it?).
Most other errors are very restricted in scope. The Shem corpus, for
example, contains 20 double-auxiliary errors, in which the auxiliary verb is
placed both before and after the verb. Twelve of these are one-o innovations
involving a variety of dierent combinations of auxiliaries, e.g. did . . . is (what
did hes showing?), does . . . be (where does these pictures be taken?), is . . . are
(why is dese clothes are blowing by?). The remaining eight, however, involve
exactly the same combination of words: what i(s) hes doing? All of these
questions have he as subject and doing as the main verb, the rst auxiliary is
rendered as [] or [z] and the second is always contracted.
These questions are productive in the sense that they are extremely unlikely
to be imitations of an adult utterance; but the fact that the same mistake recurs
again and again in the same combination of words, but not with other sub-
jects and other verbs, suggests that they are formulaic in other words, it
seems that Shem memorised his own creative utterance and reproduced it on
subsequent occasions.13
A similar example can be found in the Naomi corpus. Between the ages of
2;0 and 2;3, Naomi produced 15 WH questions in which the contracted form
of the auxiliary is was followed by a bare innitive. All of these had the form
Whats Nomi do? With all subjects other than Naomi, she consistently used the
correct form of the verb: Whats Mummy doing?, Whats sh doing?, Whats
recorder doing? and so on.
Even more striking is Naomis use of forms of self-reference. Like many
children, she sometimes referred to herself using the rst person pronoun and
sometimes used her name (which she pronounced Nomi). These two forms,
however, occurred in dierent contexts. Naomi consistently used I in declara-
tives (99 per cent of the time) and intonation questions (100 per cent) as well
DABROWSKA text Make up 26/3/04 4:35 pm Page 196
as questions with the modal auxiliaries can and could (also 100 per cent). On
the other hand, all but one of her 43 WH questions about herself contain her
name. They are all instantiations of three lexically specic formulas: Whats
Naomi do?, Whats Nomi doing? and Wheres Nomi? It is dicult to see how
such very specic restrictions in distribution could be explained without
appealing to preconstructed phrases.
Here what is construed as belonging with the last verb in the sentence, namely
needed. Because there can be any number of clauses between the question
word and the place where it logically belongs, it is dicult to account for such
sentences using schematic templates: indeed, sentences with LDDs were the
main motivation for postulating WH movement in the rst place.
Not surprisingly, questions involving LDDs are acquired late. As far as we
can tell from the transcripts, only one of the three children whose linguistic
development was discussed earlier in this chapter, Sarah, produced any such
structures only three in total, all recorded at age 4;4 or later. Given that such
questions are relatively infrequent, their absence in the transcripts of the other
two children could be an accident of sampling. However, Table 9.9, which con-
tains the relevant gures for ve other developmental corpora, suggests that
this is not the case. As we can see from the table, only two of these additional
children produced the relevant structure, and in both cases it appeared fairly
late: at 3;8 in the case in the case of Abe, who was a rather precocious child,14
and at 4;3 in Adams case. The late emergence of questions involving LDDs is
conrmed by experimental research. In a series of production studies, Crain
and Thornton (Crain and Thornton 1991; Thornton and Crain 1994) were
able to elicit this structure from most 4-year-olds, but only a few younger chil-
dren succeeded on the task.
It is also worth noting that the questions with LDDs were quite formulaic.
Of the 49 tokens produced by the children (44 by Abe, two by Adam and three
by Sarah), all but two had the form WH do you think S? There was some
variety within this schema, in that the children used dierent WH words to ask
about dierent parts of the subordinate clause.
DABROWSKA text Make up 26/3/04 4:35 pm Page 197
Note: The first three corpora contain samples from a number of children; the last
two from individual children. The figures include only questions containing a WH
word followed by an auxiliary and subject and a finite subordinate clause.
All but two of the questions, however, had think as the main verb, you as the
main verb subject, and do as the auxiliary. The same is true of the questions
elicited in the Crain and Thornton studies. Thus, by about age 4, children have
some knowledge about questions with LDDs; but it is not clear how general
such knowledge is.
Interestingly, adult LDD questions are also highly stereotypical. There are
325 tokens of this construction in the adult input in the Manchester corpus.
None of these contains more than one nite subordinate clause that is to say,
there are no sentences such as examples (14b) and (14c) above and only one
has an overt complementizer. In 96 per cent of the questions, the main verb is
either think or say; 91 per cent have you as subject (and 6 per cent have other
pronominal subjects); and 99 per cent have some form of do in the auxiliary
position. Moreover, specic main verbs appear to be associated with specic
auxiliary forms: think with do (276 out of 280 tokens), say and tell with did (31
out of 33 and six out of six tokens respectively), and pretend with shall (three
out of three tokens). The remaining four verbs (see, hope, reckon, suppose)
occurred only once in LDD questions, and hence could not occur with dier-
ent auxiliaries. In fact, even the verbs in the subordinate clauses are fairly
stereotypical, so that the following eight templates account for 50 per cent of
all LDD questions in the entire corpus (CAPITALS indicate that a particular
verb occurred in several dierent forms: for example, BE stands for is, was, are,
could be etc.):
DABROWSKA text Make up 26/3/04 4:35 pm Page 198
3.6 Conclusion
On the standard generative account, the acquisition of questions involves
learning to apply two rules, WH fronting and subjectauxiliary inversion. Both
of these rules are very general, but subject to certain constraints. Such con-
straints are thought to be unlearnable, and hence the fact that humans appar-
ently obey them is sometimes regarded as providing the clinching evidence for
the existence of innate Universal Grammar (Cf. chapter 6, section 4.1).
The evidence reviewed in this chapter, however, suggests a very dierent view
of development. We saw that early usage is highly stereotypical and that devel-
opment proceeds from invariant formulas through increasingly general formu-
laic frames to abstract templates. Furthermore, development is piecemeal, in
the sense that children learn the correct word order independently for each con-
struction,15 and the same changes occur at dierent times in dierent con-
structions. We also found no evidence of abrupt changes in behaviour which
might indicate a shift to a dierent productive mechanism: instead, we see slow,
steady progress. This suggests that the endpoint of development that is to say,
adult grammar might be rather similar to the kinds of representations that
children use; in other words, that adult knowledge comprises not constraints
on movement and suchlike, but rather fairly general templates like Can NP
VP?, Who VP? and What BE NP TRANS.V-ing? These would also allow
speakers to produce and understand a wide variety of dierent question types,
thus accounting for the observed exibility of adult behaviour. However, they
do not involve movement, and hence do not require innate constraints or an
abstract level of syntactic representation distinct from surface structure.
Some linguists have argued that speakers ability to produce and understand
questions involving LDDs constitutes evidence against schematic templates,
since such dependencies are unbounded: that is to say, there is no limit (in
principle) on the number of clauses intervening between the WH word and the
place where it logically belongs. Clearly, an adequate account of LDD ques-
tions must explain how such structures are produced and understood.
However, our examination of adult use of such structures revealed that they
are in fact highly formulaic, so the data actually supports the hypothesis that
even adults rely on lexically specic templates.
It should be stressed that the use of such frames requires grammatical
knowledge. Speakers must know that certain slots can only be lled with items
which have certain grammatical properties (for example, the VP in who VP?
must be tensed, while the VP in can NP VP? cannot); and there are often rela-
tionships between items in dierent grammatical slots (for example, in What
BE NP TRANS.V-ing? the auxiliary must agree with the subject in person and
number). The challenge for linguistics, then, is to develop a theory of grammar
which will capture such knowledge in psychologically realistic terms. The next
chapter introduces one such theory.
DABROWSKA text Make up 26/3/04 4:35 pm Page 201
Notes
of the time at the beginning of the recording period, while Naomis lowest
rates were about 40 per cent).
9. Naomis mothers questions were used for this comparison. The data ana-
lysed here consists of a corpus of the mothers questions matched in size
to the childs corpus.
10. There are only four Y/N questions with do in the rst half of the Sarah
corpus, so this part of the data was not included in Table 9.6.
11. The dierences are not statistically signicant in the mother: 2 = 1.2, df
= 2, p = 0.550. However, they are signicant in all three children: for
Naomi 2 = 12.6, df = 2, p = 0.002; for Shem, 2 = 31.7, df = 3, p < 0.001;
for Sarah2, 2 = 8.3, df = 3, p = 0.039.
12. For the mother, 2 = 4.2, df = 3, p = 0.236; for Naomi, 2 = 20.6, df = 3,
p < 0.001; for Shem, 2 = 27.3, df = 3, p < 0.001; for Sarah1, 2 = 13.5, df
= 3, p = 0.004; for Sarah2, 2 = 28.0, df = 3, p < 0.001.
13. Note that hes was not an unanalysed chunk. Shem did not use it in con-
texts which did not require an auxiliary or copula or in combination with
an auxiliary in declaratives.
14. Abes rst recorded use of a question with an LDD occurred at age 2;10,
but this was an imitation of the immediately preceding adult utterance.
15. It remains to be seen whether this is also true of later stages of language
acquisition. Development does speed up noticeably at about age 3, sug-
gesting that there may be transfer of knowledge between constructions
during this period.
DABROWSKA text Make up 26/3/04 4:35 pm Page 203
In the rst part of this book, we looked at some general properties of human
language that must be accommodated in any model of language that purports
to be psychologically realistic. Language processing is exible (we are able to
make use of whatever information is available and integrate various sources of
information as required) and very fast. It is also remarkably robust, in that it
develops even in very inauspicious circumstances and is dicult to destroy
completely, so that some linguistic abilities are preserved even after damage to
the parts of the brain which normally play a central role in language process-
ing. Last but not least, it relies to a large extent on relatively low-tech, general-
purpose mental mechanisms shared with other species, including the ability to
perceive, categorise and store information, form cross-modal associations,
and plan hierarchically organised action sequences.
In Part II so far, we have examined specic aspects of linguistic organisa-
tion in more detail, considering developmental evidence as well as research on
adult knowledge. This has enabled us to glean some hints about how these
properties are implemented in the language machine. One of the reasons why
language processing is so fast is because it relies to a considerable extent on
prefabricated units which need to be modied only minimally in order to be
used in communication. The exibility and robustness of the system are due,
in a large measure, to its redundancy: the fact that the same information is rep-
resented in dierent parts of the system, and hence there are typically dier-
ent ways of assembling the same utterance or arriving at its meaning. These
are tried out in parallel, which also makes for speed. We have also seen that
memory plays a very important role in language acquisition and production.
There is an intimate relationship between lexical and grammatical develop-
ment, between rote learning and rule extraction; and grammatical phenom-
ena, like words, are subject to frequency eects.
Much of the research discussed in Part II so far has focused on language-
specic aspects of linguistic knowledge. We have seen that such knowledge
can be quite subtle. It follows that we need a well-articulated descriptive
DABROWSKA text Make up 26/3/04 4:35 pm Page 204
framework to capture it, and a theory of acquisition which explains how humans
are able to learn the idiosyncratic features of their language as well as the more
general aspects which can be captured by means of universal principles.
This chapter is devoted to approaches to language which meet these require-
ments. While the requirements place strong constraints on linguistic theory,
they still allow for a number of alternatives. Rather than surveying all the pos-
sible options, I will focus on one particular theory, Cognitive Grammar
(Langacker 1987, 1991a, 1991b, 2000), which is the most fully elaborated of
these, although the discussion will also incorporate some key ideas from
related frameworks, notably Construction Grammar (Fillmore and Kay 1993;
Fillmore et al. 2003; Goldberg 1995) and Radical Construction Grammar
(Croft 2001). I will begin with an informal overview of this framework, and
then show how the kinds of representations that it postulates might develop
and be used in production. The chapter concludes with some general observa-
tions about the challenges facing cognitive linguistics.
Figure 10.1 The symbolic units []/[kt] (a) and [[]-]/[[kt]<[s]] (b).
Note: Boxes indicate semantic and phonological units; vertical lines correspond to
symbolic relationships. The < symbol indicates linear order.
DABROWSKA text Make up 26/3/04 4:35 pm Page 205
2.2 Imagery
The raison dtre of linguistic forms is to express meaning. Cognitivists take a
rather broad view of meaning as comprising not just content but also
imagery. This refers to our ability to construe the same content in alternate
ways by focusing on dierent aspects of a scene, viewing it from dierent per-
spectives, or conceptualising it at dierent levels of abstraction.
One of the most ubiquitous aspects of imagery is the imposition of a prole
on a base. The base is the knowledge structure within which an expression is
understood and dened; it is similar to Fillmores notion of frame (Fillmore
1975, 1978) and Lakos idealised cognitive model (Lako 1987). The prole
is a substructure within the base that an expression actually designates. For
example, the expression diameter designates, or proles, a line, but it also
evokes a larger cognitive structure a circle which is its base: it is impossible
to entertain the notion of without at the same time entertaining the
notion of . Two linguistic expressions can have the same base but dier-
ent proles and hence contrast in meaning. For example, the terms aunt and
niece designate dierent nodes in the kinship network which is their base; win,
lose, and beat prole dierent relationships in a knowledge structure we could
describe as a ; and winner, loser and runner-up prole indi-
viduals participating in these relationships. Conversely, two expressions can
have the same proles, but dier in meaning because the proles are imposed
on dierent bases. An example would be the adjective dry and the past parti-
ciple dried: both prole the same property (lack of moisture), but only the
latter has as its base a series of events in the course of which an object under-
goes a change of state (from wet to not wet).
Another dimension of imagery, crucial to the conceptualisation of rela-
tions, is trajector/landmark organisation. When we perceive relations, we tend
to impose a certain perspective on the participating entities. Specically, we
DABROWSKA text Make up 26/3/04 4:35 pm Page 206
tend to view one of the participants as the central element of a scene a gure
that stands out against a background of other entities. In CG, the most salient
entity is called the trajector and the other salient participants are called land-
marks. For instance, in (1), the Toyota is the trajector and the Mercedes the
landmark; in (2), the roles are reversed. In this example, the relationship
between the trajector and the landmark is spatial, but the terms are used to
refer to the corresponding roles in other kinds of relations as well. Complex
linguistic expressions usually have several levels of trajector/landmark organ-
isation. In (3), Chris is the trajector of put and the drink its landmark; but the
latter is also the trajector of on, while the table is its landmark.
In most cases, some element of the scene is inherently more salient than the
others and hence the most natural choice for the trajector. For example,
moving objects and agents are more likely to be construed as trajectors than
stationary objects or patients. This prototypical alignment is reected in (3),
where the agent is the clausal trajector and the displaced object the trajector
of the preposition. However, even when the relationship is inherently asym-
metrical speakers enjoy a certain degree of freedom and may choose the less
likely candidate as the gure. The humorous eect of (4), for example, is due
largely to the fact that the displaced object (the Scotch and soda) is construed
as landmark of outside, while the most natural choice for the landmark (the
container into which the Scotch is poured) becomes the trajector.
(4) He was seated at the Savoy bar when he told me this, rather feverishly
putting himself outside a Scotch and soda. (P.G. Wodehouse, The
World of Mr Mulliner)
Or consider a scene in which two cars, a Toyota and a Mercedes, crash into
one another. Although all of the sentences in (5) could be used to describe this
event, they are clearly not synonymous. Choosing one of the cars as the tra-
jector has the eect of highlighting its causal role in the crash and downplay-
ing the role played by the other car; construing both cars as trajectors results
in a more neutral description.
Similar observations can be made about the pairs of sentences given in (6)
and (7). In both cases, the two sentences describe dierent construals of the
DABROWSKA text Make up 26/3/04 4:35 pm Page 207
same event and hence dier in meaning, although the dierences are quite
subtle. Since the content words are the same in both members of each pair, it
is clear that the dierence in meaning is part of the meaning of the construc-
tion and the grammatical functors that form an integral part of it.
Yet another aspect of imagery is our ability to conceive the same content at
dierent levels of schematicity. For example, we can think of the same person
as , , or , or of the same activity as
, or . We are also able to make
categorising judgements such as (the concept is sche-
matic for ) and .
The term schema refers to the superordinate term in such relationships; the
subordinate term is called the schemas instantiation. These terms are more
general than the traditional labels superordinate and hyponym in that they
apply to all kinds of linguistic units, including phonological units (for
example, CVC is schematic for [kt], [nid], [f] and so on) and grammatical
units (see below).
Figure 10.2 A simple clause (a), a low-level schema (b), and an abstract schema (c)
DABROWSKA text Make up 26/3/04 4:35 pm Page 210
semantic pole of the schema, like that of the concrete expression, contains a
subpart corresponding to the trajector (the wanter) and the landmark (the
desired object), since they are part of the meaning of want (it is impossible to
think of wanting without simultaneously thinking of an animate being expe-
riencing this state and the object of their desire). Both of these participants
are things in the technical CG sense. The phonological pole spells out the seg-
mental content of the predicate and the linear order of the expressions corre-
sponding to the wanter and the desired object in relation to the predicate; the
actual segmental content of the expressions symbolising the trajector and
landmark is unspecied (this is represented in the diagram by three dots, which
act as place markers, or slots into which contentful expressions can be
inserted).
Finally, Figure 10.2c represents the abstract transitive construction, which
is highly schematic at both levels: the semantic representation indicates that
the construction proles a temporal relationship involving two things, and
the phonological representation spells out the linear order of the three com-
ponents. The actual segmental content of these component units is left
unspecied.
The three diagrams in Figure 10.2 represent the same content at varying
levels of abstraction: WANTER.NP want WANTED.NP is schematic for The
girl wants a cat, and NP V NP is schematic for both of these units. The crucial
point is that all three representations are symbolic units and have the same
overall structure. In this respect, the CG approach is very dierent from most
mainstream approaches to language, in which syntactic knowledge is repre-
sented in a dierent format from lexical knowledge, using constructs such as
phrase structure rules, movement rules, and various abstract principles and
constraints. In the CG framework, schemas and their instantiations are repre-
sented in the same format and dier only in the degree of specicity: the
schema is specied in less detail than its instantiations. It follows that schema
extraction involves reinforcement of shared properties and loss of detail,
rather than translation into a dierent representational format (see below).
units such as its raining cats and dogs. Likewise, the phonological form [kts]
and the semantic unit [[]-] consist of simpler elements and can form
parts of larger assemblies. Schematicity is the relation between the schema (the
superordinate term in a taxonomic hierarchy) and its instantiation (the subor-
dinate term). As in the case of composition, schematic relationship can hold
between semantic units (e.g. is schematic for , which in turn is sche-
matic for ), phonological units (CVC is schematic for the phonologi-
cal form [kt], which is schematic for actual phonetic realisations of the word),
and symbolic units (the plural construction [[]-]/[S] is sche-
matic for the expression cats, and the abstract symbolic unit []/[] is
schematic for both cat and cats).
All the examples given above involve full schematicity: the instantiations are
fully compatible with the properties of the schema, but are characterised in
more detail. This is not always the case: a linguistic expression (or phonolog-
ical form or conceptualisation) may not fully match the properties of the
schema, but still be included in the category dened by the schema. In such
cases, we speak of partial schematicity or extension. This is typically the rela-
tion between the dierent senses of a polysemous word. Consider, for
example, one of the basic senses of the verb y (move through the air by ap-
ping the wings) and an extended sense meaning utter (as in Her hair was
ying in the wind). Both senses designate situations involving a particular kind
of movement in the air, but in the extended sense, the moving object is
attached to something and is not displaced.
We saw earlier that polysemy is rampant in language (see Chapter 2, section
1; Chapter 7, section 1.2; Chapter 8, section 9), so the semantic representations
of most words are not single senses but families of related senses. This is illus-
trated in Figure 10.3, which shows some intransitive uses of the verb y and
the relationships between them. Note that each sense is related to at least one
other sense, and that there are several schemas (represented by the three boxes
in the upper part of the diagram) which capture some of the local similarities:
move through the air, move quickly, resist the force of gravity. However,
there is no single superschema which would be compatible with all the uses of
y.
Most constructional schemas, too, are complex categories in which the
various meanings are connected by relations of full or partial schematicity (cf.
Goldberg 1995; Lako 1987). For example, the caused motion construction
has at least ve interrelated senses (Goldberg 1995):
d. X Y / Z: Barbara locked
Michael out of the kitchen.
e. X Y / Z: Barbara helped Michael out of the
kitchen.
Senses (b)(e) share some similarities with the basic sense exemplied in (a)
which can be captured by means of local schemas. However, there are no fea-
tures common to all ve senses (apart from the fact that they describe some
kind of action; this, however, is a property shared with many other construc-
tions).
Finally, many words have variant pronunciations which share most but not
all features, and these form similar interconnected structures. Thus, most sym-
bolic units are not simply pairings of a phonological form with a meaning, but
rather pairings of a family of senses with one or more pronunciations sub-
networks within the larger network that constitutes the grammar of a lan-
guage.
expression (either the form or the meaning), and their task is to (re)construct
the other half: thus, production basically involves nding a phonological form
which matches ones communicative intentions, and comprehension, nding a
conceptualisation which matches the sounds one is hearing. In the simplest
case, a speakers mental lexicon will contain a unit with the appropriate
meaning, so all the speaker needs to do is to access its phonological form and
pronounce it. If the expression is also available to the listeners as a unit, all
they have to do after they have recognised it is access its semantic representa-
tion. On most occasions, however, a ready-made unit will not be available. In
such cases, the speaker will have to nd units which correspond as closely as
possible to parts of his or her communicative intentions and combine them
into a novel expression; and the listener will have to recognise the component
units, access their semantic representations, and integrate them into a coher-
ent conceptualisation. For the sake of simplicity, in the following discussion
we will look at the processing of novel expressions from the perspective of the
speaker; but many of the observations about the operations involved in sym-
bolic integration also apply to comprehension.
New expressions can be assembled by juxtaposing or superimposing estab-
lished units. Juxtaposition involves simply putting the two components next to
each other (in either order). For example, juxtaposing come here and now yields
come here now or now come here; juxtaposing Kate and would you like another
drink? yields either would you like another drink, Kate? or Kate, would you like
another drink? The relationship between the two components is paratactic
rather than truly syntactic: the fact that they are placed next to each other
signals that their meanings are to be integrated, but the grammar does not spell
out how this is to be done, so it must be inferred by the listener. Furthermore,
the grammar doesnt impose any requirements on the component expressions
(other than the fact that they should be elements capable of functioning as
utterances in their own right, given the right context). For example, now can
combine with a clause or a verb phrase (as in the examples above), with a prep-
ositional phrase (in America now) or just a noun phrase (now the gravy).
Superimposition, on the other hand, is a more constrained process in which
one component (a ller) elaborates a schematically specied subpart of
another component (the frame). For example, the expressions want and my
desk can be superimposed to derive the composite expression want my desk (cf.
Figure 10.4). Superimposition happens simultaneously at the phonological
and the semantic poles of the two components. In Figure 10.4, this is shown
by the dotted lines linking corresponding elements: elaborates the
landmark of the semantic structure [[] [ ]], and the
phonological form [ma
desk] elaborates the second slot in [wnt]. (Again,
the diagram is simplied: for example, it does not represent the internal struc-
ture of my desk.) Note that the ller must match the properties specied in the
frame: in our example the frame requires a ller designating a thing (in the
technical CG sense).
DABROWSKA text Make up 26/3/04 4:35 pm Page 216
For the sake of the following discussion, we will assume that the speaker has
the following symbolic units at his or her disposal:
The rst unit, Where AUX NONFINITE.S?, is simply the lexical representa-
tion of the interrogative pronoun where. The construction represents the
speakers knowledge about how where is used in sentences, including the fact
that it requires a preposed auxiliary. Note that this fact must be explicitly rep-
resented in the grammar, since not all question words impose this requirement.
(How come, for example, does not: we say How come she is coming?, not *How
come is she coming?) Unit (11) is the prototypical LDD question construction
DABROWSKA text Make up 26/3/04 4:35 pm Page 218
discussed in Chapter 9 (section 3.4); the symbol S-WH stands for a tensed
clause with a gap (i.e., an unlled slot). Unit (12) is a prefab containing a single
slot for a locative expression.
To derive the target question, the speaker must combine the three units. The
combination of (10) and (11) is a straightforward case of superimposition
with mutual elaboration:
The next step involves the superimposition of (12) and (13). Note that these
two components have partially incompatible specications: both contain a
location substructure, but (12) requires that the corresponding phonological
unit come after going, whereas (13) species that it should come at the begin-
ning of the sentence, before the auxiliary. To create a composite structure, the
speaker must resolve the conict, which can be done in two ways: either the
specications of (12) override those of (13), or vice versa. These two options
correspond to two dierent constructions. In ordinary questions, the WH
word is used as the framing unit and hence its requirements prevail: the result-
ing expression has interrogative word order and is used as a request for infor-
mation (14). In echo questions, the verb provides the frame, resulting in a
structure with declarative word order and a semi-declarative meaning: (15)
can be used either as a request to repeat part of the preceding utterance or
(with heavy stress on where) to express surprise, but not as a true question.
Note that the approach outlined above allows for alternative ways of assem-
bling the same expression. For example, the three components of where do you
think they are going? can be combined in a dierent order: the speaker could
begin by superimposing (11) and (12) to produce WH do you think they are
going LOCATION?, and then superimpose this and (10) to derive the target
utterance. And of course if they are going LOCATION is not available as a
unit, it would have to be constructed (for example, by superimposing they are
V-ing and NP go LOCATION).
As hinted in Chapter 9, it is likely that speakers also have some more general
knowledge about LDD questions, which might be captured by this more
abstract schema
DABROWSKA text Make up 26/3/04 4:35 pm Page 219
Such sentences, like the more prototypical LDD question discussed earlier,
consist of a WH word followed by an auxiliary and an untensed clause with a
gap, but in this case the missing element logically belongs in the doubly embed-
ded clause (the clause that functions as the complement of thought in (17) and
knew in (18)). To be able to form such questions, the speaker must extend the
category S-WH to include expressions such as Claire thought he did and Ben
knew they were going.
Figure 10.6 The emergence of a schema. (a) An unanalysed formula. The holistic
representation is open to alternative semantic analyses and may contain irrelevant
contextual detail. (b) Analysed formula. Aspects of semantic structure have been
mapped onto chunks of phonological material. The schema is implicit in the
formula, but not independently accessible. (c) A low-level schema, or the lexical
representation of one sense of the verb want. Aspects of (b) shared with other want-
formulas have been entrenched through repeated use and acquired independent
status, making the schema available for production and interpretation of novel
expressions.
DABROWSKA text Make up 26/3/04 4:35 pm Page 222
To see how this property can help in acquiring language, let us assume that
our learner, who has already learned to associate the phonological form
[ju]<[wnt]<[mlk] with a holistic representation of the milk-oering situa-
tion, acquires a second formula with a similar function, e.g. I want (the) ball.
The two formulas share some semantic substructures (they are both used in
situations when someone wants something) and the phonological chunk
[wnt], while other parts (e.g. the visual and motor images associated with ball,
the phonological chunk [bl]) are not shared. Once the learner has acquired
two (or more) formulas with partially overlapping representations, interesting
things start happening. The parts shared by several formulas are activated,
and hence entrenched, every time the learner hears (and later, produces) any
formula that contains them. In our example, the learners representations of
the phonological chunk [wnt] and semantic relationship between the experi-
encer and the desired object become active every time either formula is used;
and of course [mlk] and [bl] also occur in other expressions, not just as
objects of the verb want, so they, too, are activated in a variety of situations.
The elements making up the substructures shared by several formulas are fre-
quently co-activated, and eventually become units in their own right. In this
way, shared representation results in analysis: the internal structure of the
formula becomes visible to the learner (Figure 10.6b). In other words, the
component parts which are implicit in the unanalysed formula gradually
become more salient because certain congurations of semantic and phono-
logical material frequently co-occur. At this stage, the learner may be said to
have a mini-grammar: a grammar of the expression you want milk.
This change has several important consequences. First, as the formula is
analysed into its component units, meaning becomes more compositional. In
our example, whatever meaning You want milk might have had above and
beyond that contributed by its components is not strengthened as a result of
sharing a representation with other formulas, and gradually becomes less
salient; as a result, the meaning of the component units (notably the relational
expression) is distilled from the formula. The contextual meaning may not dis-
appear altogether: adult lexicons contain many multi-word units which are
not fully compositional. These include not only idioms, but also regular
expressions whose meanings are more specic than the sum of the meanings
of their parts, such as I told you so (I told you this would happen and you
should have listened to my advice/warning) or forget it (forget it because it is
DABROWSKA text Make up 26/3/04 4:35 pm Page 223
not important). It may also survive as knowledge about the kinds of situa-
tions in which it is appropriate to use a particular expression (e.g. formal v.
informal, when speaking to a social superior, when speaking to a member of
the opposite sex, and so on) and about what we call connotation (knowledge
about the emotional associations of a particular expression).5
Another important consequence of analysis is that it is now easier for the
learner to acquire new formulas, since part of their representation is already
there. Analysis is basically a chunking process: complex congurations of sub-
structures become units in their own right. The child might have needed 50
repetitions to learn you want milk, 30 repetitions to learn I want (the) ball and
10 repetitions to learn you want yoghurt; subsequently he will be able to repro-
duce other instantiations of this lexically based pattern after just one or two
learning episodes (assuming he already knows the nominals).
Once a formula has been analysed into its component units, the learner has
acquired some knowledge about want constructions which one might describe
as an incipient schema. It is, however, a passive generalisation, not readily
available for use in production of novel expressions: it has a similar status to
the -en plural in English (which is found in only three nouns: oxen, children,
brethren) or the patterns underlying the smaller classes of irregular verbs (e.g.
the buyboughtbought pattern, which applies to six or seven verbs only).
There are good reasons for believing that such generalisations are captured in
our mental grammars, but they are not normally applied to new nouns and
verbs.
As observed earlier, high-frequency forms are more entrenched, and hence
easier to access. Since components shared by several stored units are necessar-
ily accessed more frequently than any of the units which contain them, it
follows that they will become more entrenched than the larger units, and, after
a time, they will become accessible as units in their own right (Figure 10.6c).
This has little impact on high-frequency formulas, which are easy to access
anyway, but has important consequences for other aspects of language use.
First, the learner now has an alternative method of retrieving low-frequency
formulas: they can be activated by accessing the component parts and allow-
ing them to activate the whole. Secondly, and much more importantly, the
schema can be combined with other established units in order to form novel
expressions in other words, it becomes productive.
It should be clear from the preceding discussion that the accessibility of a
component unit, and therefore the potential to use it in novel expressions, is a
matter of degree. The more accessible the schema, the easier it is to apply pro-
ductively, and hence the greater the likelihood that it will be used productively.
Schema accessibility depends on two main factors: specicity (as explained
earlier, more specic units are easier to access) and entrenchment, which is a
function of frequency (other things being equal, more frequent units are easier
to access).
However, even schemas which are generalisations over a relatively small
DABROWSKA text Make up 26/3/04 4:35 pm Page 224
5. More on regularity
The account outlined above emphasises stored units and low-level patterns
which are learned piecemeal. This seems to be at odds with a basic tenet of
modern linguistics, namely, that languages are highly systematic: in each lan-
guage, there are some broad regularities shared by many dierent elements,
characteristic of the language as a whole: word-order patterns, agreement rules,
structural parallels between dierent types of constructions (for example,
DABROWSKA text Make up 26/3/04 4:35 pm Page 226
interrogatives, relatives) etc. Where do such patterns come from, and what is
their status? If each verb is a mini-grammar acquired independently of other
verbs ([[]--[]]/[[]<[hv]<[]], [[]--
[]]/ [[]<[it]<[]], [[]--[]]/[[]<[rat]<[]], etc.), why
does English have a xed word order (SVO) shared by all (transitive) verbs, and
why do all verbs have to agree with their subjects? Why dont dierent verbs
impose dierent word-order and agreement requirements?
Before we address this question, it is worth pointing out that modern lin-
guistics has tended to overemphasise regularity. There is, in fact, a great deal
of irregularity in language: very few rules are completely exceptionless. This is
most evident in morphology, where the existence of minor patterns and out-
right exceptions is quite commonplace. What is more, many languages have
classes of words to which the normal morphological rules do not apply, or
apply only optionally. It was noted in Chapter 8, for example, that some Polish
nouns are not inected for case in other words, the usual rules of grammar,
which require inected forms in certain contexts, do not apply to these nouns.
One can, of course, say that the rules do apply, but their operation is invisible:
the words are inected for whatever case is required, but the inections are
realised as zero. This, however, is regularity by at: the linguists grammar is
regular but it is not clear that the language itself is.
Similar observations apply to syntax. The English passive construction is
sometimes described as a product of fully general rules which convert an
underlying structure in which the direct object follows the verb into one with
passive word order. Not all transitive verbs in English, however, can be used
in the passive construction: cf. the unacceptability of *A book is had by her,
*You are suited by this dress, *A good life was lived by Robin. One might wish
to claim that have, suit and live are not truly transitive but this amounts to
saying that a transitive verb is a verb which passivises, which makes the claim
that the passive construction is fully general vacuous.
Word-order and agreement patterns may also be to some extent lexically
specic. For example, in French, most adjectives come after the noun, but
some precede it, and a few can occur in either position, depending on their
meaning (cf. femme ancienne old woman/wife and ancienne femme ex-wife).
Even English, with its relatively rigid word order, is not fully consistent: in
addition to the normal adjectivenoun order, we also have nounadjective in
expressions such as secretary general, court martial, proof positive, sister-
german, battle royal. Or consider British Sign Language, which has three types
of verbs: the so-called plain verbs are not marked for agreement and are
strongly associated with SVO word order; agreement verbs, as their name sug-
gests, do require agreement (with both subject and object), and normally
appear in sentence-nal position; and spatial verbs are marked for location
and movement rather than subject and object (Sutton-Spence and Woll 1999).
Thus, in this case we have three dierent word-order and agreement patterns
in the same language.
DABROWSKA text Make up 26/3/04 4:35 pm Page 227
On the other hand, it is undeniable that there are many high-level regular-
ities in all human languages, and these require an explanation. As noted in
Chapter 6, grammars are to some degree shaped by external discourse pres-
sures and processing constraints: some utterances are preferred by language
users because they are easier to process or are less likely to lead to ambiguity,
and hence their use spreads. Since the same pressures act on dierent construc-
tions, the result is that individual constructions converge towards the same
points. Such functional pressures still leave considerable variety (since there is
generally more than one workable solution to a particular communication
problem), but they do reduce it since they eliminate some variants.
There are also pressures for systematicity. Systems in which there are several
competing patterns are inherently unstable, as it is virtually inevitable that
speakers will occasionally overgeneralise some of them, sometimes because of
memory failure, and sometimes to achieve a special eect. Such innovations
may be picked up by other speakers, which results in uctuation in the fre-
quency of the individual variants. However, once a particular variant gains a
clear advantage over the other(s), whether by chance or as a result of a lan-
guage fad, it will tend to spread (since, other things being equal, more
entrenched patterns are easier to access) and may eventually monopolise the
grammar. Of course, not all language change is regularising: processes which
result in the emergence of irregular structures are also very common other-
wise all mature languages would be completely regular. However, research on
the emergence of creole languages suggests that regularisation processes pre-
dominate in the early stages of the development of a language (Aitchison
1989; Mhlhusler 1986).
6. Future directions
The approach to language outlined above has much to recommend it, and has
been applied successfully to a variety of linguistic domains (see for example
Casad 1996b; Casad and Palmer 2003; Cienki et al. 2001; Foolen and van der
Leek 2000; Langacker 1991a; Rudzka-Ostyn 1988). However, Cognitive
Grammar is a relatively young endeavour, and is only beginning to address
certain issues, among them phenomena such as extraction, binding and
control, which have been the mainstays of generative linguistics. While there
is some promising work in these areas (see e.g. Deane 1993; Langacker 1995;
van Hoek 1997), it remains to be seen whether CG can provide a workable
alternative to generative accounts of these aspects of a speakers linguistic
knowledge.
Furthermore, although cognitive linguists are ocially committed to devel-
oping an account of language that is usage-based and rmly grounded in
human cognition, in practice, only a few have begun to go beyond the tradi-
tional introspective methods. It is to be hoped that more researchers will move
DABROWSKA text Make up 26/3/04 4:35 pm Page 228
Notes
child extracts does not contain the initial auxiliary, since it is phonologi-
cally indistinct and often omitted in informal speech. It is also possible, of
course, that the child does represent the full adult form (do you want milk?)
or just want milk. This will obviously aect subsequent development: if the
child extracts the larger unit, he will have more analysis to do, and if the
original unit is want milk, he will have to learn some larger units at a later
stage.
5. These non-compositional aspects of meaning are usually regarded as an
extra that somehow gets added to the meanings contributed by the lexical
items when utterances are interpreted in context. However, from a devel-
opmental point of view, it is the other way around: the contextual meaning
is more basic, and the meanings of individual lexical items are carved out
of it.
6. Gleitman makes a very similar point in her argument for syntactic boot-
strapping (the hypothesis that children discover verb meanings by observ-
ing which syntactic frames verbs can occur in: see Gleitman 1990;
Gleitman and Gillette 1995). However, the above discussion suggests that
our knowledge about the distributional properties of words is much more
specic than a purely syntactic subcategorisation frame.
DABROWSKA text Make up 26/3/04 4:35 pm Page 230
Bibliography
Bibliography 231
Baldo, J. V., Elder, J. T., Larsen, J., Dronkers, N. F., Redfern, B. and Ludy, C.
(2001), Is cognition intact in patients with aphasia?, Brain and Language, 79,
647.
Barsalou, L. W. (1992), Cognitive Psychology: An Overview for Cognitive Scientists,
Hillsdale, NJ: Lawrence Erlbaum.
Barsalou, L. W. (1993), Flexibility, structure and linguistic vagary in concepts: mani-
festations of a compositional system of perceptual symbols, in A. C. Collins, S. E.
Gathercole and M. A. Conway (eds), Theories of Memories, London: Lawrence
Erlbaum, pp. 29101.
Barsalou, L. W. (1999), Perceptual symbol systems, Behavioral and Brain Sciences, 22,
577609.
Barsalou, L. W., Yeh, W., Luka, B. J., Olseth, K. L., Mix, K. S. and Wu, L. (1993),
Concepts and meaning, Papers from the Parasession on Conceptual Representations,
Chicago Linguistic Society, 29, 2361.
Basso, A., Della Sala, S. and Fabrola, M. (1987), Aphasia arising from purely deep
lesions, Cortex, 23, 2944.
Bates, E. (1994), Modularity, domain specicity and the development of language,
Discussions in Neuroscience, 10, 13649.
Bates, E. (1997), On language savants and the structure of the mind, International
Journal of Bilingualism, 1, 16386.
Bates, E. (1999), Language and the infant brain, Journal of Communication Disorders,
32, 195205.
Bates, E. and Goodman, J. (1997), On the inseparability of grammar and the lexicon:
evidence from acquisition, aphasia, and real-time processing, Language and
Cognitive Processes, 12, 50784.
Bates, E. and Goodman, J. (1999), On the emergence of grammar from the lexicon,
in B. MacWhinney (ed.), The Emergence of Language, Mahwah, NJ: Lawrence
Erlbaum, pp. 2979.
Bates, E., Bretherton, I. and Snyder, L. (1988), From First Words to Grammar:
Individual Dierences and Dissociable Mechanisms, Cambridge: Cambridge
University Press.
Bates, E., Dale, P. S. and Thal, D. (1995), Individual dierences and their implications
for theories of language development, in P. Fletcher and B. MacWhinney (eds), The
Handbook of Child Language, Oxford: Blackwell, pp. 96151.
Bates, E., Wulfeck, B. and MacWhinney, B. (1991), Cross-linguistic research in
aphasia: an overview, Brain and Language, 41, 12348.
Bates, E., Harris, C., Marchman, V. and Wulfeck, B. (1995), Production of complex
syntax in normal ageing and Alzheimers disease, Language and Cognitive
Processes, 10, 487593.
Bates, E., Thal, D., Trauner, D., Fenson, J., Aram, D., Eisele, J. and Nass, R. (1997),
From rst words to grammar in children with focal brain injury, Developmental
Neuropsychology, 13, 44776.
Bavin, E. L. (1990), Locative terms and Warlpiri acquisition, Journal of Child
Language, 17, 4366.
Behrens, H. (2001), Learning multiple regularities: evidence from overgeneralization
errors in the German plural, in A. H.-J. Do, L. Dominguez and A. Johansen (eds),
Proceedings of the 26th Annual Boston University Conference on Language
Development, Somerville, MA: Cascadilla Press, pp. 7283.
DABROWSKA text Make up 26/3/04 4:35 pm Page 232
Bibliography 233
Bibliography 235
Bibliography 237
Corina, D. (1996), Sign language and the brain: apes, apraxia and aphasia, Behavioral
and Brain Sciences, 19, 6334.
Corina, D. (1998), Aphasia in users of signed languages, in P. Coppens, Y. Lebrun and
A. Basso (eds), Aphasia in Atypical Populations, Mahwah, NJ: Lawrence Erlbaum,
pp. 261309.
Cottrell, G. W. and Plunkett, K. (1994), Acquiring the mapping from meanings to
sounds, Connection Science, 6, 379412.
Crain, S. (1991), Language acquisition in the absence of experience, Behavioral and
Brain Sciences, 14, 597650.
Crain, S. and Lillo-Martin, D. (1999), An Introduction to Linguistic Theory and
Language Acquisition, Malden, MA: Blackwell.
Crain, S. and Thornton, R. (1991), Recharting the course of language acquisition:
studies in elicited production, in N. A. Krasnegor, D. Rumbargh, M. Studdert-
Kennedy and R. Schiefelbusch (eds), Biological and Behavioral Determinants of
Language Development, Hillsdale, NJ: Lawrence Erlbaum, pp. 32137.
Croft, W. (2001), Radical Construction Grammar: Syntactic Theory in Typological
Perspective, Oxford: Oxford University Press.
Crosson, B., Parker, J., Albert, K., Warren, R., Kepes, J. and Tully, R. (1986), A case
of thalamic aphasia with postmortem verication, Brain and Language, 29, 30114.
Culicover, P. W. (1999), Syntactic Nuts: Hard Cases, Syntactic Theory and Language
Acquisition, Oxford: Oxford University Press.
Culp, R., Watkins, R., Lawrence, H., Letts, D., Kelly, D. and Rice, M. (1991),
Maltreated childrens language and speech development: abused, neglected and
abused and neglected, First Language, 11, 37790.
Curtiss, S. and de Bode, S. (1998), Linguistic outcomes for hemispherectomized chil-
dren, in A. Greenhill, M. Hughes, H. Littleeld and H. Walsh (eds), Proceedings of
the Twenty-Second Boston University Conference on Language Development, Boston:
Cascadilla Press, pp. 12133.
Curtiss, S. and Tallal, P. (1991), On the nature of the impairment in language-impaired
children, in J. F. Miller (ed.), Research on Child Language Disorders: A Decade of
Progress, Austin, TX: Pro-ed, pp. 189210.
Curtiss, S., Katz, W. and Tallal, P. (1992), Delay vs. deviance in the language acquisi-
tion of language impaired children, Journal of Speech and Hearing Research, 35,
37283.
Cutler, A. and McQueen, J. M. (1995), The recognition of lexical units in speech, in
B. de Gelder and J. Morais (eds), Speech and Reading: A Comparative Approach,
Hove: Lawrence Erlbaum, pp. 3347.
Dabrowska, E. (1993), O jezykowej idealizacji swiata, Bulletin de la Socit Polonaise
de Linguistique, XLIX, 3542.
Dabrowska, E. (1996), The linguistic structuring of events: a study of Polish perfec-
tivizing prexes, in R. Dirven and M. Ptz (eds), The Construal of Space in
Language and Thought, Berlin: Mouton de Gruyter, pp. 46790.
Dabrowska, E. (1997), The LAD goes to school: a cautionary tale for nativists,
Linguistics, 35, 73566.
Dabrowska, E. (2000), From formula to schema: the acquisition of English ques-
tions, Cognitive Linguistics, 11, 120.
Dabrowska, E. (2001), Learning a morphological system without a default: the Polish
genitive, Journal of Child Language, 28, 54574.
DABROWSKA text Make up 26/3/04 4:35 pm Page 238
Bibliography 239
Dennis, M. (1980), Capacity and strategy for syntactic comprehension after left of
right hemidecortication, Brain and Language, 10, 287317.
Dennis, M. and Kohn, B. (1975), Comprehension of syntax in infantile hemiplegics
after cerebral hemidecortication, Brain and Language, 2, 47586.
Dennis, M. and Whitaker, H. (1976), Language acquisition following hemidecortica-
tion: linguistic superiority of the left over the right hemisphere, Brain and Language,
3, 40433.
Dick, F., Bates, E., Utman, J. A., Wulfeck, B., Dronkers, N. and Gernsbacher, M. A.
(2001), Language decits, localization, and grammar: evidence for a distributive
model of language breakdown in aphasics and normals, Psychological Review, 108,
75988.
Dickinson, D. D. (1984), First impressions: childrens knowledge of words gained
from a single exposure, Applied Psycholinguistics, 5, 35973.
Diessel, H. and Tomasello, M. (2000), The development of relative clauses in sponta-
neous child speech, Cognitive Linguistics, 11, 13151.
Diessel, H. and Tomasello, M. (2001), The acquisition of nite complement clauses in
English: a corpus-based analysis, Cognitive Linguistics, 12, 97141.
Dockrell, J. and Campbell, R. (1986), Lexical acquisition strategies in the preschool
child, in S. Kuczaj and M. Barrett (eds), The Development of Word Meaning, New
York: Springer-Verlag, pp. 12154.
Drews, E. (1999), Pitfalls in tracking the psychological reality of lexically based and
rule-based inection, Behavioral and Brain Sciences, 22, 10223.
Dronkers, N. N., Redfern, B. B. and Knight, R. T. (2000), The neural architecture of
language disorders, in M. S. Gazzaniga (ed.), The New Cognitive Neurosciences (2nd
edn), Cambridge, MA: MIT Press, pp. 94958.
Du Bois, J. (1987), The discourse basis of ergativity, Language, 63, 80555.
Dunlea, A. (1989), Vision and the Emergence of Meaning: Blind and Sighted Childrens
Early Language, Cambridge: Cambridge University Press.
Eberhard, K. M., Spivey-Knowleton, M. J., Sedivy, J. C. and Tanenhaus, M. K. (1995),
Eye-movements as a window into real-time spoken language comprehension in
natural contexts, Journal of Psycholinguistic Research, 24, 40934.
Ebersberger, I., Metzier, D., Schwarz, C. and Paabo, S. (2002), Genome-wide compar-
ison of DNA sequences between humans and chimpanzees, American Journal of
Human Genetics, 70, 14907.
Eeg-Olofsson, M. and Altenberg, B. (1993), Discontinuous recurrent word combina-
tions in the LondonLund corpus, in U. Fries, G. Tottie and P. Schneider (eds),
Creating and Using English Language Corpora. Papers from the Fourteenth
International Conference on English Language Research on Computerized Corpora,
Zrich, Amsterdam: Rodopi, pp. 6377.
Ellis, N. C. (2000), Frequency eects in language processing: a review with implica-
tions for theories of implicit and explicit language acquisition, Studies in Second
Language Acquisition, 24, 14388.
Elman, J. (1993), Learning and development in neural networks: the importance of
starting small, Cognition, 48, 7199.
Elman, J. L. (1990), Finding structure in time, Cognitive Science, 14, 179211.
Elman, J. L. (1999), The emergence of language: a conspiracy theory, in B.
MacWhinney (ed.), The Emergence of Language, Mahwah, NJ: Lawrence Erlbaum,
pp. 127.
DABROWSKA text Make up 26/3/04 4:35 pm Page 240
Elman, J. L., Bates, E., Johnson, M. H., Karmilo-Smith, A., Parisi, D. and Plunkett,
K. (1996), Rethinking Innateness: A Connectionist Perspective on Development,
Cambridge, MA: MIT Press.
Elstner, W. (1983), Abnormalities in the verbal communication of visually-impaired
children, in A. E. Mills (ed.), Language Acquisition in the Blind Child, London:
Croom Helm, pp. 1841.
Enard, W., Przeworski, M., Fisher, S. E., Lai, C. S. L., Wiebe, V., Kitano, T., Monaco,
A. P. and Paabo, S. (2002), Molecular evolution of FOXP2, a gene involved in
speech and language, Nature, 418, 86972.
Erman, B. and Warren, B. (2000), The idiom principle and the open choice principle,
Text, 20, 2962.
Ewers, H. (1999), Schemata im mentalen Lexikon: Empirische Untersuchungen zum
Erwerb der deutschen Pluralbildung, in J. Meibauer and M. Rothweiler (eds), Das
Lexikon im Spracherwerb, Tbingen: Francke, pp. 10627.
Farrar, M. J. (1990), Discourse and the acquisition of grammatical morphemes,
Journal of Child Language, 17, 60724.
Fenson, L., Dale, P. S., Resnick, J. S., Bates, E., Thal, D. and Pethick, S. (1994),
Variability in early communicative development, Monographs of the Society for
Research in Child Development, 59 (Serial No. 242).
Ferreira, F. (2003), The misinterpretation of noncanonical sentences, Cognitive
Psychology, 47, 164203.
Ferreira, F. and Clifton, C., Jr (1986), The independence of syntactic processing,
Journal of Memory and Child Language, 25, 34868.
Fiez, J. A., Tallal, P., Raichle, M. E., Miezin, F. M., Katz, W. F. and Petersen, S.
E. (1995), PET studies of auditory and phonological processing: eects of stim-
ulus characteristics and task demands, Journal of Cognitive Neuroscience, 7,
35775.
Fillenbaum, S. (1974), Pragmatic normalization: further results for some conjunctive
and disjunctive sentences, Journal of Experimental Psychology, 102, 5748.
Fillmore, C. (1978), On the organization of semantic information in the lexicon,
Papers from the Parasession on the Lexicon, Chicago Linguistic Society, 111.
Fillmore, C., Kay, P., Michaelis, L. A. and Sag, I. A. (2003), Construction Grammar,
Stanford, CA: CSLI Publications.
Fillmore, C. J. (1975), An alternative to checklist theories of meaning, Proceedings of
the First Annual Meeting of the Berkeley Linguistic Society, 12331.
Fillmore, C. J. and Kay, P. (1993), Construction grammar. Unpublished manuscript.
Fisher, S. E., Lai, C. S. L. and Monaco, A. P. (2003), Deciphering the genetic basis of
speech and language disorders, Annual Review of Neuroscience, 26, 5780.
Fletcher, S., Rowland, C. and Jackson, K. (2003), Why modelling does matter: the use
of elicitation to study wh-question production in children. Paper presented at the
Child Language Seminar, Newcastle, 911 July.
Fodor, J. A. (1970), Three reasons for not deriving kill from cause to die,
Linguistic Inquiry, 1, 42938.
Fodor, J. A. (1983), The Modularity of Mind, Cambridge, MA: MIT Press.
Fodor, J. D., Fodor, J. A. and Garrett, M. (1975), The unreality of semantic represen-
tations, Linguistic Inquiry, 4, 51531.
Fodor, J. A., Garrett, M., Walker, E. and Parkes, C. (1980), Against denitions,
Cognition, 8, 263367.
DABROWSKA text Make up 26/3/04 4:35 pm Page 241
Bibliography 241
Foley, W. A. and Van Valin, R. D. (1984), Functional Syntax and Universal Grammar,
Cambridge: Cambridge University Press.
Foolen, A. and van der Leek, F. (2000), Constructions in Cognitive Linguistics: Selected
Papers from the Fifth International Cognitive Linguistics Conference, Amsterdam
1997, Amsterdam: John Benjamins.
Fox, P. T., Petersen, S., Posner, M. and Raichle, M. E. (1988), Is Brocas area language
specic?, Neurology, 38 (suppl.), 172.
Fraiberg, S. (1977), The acquisition of language, in S. Fraiberg (ed.), Insights from the
Blind, London: Condor Book Souvenir Press, pp. 22147.
Frazier, L. (1987), Sentence processing: a tutorial review, in M. Coltheart (ed.),
Attention and Performance XII: The Psychology of Reading, Hove: Lawrence
Erlbaum, pp. 55986.
Friederici, A. D. (1995), The time course of syntactic activation during language pro-
cessing: a model based on neuropsychological and neurophysiological data, Brain
and Language, 50, 25981.
Gardner, H. (1974), The Shattered Mind, New York: Vintage.
Garnsey, S. M., Pearlmutter, N. J., Myers, E. and Lotocky, M. A. (1997), The contri-
bution of verb bias and plausibility to the comprehension of temporarily amgigu-
ous sentences, Journal of Memory and Language, 37, 5893.
Gasser, M. and Smith, L. B. (1998), Learning nouns and adjectives: a connectionist
account, Language and Cognitive Processes, 13, 269306.
Gathercole, S. E. and Baddeley, A. D. (1990), Phonological memory decits in lan-
guage disordered children: is there a causal relationship?, Journal of Memory and
Language, 29, 33660.
Gawlitzek-Maiwald, I. (1994), How do children cope with variation in the input? The
case of German plurals and compounding, in R. Tracy and E. Lattey (eds), How
Tolerant is Universal Grammar? Essays on Language Variability and Language
Variation, Tbingen: Niemeyer, pp. 22566.
Geeraerts, D. (1988), Where does prototypicality come from?, in B. Rudzka-
Ostyn (ed.), Topics in Cognitive Linguistics, Amsterdam: John Benjamins, pp.
20729.
Gibbs, R. W. (1980), Spilling the beans on understanding and memory for idioms in
conversation, Memory and Cognition, 8, 14956.
Gibbs, R. W. and Colston, H. L. (1995), The cognitive psychological reality of image
schemas and their transformations, Cognitive Linguistics, 6, 34778.
Gibbs, R. W., Jr and Gonzales, G. P. (1985), Syntactic frozenness in processing and
remembering idioms, Cognition, 20, 24359.
Givn, T. (1989), Mind, Code and Context: Essays in Pragmatics, Hillsdale, NJ:
Lawrence Erlbaum.
Gleitman, L. R. (1981), Maturational determinants of language growth, Cognition,
10, 10314.
Gleitman, L. R. (1990), The structural sources of verb meanings, Language
Acquisition, 1, 355.
Gleitman, L. R. and Gillette, J. (1995), The role of syntax in verb learning, in P.
Fletcher and B. MacWhinney (eds), The Handbook of Child Language, Oxford:
Blackwell, pp. 41327.
Goldberg, A. E. (1995), Constructions: A Construction Grammar Approach to
Argument Structure, Chicago: University of Chicago Press.
DABROWSKA text Make up 26/3/04 4:35 pm Page 242
Goldeld, B. A. and Reznick, J. S. (1990), Early lexical acquisition: rate, content and
vocabulary spurt, Journal of Child Language, 17, 17184.
Goldman-Eisler, F. (1968), Psycholinguistics: Experiments in Spontaneous Speech,
London: Academic Press.
Goodall, J. (1986), The Chimpanzees of Gombe: Patterns of Behavior, Cambridge, MA:
Harvard University Press.
Goodglass, H. (2001), The Assessment of Aphasia and Related Disorders, Philadelphia:
Lippincott, Williams and Williams.
Goodglass, H. and Kaplan, E. (1972), The Assessment of Aphasia and Related
Disorders, Philadelphia: Lea and Febiger.
Gordon, P. and Chafetz, J. (1990), Verb-based versus class-based accounts of action-
ality eects in childrens comprehension of passives, Cognition, 36, 22754.
Graham, M. D. (1968), Multiply Impaired Blind Children: A National Problem, New
York: American Foundation for the Blind.
Grant, J., Valian, V. and Karmilo-Smith, A. (2002), A study of relative clauses in
Williams syndrome, Journal of Child Language, 29, 40316.
Grodzinsky, Y. and Finkel, L. (1998), The neurology of empty categories, Journal of
Cognitive Neuroscience, 10, 28192.
Grosjean, F. (1980), Spoken word recognition and the gating paradigm, Perception
and Psychophysics, 28, 299310.
Grossman, M. (1980), A central processor for hierarchically structured material: evi-
dence from Brocas aphasia, Neuropsychologia, 18, 299309.
Gullo, D. F. (1981), Social class dierences in preschool childrens comprehension of
wh-questions, Child Development, 52, 73640.
Haiman, J. (1985), Iconicity in Syntax, Cambridge: Cambridge University Press.
Hall, W. S., Nagy, W. E. and Linn, R. (1984), Spoken Words: Eects of Situation
and Social Group on Oral Word Usage and Frequency, Hillsdale, NJ: Lawrence
Erlbaum.
Hampson, J. and Nelson, K. (1993), The relation of maternal language to variation in
rate and style of language acquisition, Journal of Child Language, 20, 31342.
Hamsher, K. (1998), Intelligence and aphasia, in M. T. Sarno (edn), Acquired Aphasia
(3rd edn), San Diego: Academic Press, pp. 34173.
Hare, M. and Elman, J. L. (1995), Learning and morphological change, Cognition,
56, 6198.
Hare, M., Elman, J. L. and Daugherty, K. G. (1995), Default generalization in con-
nectionist networks, Language and Cognitive Processes, 10, 60130.
Hart, B. and Risley, R. R. (1995), Meaningful Dierences in the Everyday Experience
of Young American Children, Baltimore: Paul Brooks.
Hart, B. and Risley, R. R. (1999), The Social World of Children Learning to Talk,
Baltimore: Paul Brookes.
Hauser, M. D., Chomsky, N. and Fitch, W. T. (2002), The faculty of language: what is
it, who has it, and how did it evolve, Science, 298, 156979.
Hawkins, J. A. (1994), A Performance Theory of Order and Constituency, Cambridge:
Cambridge University Press.
Heibeck, T. H. and Markman, E. M. (1987), Word learning in children: an examina-
tion of fast mapping, Child Development, 58, 102134.
Heine, B. (1997), Cognitive Foundations of Grammar, New York and Oxford: Oxford
University Press.
DABROWSKA text Make up 26/3/04 4:35 pm Page 243
Bibliography 243
Johnson, M. (1987), The Body in the Mind: The Bodily Basis of Reason and
Imagination, Chicago: University of Chicago Press.
Johnson, M. H. (1993), Constraints on cortical plasticity, in M. H. Johnson (ed.),
Brain Development and Cognition: A Reader, Cambridge: Blackwell, pp. 70322.
Johnson, M. H. (1998), The neural basis of cognitive development, in D. Kuhn and
R. S. Siegler (eds), Handbook of Child Psychology. Vol. 2: Cognition, Perception and
Language, New York: John Wiley and Sons, pp. 150.
Johnston, J. R. (1997), Specic language impairment, cognition and the biological
basis of language, in M. Gopnik (ed.), The Inheritance and Innateness of Grammars,
New York: Oxford University Press, pp. 16180.
Johnston, J. R. and Slobin, D. I. (1979), The development of locative expressions in
English, Italian, Serbo-Croatian and Turkish, Journal of Child Language, 6, 52945.
Jones, D., von Stienbrugge, W. and Chicralis, K. (1994), Underlying decits in
language-disordered children with central auditory processing diculties, Applied
Psycholinguistics, 15, 31128.
Jones, M. J. (1996), A longitudinal and methodological investigation of auxiliary verb
development. Unpublished thesis, University of Manchester.
Jones, S. S. and Smith, L. B. (1993), The place of perception in childrens concepts,
Cognitive Development, 8, 11339.
Jou, J. and Harris, R. J. (1991), Processing inections: dynamic processes in sentence
comprehension, Journal of Experimental Psychology: Learning Memory and
Cognition, 17, 108294.
Kaplan, J. (1998), Message posted to Funknet, Funknet Archive, 13 October 1998.
Retrieved 8 December 2003 from https://mailman.rice.edu/pipermail/funknet/1998-
October.txt.
Karmilo-Smith, A. (1998), Is atypical development necessarily a window on the
normal mind/brain? The case of Williams syndrome, Developmental Science, 1,
2737.
Karmilo-Smith, A., Grant, J., Bethoud, I., Davies, M., Howlin, P. and Udwin, O.
(1997), Language and Williams syndrome: how intact is intact?, Child
Development, 68, 24662.
Karmilo-Smith, A., Tyler, L. K., Voice, K., Sims, K., Udwin, O., Howlins, P. and
Davies, M. (1998), Linguistic dissociations in Williams syndrome: evaluating recep-
tive syntax in on-line and o-line tasks, Neuropsychologia, 36, 34351.
Katz, J. (1972), Semantic Theory, New York: Harper and Row.
Katz, J. (1987), Common sense in semantics, in E. Lepore (ed.), New Directions in
Semantics, London: Academic Press, pp. 157234.
Katz, J. J. and Fodor, J. A. (1963), The structure of semantic theory, Language, 39,
170210.
Kegl, J., Senghas, A. and Coppola, M. (1999), Creation through contact: sign lan-
guage emergence and sign language change in Nicaragua, in M. DeGraf (ed.),
Language Creation and Language Change: Creolization, Diachrony and Development,
Cambridge, MA: MIT Press, pp. 179238.
Keil, K. and Kaszniak, A. W. (2002), Examining executive function in individuals with
brain injury: a review, Aphasiology, 16, 30535.
Kertesz, A. and Osmn-Sgi, J. (2001), Manifestations of aphasic symptoms in
Hungarian, Journal of Neurolinguistics, 14, 31319.
Kidd, E., Bavin, E. L. and Rhodes, B. (2001), Two-year-olds knowledge of verbs and
DABROWSKA text Make up 26/3/04 4:35 pm Page 245
Bibliography 245
Kutas, M. and King, J. W. (1995), The potentials for basic sentence processing: dier-
entiating integrative processes, in T. Inui and J. L. McClelland (eds), Attention and
Performance XVI: Information Integration in Perception and Communication,
Cambridge, MA: MIT Press, pp. 50146.
Labov, W. and Labov, T. (1978), Learning the syntax of questions, in R. N. Campbell
and P. T. Smith (eds), Recent Advances in the Psychology of Langauge (Vol. 4B), New
York: Plenum Press, pp. 144.
Lachter, J. and Bever, T. (1988), The relation between linguistic structure and associa-
tive theories of language learning: a constructive critique of some connectionist
learning models, Cognition, 28, 195247.
Lai, C. S. L., Fisher, S. E., Hurst, J. A., Vargha-Khadem, F. and Monaco, A. P. (2001),
A forkhead-domain gene is mutated in a severe speech and language disorder,
Nature, 413, 51923.
Lako, G. (1987), Women, Fire and Dangerous Things: What Categories Reveal about
the Mind, Chicago: Chicago University Press.
Lako, G. and Johnson, M. (1980), Metaphors We Live By, Chicago: Chicago
University Press.
Landau, B. and Gleitman, L. R. (1985), Language and Experience: Evidence from the
Blind Child, Cambridge, MA: Harvard University Press.
Langacker, R. W. (1987), Foundations of Cognitive Grammar. Vol. 1: Theoretical
Prerequisites, Stanford, CA: Stanford University Press.
Langacker, R. W. (1991a), Foundations of Cognitive Grammar. Vol. 2: Descriptive
Application, Stanford, CA: Stanford University Press.
Langacker, R. W. (1991b), Concept, Image and Symbol: The Cognitive Basis of
Grammar, Berlin: Mouton de Gruyter.
Langacker, R. W. (1995), Raising and transparency, Language, 71, 162.
Langacker, R. W. (2000), A dynamic usage-based model, in M. Barlow and S. Kemmer
(eds), Usage-Based Models of Language, Stanford, CA: CSLI Publications, pp. 163.
Lebrun, Y. and Hoops, R. (1974), Intelligence and Aphasia, Amsterdam: Swets and
Zeitlinger B.V.
Levine, S. C. and Carey, S. (1982), Up front: the acquisition of a concept and a word,
Journal of Child Language, 9, 64557.
Levinson, S. C. (1994), Vision, shape and linguistic description: Tzeltal body-part ter-
minology and object description, Linguistics, 94, 791855.
Lieberman, P. (1991), Uniquely Human: The Evolution of Speech, Thought, and Seless
Behavior, Cambridge, MA: Harvard University Press.
Lieberman, P. (1998), Eve Spoke: Human Language and Human Evolution, London:
Picador.
Lieberman, P. (2002), On the nature and evolution of the neural bases of human lan-
guage, Yearbook of Physical Anthropology, 45, 3662.
Lieberman, P., Friedman, J. and Feldman, L. S. (1990), Syntax comprehension decits
in Parkinsons disease, Journal of Nervous and Mental Disease, 178, 3605.
Lieberman, P., Kako, E. T., Friedman, J., Tajchman, G., Feldman, L. S. and Jiminez,
E. B. (1992), Speech production, syntax comprehension, and cognitive decits in
Parkinsons disease, Brain and Language, 43, 16989.
Lieven, E. V. (1994), Crosslinguistic and crosscultural aspects of language addressed
to children, in C. Gallaway and B. J. Richards (eds), Input and Interaction in
Language Acquisition, Cambridge: Cambridge University Press, pp. 5673.
DABROWSKA text Make up 26/3/04 4:35 pm Page 247
Bibliography 247
Bibliography 249
(ed.), Phraseology: Theory, Analysis and Applications, Oxford: Clarendon Press, pp.
79100.
Mueller, R. A. G. and Gibbs, R. W. (1987), Processing idioms with multiple meanings,
Journal of Psycholinguistic Research, 16, 6381.
Mhlhusler, P. (1986), Pidgin and Creole Linguistics, Oxford: Blackwell.
Mulford, R. (1988), First words of the blind, in M. D. Smith and J. L. Locke (eds),
The Emergent Lexicon: The Childs Development of a Linguistic Vocabulary, New
York: Academic Press, pp. 293338.
Mller, R.-A. (1992), Modularism, holism, connectionism: old conicts and new per-
spectives in aphasiology and neuropsychology, Aphasiology, 6, 44375.
Mller, R.-A. (1996), Innateness, autonomy, universality? Neurobiological approaches
to langauge, Behavioral and Brain Sciences, 19, 61175.
Murray, A. D., Johnson, J. and Peters, J. (1990), Fine-tuning of utterance length to
preverbal infants: eects on later language development, Journal of Child Language,
17, 51125.
Nagell, K., Olguin, K. and Tomasello, M. (1993), Processes of social learning in the
tool use of chimpanzees (Pan troglodytes) and human children (Homo sapiens),
Journal of Comparative Psychology, 107, 17486.
Nagrko, A. (1998), Zarys gramatyki polskiej (ze sl-owotwrstwem), Warsaw: PWN.
Naigles, L. R. (1990), Children use syntax to learn verb meaning, Journal of Child
Language, 17, 35774.
Nelson, K. (1973), Structure and strategy in learning to talk, Monographs of the
Society for Research in Child Development, 38 (Serial No. 149), 1135.
Nelson, K. (1981), Individual dierences in language development: implications for
development and language, Developmental Psychology, 17, 17087.
Neville, H. J. and Bavelier, D. (2000), Specicity and plasticity in neurocognitive devel-
opment in humans, in M. S. Gazzaniga (ed.), The New Cognitive Neurosciences,
Cambridge, MA: MIT Press, pp. 8398.
Neville, H. J., Mills, D. L. and Bellugi, U. (1993), Eects of altered auditory sensitiv-
ity and age of language acquisition on the development of language-relevant neural
systems: preliminary studies of Williams syndrome, in S. Broman and J. Grafman
(eds), Cognitive Decits in Developmental Disorders: Implications for Brain Function,
Hillsdale, NJ: Lawrence Erlbaum, pp. 6783.
New, H. (1972), Ranking of constraints on /t, d/ deletion in American English: a sta-
tistical analysis, in W. Labov (ed.), Locating Language in Time and Space, New
York: Academic Press, pp. 3754.
Ni, W., Crain, S. and Shankweiler, D. (1996), Sidestepping garden paths: assessing the
contributions of syntax, semantics, and plausibility in resolving ambiguities,
Language and Cognitive Processes, 11, 283334.
Obler, L. K. and Gjerlow, K. (1999), Language and the Brain, Cambridge: Cambridge
University Press.
Ogden, J. A. (1988), Language and memory functions after long recovery periods in
left hemispherectomised subjects, Neuropsychologia, 26, 64559.
Osterhout, L. and Nicol, J. (1999), On the distinctiveness, independence and time
course of the brain responses to syntactic and semantic anomalies, Language and
Cognitive Processes, 14, 283317.
Paivio, A. (1986), Mental Representations: A Dual Coding Approach, New York:
Oxford University Press.
DABROWSKA text Make up 26/3/04 4:35 pm Page 251
Bibliography 251
Bibliography 253
Bibliography 255
Bibliography 257
to structural and meaning violations when monitoring for nouns and verbs in sen-
tences, Neuropsychology, 10, 50416.
Wegener, H. (1994), Variation in the acquisition of German plural morphology by
second language learners, in R. Tracy and E. Lattey (eds), How Tolerant is Universal
Grammar? Essays on Language Variability and Language Variation, Tbingen:
Niemeyer, pp. 26794.
Wegener, H. (1999), Die Pluralbildung im Deutschen: ein Versuch im Rahmen der
Optimalittstheorie, Linguistik online, 4, http://viadrina.euv-frankfurt-o.de/~
wjournal/3{I99/wegener.html.
Wells, C. G. (1979), Learning and using the auxiliary verb in English, in V. Lee (ed.),
Language Development, London: Croom Helm, pp. 25070.
Wells, G. (1985), Language Development in the Preschool Years, Cambridge:
Cambridge University Press.
Westfal, S. (1956), A Study in Polish Morphology: The Genitive Singular Masculine, The
Hague: Mouton.
Whiten, A., Custance, D. M., Gomez, J. C., Teixidor, P. and Bard, K. A. (1996),
Imitative learning of articial fruit processing in children (Homo sapiens) and chim-
panzees (Pan troglodytes), Journal of Comparative Psychology, 110, 314.
Wilson, B. and Peters, A. M. (1988), What are you cookin on a hot? A three-year-old
blind childs violation of universal constraints on constituent movement, Language,
64, 24973.
Wise, R. J. S., Chollet, F., Hadar, U., Friston, K., Honer, E. and Frackowiak, R.
(1991), Distribution of cortical networks in word comprehension and word
retrieval, Brain, 114, 180317.
Worden, R. (1998), The evolution of language from social intelligence, in J. R.
Hurford, M. Studdert-Kennedy and C. Knight (eds), Approaches to the Evolution of
Language: Social and Cognitive Bases, Cambridge: Cambridge University Press, pp.
14866.
Wulfeck, B. and Bates, E. (1991), Dierential sensitivity to errors of agreement and
word order in Brocas aphasia, Journal of Cognitive Neuroscience, 3, 25872.
Wulfeck, B., Bates, E. and Capasso, R. (1991), A crosslinguistic study of grammatical-
ity judgments in Brocas aphasia, Brain and Language, 41, 31136.
Yamada, J. E. (1990), Laura: A Case for Modularity of Language, Cambridge, MA:
MIT Press.
Zattore, R., Evans, A. C., Meyer, E. and Gjedde, A. (1992), Lateralization of phonetic
and pitch discrimination in speech processing, Science, 256, 8469.