Movatterモバイル変換


[0]ホーム

URL:


SEP home page
Stanford Encyclopedia of Philosophy

Innateness and Language

First published Wed Jan 16, 2008

The philosophical debate over innate ideas and their role in the acquisition of knowledge has a venerable history. It is thus surprising that very little attention was paid until early last century to the questions of howlinguistic knowledge is acquired and what role, if any, innate ideas might play in that process.

To be sure, many theorists have recognized the crucial part played bylanguage in our lives, and have speculated about the (syntactic and/or semantic) properties of language that enable it to play that role. However, few had much to say about theproperties of us in virtue of which we can learn and use a natural language. To the extent that philosophers before the 20th century dealt with language acquisition at all, they tended to see it as a product of our general ability to reason — an ability that makes us special, and that sets us apart from other animals, but that is not tailored for language learning in particular.

In Part 5 of theDiscourse on the Method, for instance, Descartes identifies the ability to use language as one of two features distinguishing people from “machines” or “beasts” and speculates that even the stupidest people can learn a language (when not even the smartest beast can do so) because human beings have a “rational soul” and beasts “have no intelligence at all.” (Descartes 1984: 140-1.) Like other great philosopher-psychologists of the past, Descartes seems to have regarded our acquisition of concepts and knowledge (‘ideas’) as the main psychological mystery, taking language acquisition to be a relatively trivial matter in comparison;as he puts it, albeit ironically, “it patently requires very little reason to be able to speak.” (1984: 140.)

All this changed in the early twentieth century, when linguists, psychologists, and philosophers began to look more closely at the phenomena of language learning and mastery. With advances in syntax and semantics came the realization that knowing a language was not merely a matter of associating words with concepts. It also cruciallyinvolves knowledge ofhow to put words together, for it's typically sentences that we use to express our thoughts, not words inisolation.

If that's the case, though, language mastery can be no simple matter.Modern linguistic theories have shown that human languages are vastlycomplex objects. The syntactic rules governing sentence formation andthe semantic rules governing the assignment of meanings to sentences and phrases are immensely complicated, yet language users apparently apply them hundreds or thousands of times a day, quite effortlessly and unconsciously. But if knowing a language is a matter of knowing all these obscure rules, then acquiring a language emerges as the monumental task oflearning them all. Thus arose the question that has driven much of modern linguistic theory: How could mere children learn the myriad intricate rules that govern linguisticexpression and comprehension in their language — and learn themsolely from exposure to the language spoken around them?

Clearly, there is something very special about the brains of human beings that enables them to master a natural language — a feat usually more or less completed by age 8 or so. §2.1 of this article introduces the idea, most closely associated with the work ofthe MIT linguist Noam Chomsky, that what is special about human brains is that they contain a specialized ‘language organ,’ an innate mental ‘module’ or ‘faculty,’ that is dedicated to the task of mastering a language.

On Chomsky's view, the language faculty contains innate knowledge of various linguistic rules, constraints and principles; this innate knowledge constitutes the ‘initial state’ of the languagefaculty. In interaction with one's experiences of language during childhood — that is, with one's exposure to what Chomsky calls the ‘primary linguistic data’ or ‘pld’ (see §2.1) — it gives rise to anew body of linguistic knowledge, namely, knowledge of a specific language (like Chinese or English). This ‘attained’ or ‘final’ state of the language faculty constitutes one's ‘linguistic competence’ and includes knowledge of the grammar of one's language. This knowledge, according to Chomsky, is essential to our ability to speak and understand a language (although, of course, it is not sufficient for this ability: much additional knowledge is brought to bear in ‘linguistic performance,’ that is, actual language use).[1]

§§2.2-2.5 discuss the main arguments used by Chomsky and others to support this ‘nativist’ view that what makes language acquisition possible is the fact that much of our linguisticknowledge is unlearned; it is innate or inborn, part of the initial state of the language faculty.[2] Section 3 presents a number of other avenues of research that have been argued to bear on the innateness of language, and shows how recent empirical research about language learning and the brain may challenge the nativist position. Because much of this material is very new, and because my conclusions (many of which are tentative) are highly controversial, more references to the empirical literature than are normal in an encyclopedia article are included. The reader is encouraged to follow up on the research cited and assess the plausibility of linguistic nativism for him or herself: whether language is innate or not is, after all, an empirical issue.


1. Chomsky's Case against Skinner

The behaviorist psychologist B.F. Skinner was the first theorist to propose a fully fledged theory of language acquisition in his book,Verbal Behavior (Skinner 1957). His theory of learning was closely related to his theory of linguistic behavior itself. He argued that human linguistic behavior (that is, our own utterances and our responses to the utterances of others) is determined by two factors: (i) the current features of the environment impinging on thespeaker, and (ii) the speaker's history of reinforcement (i.e., the giving or withholding of rewards and/or punishments in response to previous linguistic behaviors). Eschewing talk of the mental as unscientific, Skinner argued that ‘knowing’ a language isreally just a matter of having a certain set of behavioral dispositions: dispositions to say (and do) appropriate things in response to the world and the utterances of others. Thus, knowing English is, in small part, a matter of being disposed to utter “Please close the door!” when one is cold as a result of a draught from an open door, and of being disposed (other things being equal) to utter “OK” and go shut a door in responseto someone else's utterance of that formula.

Given his view that knowing a language is just a matter of having a certain set of behavioral dispositions, Skinner believed that learning a language just amounts to acquiring that set of dispositions. He argued that this occurs through a process that he calledoperant conditioning. (‘Operants’ are behaviors that have no discernible law-like relation to particular environmental conditions or ‘eliciting stimuli.’ They areto be contrasted with ‘respondents,’ which are reliable or reflex responses to particular stimuli. Thus, blinking when someone pokes at your eye is a respondent; episodes of infant babbling are operants.) Skinner held that most human verbal behaviorsare operants: they start off unconnected with any particular stimuli.However, they can acquire connections to stimuli (or other behaviors)as a result of conditioning. In conditioning, the behavior in question is made more (or in some paradigms less) likely to occur in response to a given environmental cue by the imposition of an appropriate ‘schedule of reinforcement’: rewards or punishments are given or withheld as the subject's response to the cue varies over time.

According to Skinner, language is learned when children's verbal operants are brought under the ‘control’ of environmentalconditions as a result of training by their caregivers. They are rewarded (by, e.g., parental approval) or punished (by, say, a failure of comprehension) for their various linguistic productions and as a result, their dispositions to verbal behavior gradually converge on those of the wider language community. Likewise, Skinner held, ‘understanding’ the utterances of others is a matter of being trained to perform appropriate behaviors in response to them: one understands ‘Shut the door!’ to the extent that one responds appropriately to that utterance.

In his famous review of Skinner's book, Chomsky (1959) effectively demolishes Skinner's theories of both language mastery and language learning. First, Chomsky argued, mastery of a language is not merely a matter of having one's verbal behaviors ‘controlled’ byvarious elements of the environment, including others' utterances. For language use is (i) stimulus independent and (ii) historically unbound. Language use isstimulus independent: virtually anywords can be spoken in response to any environmental stimulus, depending on one's state of mind. Language use is alsohistorically unbound: what we say is not determined by our history of reinforcement, as is clear from the fact that we can and do say things that we have not been trained to say.

The same points apply to comprehension. We can understand sentences we have never heard before, even when they are spoken in odd or unexpected situations. And how we react to the utterances of others is again dependent largely on our state of mind at the time, rather than any past history of training. There are linguistic conventions in abundance, to be sure, but as Chomsky rightly pointed out, human ‘verbal behavior’ is quite disanalogous to a pigeon's disk-pecking or a rat's maze-running.. Mastery of language is not a matter of having a bunch of mere behavioral dispositions. Instead, itinvolves a wealth of pragmatic, semantic and syntactic knowledge. What we say in a given circumstance, and how we respond to what others say, is the result of a complex interaction between our history, our beliefs about our current situation, our desires,and our knowledge of how our language works. Skinner's firstbig mistake, then, was in failing to recognize that language mastery involves knowledge (or, as Chomsky later called it ‘cognizance’) of linguistic rules and conventions.

His second big mistake was related to this one: he failed to recognize that acquiring mastery of a language is not a matter of being trained what to say. It's simply false, says Chomsky, that “a careful arrangement of contingencies of reinforcement by theverbal community is a necessary condition of language learning.” (1959:39) First, children learning language do not appear to be being ‘conditioned’ at all! Explicit training (such as a dog receives when learning to bark on command) issimply not a feature of language acquisition. It's only comparativelyrarely that parents correct (or explicitly reward) their children's linguistic sorties; children learn much of what they know about language from watching TV or passively listening to adults; immigrantchildren learn a second language to native speaker fluency in the school playground; and even very young children are capable of linguistic innovation, saying things undreamt of by their parents. AsChomsky concludes: “It is simply not true that children can learn language only through ‘meticulous care’ on the partof adults who shape their verbal repertoire through careful differential reinforcement.” (1959:42)

Secondly, Chomsky argued — and here we see his first invocationof the famous ‘poverty of the stimulus’ argument, to be discussed in more detail in §2.2 below — it is unclear that conditioning couldeven in principle give rise to a setof dispositions rich enough to generate the full range of a person's linguistic behavior. In order, for example, to acquire the appropriate set of dispositions concerning the wordcar, onewould have to be trained on vast numbers of sentences containing thatword: one would have to hearcar in object position andcar in subject position;car modified by adjectivesandcar unmodified;car embedded in opaque contexts(e.g. in propositional attitude ascriptions) andcar used transparently; and so on. But the ‘primary linguistic data,’ usually referred to as the ‘pld’ and comprising the set of sentences to which a child is exposed during language learning (plus any analysis performed by the child onthose sentences; see below), simply cannot be assumed to contain enough of these ‘minimally differing sentences’ to fully determine a person's dispositions with respect to that word. Instead,Chomsky argued, what determines one's dispositions to usecar is one's knowledge of that word's syntactic and semanticproperties (e.g.,car is a noun referring to cars), togetherwith one's knowledge of how elements with those properties function in the language as a whole. So even if language mastery were (in part) a matter of having dispositions concerningcar, the mechanism of conditioning would be unable to give rise to them. The training set to which children have access is simply too limited: it doesn't contain enough of the right sorts of exemplars.

In sum: Skinner was mistaken on all counts. Language mastery is not merely a matter of having a set of bare behavioral dispositions. Instead, it involves intricate and detailed knowledge of the properties of one's language. And language learning is not a matter of being trained what to say. Instead, children learn language just from hearing it spoken around them, and they learn it effortlessly, rapidly, and without much in the way of overt instruction.

These insights were to drive linguistic theorizing for the next fiftyyears, and it's worth emphasizing just how radical and exciting they were at the time. First, the idea that explaining language use involves attributingknowledge to speakers flouted the prevailing behaviorist view that talking about mental states was unscientific because mental states are unobservable. It also raised several pressing empirical question that linguists are still debating. For example, what is thecontent of speakers' knowledge of language?[3] What sorts of facts about language are represented in speakers' heads? And how does this knowledge actually function in the psychological processes of language production and comprehension: what are themechanisms of language use?

Secondly, the idea that children learn language essentially on their own was a radical challenge to the prevailing behaviorist idea that all learning involves reinforcement. In addition, it made clear our need for a more ‘cognitive’ or ‘mentalistic’ conception of how language learning occurs, and vividly raised the question — our focus in this article — of what might be thepreconditions for that process. As we will see in the next section, Chomsky was ready with a theory addressing each of these points.

2. Arguments for the Innateness of Language

2.1 What do Children Learn when they Learn Language?

At the same time as the behaviorist program in psychology was waning under pressure from Chomsky and others, linguists were abandoning what is known as ‘American Structuralism’ in the theory of syntax. Like the behaviorists, the structuralists (e.g., Harris, 1951) refused to postulate irreducibly theoretical entities; they insisted that syntactic categories (such as ‘noun phrase’(‘NP’) or ‘verb phrase’ (‘VP’), etc.) be reducible to properties of actual utterances (collected in ‘corpora’ — lists of things people have said). In his landmark book,Syntactic Structures (1957), however, Chomsky argued that because corpora can contain only finitely many sentences, no attempt at reduction can succeed. Linguists need theoretical constructs that capture regularities going beyond the setof actual utterances, and that allow them to predict the properties of novel utterances. But if the category NP, for instance, is to include noun phrases that haven't been uttered yet, the meaning ofnoun phrase can't be exhausted by what's in the corpus: the structuralists' positivistic strictures on theoretical kinds are misguided.

In addition, the structuralists had attempted to capture the syntactic properties of languages in terms of simple rewrite rules known as ‘phrase structure rules.’ Phrase structure rulesdescribe the internal syntactic structures of sentence types; interpreted as rewrite rules, they can be used to generate or construct sentences. Thus, the rule S → NP VP, for instance, says that a sentence symbol S can be rewritten as the symbol NP followed by the symbol VP, and tells you that a sentence consists of a noun phrase followed by a verb phrase. (This information can be represented via a tree-diagram, as in Fig. 1a, or by a phrasemarker (or labeled bracketing), as in Fig. 1b.)

S-NP-VP Tree
(a)(b)

Figure 1. Phrasemarkers representing a sentence as consisting of a noun phrase and a verb phrase via (a) a tree diagram or (b) a labeledbracketing.

Other rules, (such as NP → Det N, VP → V NP, Det →a, the, …, etc., V→hit, kiss…, etc.; N →boy, girl,…, etc.) are subsequently applied, and (with still further rules not discussed here) allow for the generation of sentences such asThe boy kissed the girl, The girl hits the boy, and so on.

Chomsky argued (on technical grounds; see Chomsky 1957, ch.1) that grammars must be enriched with a second type of rule, known as ‘transformations.’ Unlike phrase structure rules, transformations operate on whole sentences (or more strictly, their phrasemarkers); they allow for the generation of new sentences (/phrasemarkers) out of old ones. ThePassive transformationdescribed in Chomsky 1957:112, for instance, specifies how to turn anactive sentence (/phrasemarker) into a passive one. Simplifying somewhat, you take an active phrasemarker of the formNP — AuxV — NP, likeKate is biting Mark, and rearrange its elementsx1x2x3x4 as follows:x4x2 +be +enx3 +byx1 to getMark bite (+is +en)by Kate. The parenthetical+ en and+ is invoke further operations on the verbbite that transform it intois being bitten, and ultimatelyKate is biting Markis ‘transformed’ intoMark is being bitten by Kate.

Only a grammar containing both phrase structure and transformation rules, Chomsky argued, could generate a natural language — ‘generate’ in the sense that by stepwise application of the rules, one could in principle build up from scratch all and only the sentences that the language contains. Hence, Chomsky urged the development of generative grammars of this type.

Syntactic theory has now gone well beyond this early vision — both phrase structure and transformation rules were abandoned in successive linguistic revolutions wrought by Chomsky and his studentsand colleagues (see Newmeyer 1986, 1997 for a history of generative linguistics).

But what has not changed — and what is important for our purposes — is that in every version of the grammar of (say) English, the rules governing the syntactic structure of sentences andphrases are stated in terms of syntactic categories that are highly abstracted from the properties of utterances that are accessible to experience. As an example of this, consider the notion of atrace. Traces are symbols that appear in phrasemarkers and mark the path of an element as it is moved from one position to another at various stages of a sentence's derivation, as in (1), where ti markes the NPJacob's position at an earlier stage in the derivation.

  1. Jacobi seems [ti to have vanished]

But while traces are vital to the statement of many syntactic rules and regularities, they are ‘empty categories’ — they are not audible in the sentence as spoken. (See Chomsky 1981 and Lasnik and Uriagereka 1986 for more on traces and other empty categories.) Traces (and other similarly abstract properties of languages) thus raise a question for the theory of language acquisition. For if, as Chomsky maintains, mastery of language involves knowledge of rules stated in terms of sentences' syntactic properties, and if those properties are not so to speak ‘present’ in the data, but are rather highly abstract and‘unobservable,’ then it becomes hard to see how children could possibly acquire knowledge of the rules concerning them. As a consequence, children's feat in learning a language appears miraculous: how could a child learn the myriad rules governing linguistic expression given only her exposure to the sentences spokenaround her?[4]

In response to this question, most 20th century theorists followed Chomsky in holding that language acquisition could not occur unless much of the knowledge eventually attained were innate or inborn. The gap between what speaker-hearers know about language (its grammar, among other things) and the data they have access to during learning (thepld) is just too broad to be bridged by any process of learning alone. It follows that since children patently do learn language, they are not linguistic ‘blank slates.’ Instead, Chomsky and his followers maintained, human children are born knowing the ‘Universal Grammar’ or ‘UG,’a theory describing the most fundamental properties of all natural languages (e.g., the facts that elements leave traces behind when they move, and that their movements are constrained in various ways).Learning a particular language thus becomes the comparatively simple matter of elaborating upon this antecedently possessed knowledge, andhence appears a much more tractable task for young children to attempt.

Over the years, two conceptions of the innate contribution to language learning and its elaboration during the learning process have been proposed. In earlier writings (e.g., Chomsky 1965), Chomskysaw learning a language as basically a matter of formulating and testing hypotheses about its grammar — unconsciously, of course. He argued that in order to acquire the correct grammar, the child must innately know a “a linguistic theory that specifies the form of the grammar of a possible human language” (1965:25)— she must know UG in other words. He saw this knowledge as being embodied in a suite of innate linguistic abilities, concepts, and constraints on the kinds of grammatical rules learners can propose for testing. On this view (1965:30-31), the inborn UG includes (i) a way of analyzing and representing the incoming linguistic data; (ii) a set of linguistic concepts with which to state grammatical hypotheses; (iii) a way of telling how the data bear on those hypotheses (an ‘evaluation metric’); and (iv) a very restrictive set of constraints on the hypotheses that areavailable for consideration. (i) through (iv) constitute the ‘initial state’ of the language faculty, and the child arrives at the final state (knowledge of her language) by performing what is basically a kind of scientific inquiry into its nature.

By the 1980's, a less intellectualized conception of how language is acquired began to supplant the hypothesis-testing model. Whereas the early model saw the child as a ‘little scientist,’ actively (if unconsciously) figuring out the rules of grammar, the new ‘parameter-setting’ model conceived language acquisition as a kind of growth or maturation; language acquisition is something that happens to you, not something you do. The innate UGwas no longer viewed as a set of tools for inference; rather, it was conceived as a highly articulated set of representationsof actual grammatical principles. Of course, since not everyone ends up speaking the same language, these innate representations must allow for some variation. This is achieved in this modelviathe notion of a ‘parameter’: some of the innately represented grammatical principles contain variables that may take one of a certain (highly restricted) range of values. These different‘parameter settings’ are determined by the child's linguistic experience, and result in the acquisition of different languages. Thus, Chomsky (1988:61-62) compared the learner to a switchbox: just as a switchbox's circuitry is all in place but for some switches that need to be flicked to one position or another, thelearner's knowledge of language is basically all in place, but for some linguistic ‘switches’ that are set by linguistic experience.

To illustrate how parameter setting works, consider a simplified example (discussed in more detail in Chomsky 1990:644-45). All languages require that sentences have subjects, but whereas some languages (like English) require that the subject be overt in the utterance, other languages (like Spanish) allow you to leave the subject out of the sentence when it is written or spoken. Thus, a Spanish speaker who wanted to say that he speaks Spanish could sayHablo español (leaving out the first personal pronounyo) without violating the rules of Spanish, whereas an English speaker wanting to express that thought could not say *Speak Spanish without violating the rules of English: to speak grammatically, he must sayI speak Spanish. The parameter-setting model accommodates this sort of difference by proposing that there is a ‘Null Subject Parameter,’ whichis set differently in English and Spanish speakers: Spanish speakers set it to ‘Subject Optional,’ whereas in English speakers, it is set to ‘Subject Obligatory.’ How? One proposal is that the parameter is set by default to ‘Subject Obligatory’ and that hearing a subjectless sentence causes it to be set to ‘Subject Optional.’ Since children learning Spanish frequently hear subjectless sentences, whereas those learningEnglish do not, the parameter setting is switched in the Spanish learner, but remains set at the default for the English learner. (Roeper and Williams 1987 is thelocus classicus for parameter-setting models; Ayoun 2003 is more up-to-date; Pinker, 1997: ch.3 provides a helpful, non-technical overview.)

These two approaches to language acquisition clearly differ significantly in their conception of the nature of the learning process and the learner's role in it, but we are not concerned to evaluate their respective merits here. Rather, the important point for our purposes is that they both attribute substantial amounts of innate informationabout language to the language learner. In what follows, we will look in more detail at the various argumentsthat have been used to support this ‘nativist’ theory of language acquisition. We will focus on the following question:

What evidence is there that children come to the language learning taskequipped with a specialized store of inborn linguistic information,such as that specified in the linguist's theory ofUniversal Grammar?

Terminological Note: As Chomsky acknowledges (e.g., 1986:28-29), ‘Universal Grammar’ is used with a systematic ambiguity in his writings. Sometimes, the term refers to the inborn knowledge of language that learners are hypothesized to possess — the content of the ‘initial state’ of the language faculty —whatever that knowledge (/content) turns out to be. Other times, ‘Universal Grammar’ is used to refer to certainspecific proposals as to the content of our innate linguistic knowledge, such as the Government-Binding theorist's claimthat we have inborn knowledge of such things as the Principle of Structure Dependence, Binding theory, Theta theory, the Empty Category Principle, etc.

This ambiguity is important when one is evaluating Chomskyan claims that we have innate knowledge of UG. For on the first reading of ‘Universal Grammar’ distinguished above, that claim will be true so long asany form of nativism turns out to be trueof language learners (i.e., so long as they possessany inborn knowledge about language). On the second reading, however, it is possible that learners have innate knowledge of languagewithout that knowledge's being knowledge of UG (as currentlydescribed by linguists): learners might know things about language, yetnot know Binding Theory, or the Principle of Structure Dependence, etc.

In this entry, ‘Universal Grammar’ will always be used inthe second of these senses, to refer to a specific theory as to the content of learners' innate knowledge of language. Where the issue concerns merely their havingsome or other innate knowledge about language (and is neutral on the question of whether any particular theory about that knowledge is true), I will talk of ‘innate linguistic information.’ Clearly, an argument to the effect that speakers have inborn knowledge of UG entails the claim that they have innate linguistic information at their disposal.The reverse, however, is not the case: there might be reason to thinkthat a speaker knows something about language innately, without its constituting reason to think that what they know is Universal Grammaras described by Chomksyan linguists; Chomksy might be right that we have innate knowledge about language, but wrong about what thecontent of that knowledge is. These issues will be clarified, as necessary, below.

2.2 Chomsky's ‘Poverty of the Stimulus’ Argument for the Innateness of Language

As we saw in §1.1, one of the conclusions Chomsky drew from his (1959) critique of the Skinnerian program was that language cannot belearned by mere association of ideas (such as occurs in conditioning). Since language mastery involves knowledge of grammar, and since grammatical rules are defined over properties of utterancesthat are not accessible to experience, language learning must be morelike theory-building in science. Children appear to be ‘little linguists,’ making highly theoretical hypotheses about the grammar of their language and testing them against the data provided by what others say (and do):

It seems plain that language acquisition is based on the child'sdiscovery of what from a formal point of view is a deep and abstracttheory — a generative grammar of his language — many ofthe concepts and principles of that are only remotely related toexperience by long and intricate chains of quasi-inferentialsteps. (Chomsky 1965:58)

However, argued Chomsky, just as conditioning was too weak a learningstrategy to account for children's ability to acquire language, so too is the kind of inductive inference or hypothesis-testing that goes on in science. Successful scientific theory-building requires huge amounts ofdata, both to suggest plausible-seeming hypotheses and to weed out any false ones. But the data children haveaccess to during their years of language learning (the ‘primarylinguistic data’ or ‘pld’) are highly impoverished, in two important ways:

  1. they constitute a small finite sample of the infinitely manysentences natural languages contain
  2. they do not reliably contain the kinds of sentences thatlearners need to falsify incorrect hypotheses

The first type of inadequacy is, of course, endemic to any kind of empirical inquiry: it is simply the problem of the underdeterminationof theories by their evidence. Cowie has argued elsewhere that underdetermination per se cannot be taken to be evidence for nativism: if it were, we would have to be nativists abouteverything that people learn (Cowie 1994; 1999). What of thesecond kind of impoverishment? If the evidence about language available to children does not enable them to reject false hypotheses, and if they nonetheless hit on the correct grammar, then language learning could not be a kind of scientific inquiry, which depends in part on being able to find evidence to weed out incorrect theories. And indeed, this is what Chomsky argues: since thepld are not sufficiently rich or varied to enable a learner to arrive at the correct hypothesis about the grammar of the languageshe is learning, language could not be learned from thepld.

For consider: The fact (i) that thepld are finite whereas natural languages are infinite shows that children must be generalizing beyond the data when they are learning their language's grammar: they must be proposing rules that cover as-yet unheard utterances. This, however, opens up room for error. In order to recover from particular sorts of error, children would need access toparticular kinds of data. If those data don't exist, as (ii) asserts,then children would not be able to correct their mistakes. Thus, since children do eventually converge on the correct grammar for their language, they mustn't be making those sorts of errors in the first place: something must be stopping them from making generalizations that they cannot correct on the basis of thepld.

Chomsky (e.g., 1965: 30-31) expresses this last point in terms of theneed forconstraints — on grammatical concepts, on thehypothesis space, on the interpretation of data — and proposes that it is innate knowledge of UG that supplies the needed limitations. On this view, children learning language are not open-minded or naïve theory generators — they are not ‘little scientists.’ Instead, the human language-learningmechanism (the ‘language acquisition device’ or ‘LAD’) embodies built-in knowledge about human languages,knowledge that prevents learners from entertaining most possible grammatical theories. As Chomsky puts it:

A consideration of…the degenerate quality andnarrowly limited extent of the available data …leave[s] little hope that much of the structure of the language can belearned by an organism initially uninformed as to its generalcharacter. (1965:58)

Chomsky rarely states the argument from the poverty of the stimulus in its general form, as Cowie has done here. Instead, he typically presents itvia an example. One of these concerns learning how to form ‘polar interrogatives,’ i.e., questions demandingyes orno by way of answer,via a mechanism known as ‘auxiliary fronting.’[5] Suppose that a child heard pairs of sentences like the following:

1a.Jacob is happy today
1b.Is Jacob happy today?
2a.The girls are dancing
2b.Are the girls dancing?

She wants to figure out the rule you use to turn declaratives like (1a) and (2a) into interrogatives like (1b) and (2b). Here are two possibilities:

H1.Find the first occurrence ofis in thesentence and move it to the front.
H2.Find the first occurrence ofis followingthe subject nounphrase (‘NP’) of the sentence, and move itto the front.

Both hypotheses are adequate to account for the data the learner has so far encountered. To any unbiased scientist, though, H1 would surely appear preferable to H2, for it issimpler — itis shorter, for one thing, and does not refer to theoretical properties, like being a NP, being instead formulated in terms of ‘observable’ properties like word order. Nonetheless, H1 is false, as is evident when you look at examples like (3):

3a.[The girl who is in the jumping castle]NP isKayley's daughter
3b.*Is [the girl who in the jumping castle]NP is Kayley'sdaughter?
3c.Is [the girl who is in the jumping castle]NP Kayley'sdaughter?

H1 generates the ungrammatical question (3b), whereas H2 generates the correct version, (3c).[6] Now, you and I and every other English speaker know (in some sense — see §3.2.1a) that H1 is false and H2 is correct.That we know this is evident, Chomsky argues, from the fact that we all know that (3b) is not the right way to say (3c). The question is how we could have learnt this.

Suppose, for example, that based on her experience of (1) and (2), a child were to adopt H1. How would she discover her error? There wouldseem to be two ways to do this. First, she could use H1in her own speech, utter a sentence like (3b), and be corrected by her parents or caregivers; second, she could hear a sentence like (3c) uttered bya competent speaker, and realize that that sentence is not generated by her hypothesis, H1. But typically parents don't correct their children's ill formed utterances (see §2.2.1(c) for more on this), andworse, according to Chomsky, sentences like (3c) — sentences that are not generated by the incorrect rule H1 and hence would falsify it— do not occur often enough in thepld to guarantee that every native English speaker will be able to get it right.

So in answer to the question: how do we learn that H2 isbetter than H1, Chomsky argued that we don'tlearnthis at all! A better explanation of how we all know thatH2 is right and H1 is wrong is that wewereborn knowing this fact. Or, more accurately, we wereborn knowing a certain principle of UG (the ‘Principle ofStructure Dependence’), which tells us that rules likeH1 are not worth pursuing, their ostensible‘simplicity’ notwithstanding, and that we should alwaysprefer rules, like H2, which are stated in terms ofsentences'structural properties. In sum, we know thatH2 is a better rule than H1, but we didn't learnthis from our experience of the language. Rather, this fact is aconsequence of our inborn knowledge of UG.

Chomskyans contest that there are many other cases in which speaker-hearers know grammatical rules, the critical evidence in favor of which is missing from thepld. Kimball 1973:73-5, for instance, argues that complex auxiliary sequences likemight have been are “vanishingly rare” in thepld, hence that children acquire competence with these constructions (in the sense of knowing the order in which to put the modal, perfect and progressive elements) without relevant experience.(Pullum and Scholz 2002, discuss two other well known examples.) Nativists thus conclude that numerous other principles of UG are innately known as well. Together, these UG principles place strong constraints on learners' linguistic theorizing, preventing them from making errors for which there are no falsifying data.

So endemic is the impoverishment of thepld, according to Chomskyans, that it began to seem as if the entire learning paradigm were inapplicable to language. As more and more and stricter and stricter innate constraints needed to be imposed on the learner's hypothesis space to account for their learning rules in the absence of relevant data, notions like hypothesis generation and testing seemed to have less and less purchase. This situation fuelled the recent shift away from hypothesis testing models of language acquisition and towards parameter setting models discussed in §2.1 above.

2.2.1 Criticisms of the Poverty of the Stimulus Argument

Many, probably most theorists in modern linguistics and cognitive science have accepted Chomsky's poverty of the stimulus argument for the innateness of UG. As a result, a commitment to linguistic nativism has underpinned most research into language acquisition overthe last 40-odd years. Nonetheless, it is important to understand what criticisms have been leveled against the argument, which I schematize as follows for convenience:

The General Form of the Argument from the Poverty of the Stimulus

  1. Mastery of a language consists (in part) of knowing itsgrammar.
  2. In order to learn a certain rule of grammar,G children would haveto have access to certain sorts of data,D, which falsify competinghypotheses.
  3. The primary linguistic data(pld) do not containD.

So

  1. G could not be learned.
  2. This situation is quite general: many rules of grammar areunlearnable from thepld.

So

  1. UG is innately known.
2.2.1(a) Premiss 1: Knowledge of grammar

In the 1970's, philosophers contested Chomsky's use of the word ‘know’ to describe speakers' relations to grammar, arguing that unlike standard cases of propositional knowledge, most speakers are utterly unaware of grammatical rules (e.g., “Anaphors are bound, and pronominals and R-expressions are freein their binding domains”) and many probably wouldn't understand them even if told what they are (Stich 1971). In response, Chomsky (e.g., 1980:92) began to use a technical term, ‘cognize,’ to describe the speaker-grammar relation, avoiding the philosophically loaded term, ‘knowledge.’

However, while it is certainly legitimate to propose a special relationship between speakers and grammars, unanswered questions remain about the precise nature of cognizance. Is it a representational relation, like belief? If not, what does ‘learning a grammar’ amount to? If so, are speakers' representations of grammar ‘explicit’ or ‘implicit’ or ‘tacit’ — and what, exactly, do any of these terms mean? (See the papers collected in MacDonald 1995, for discussion of this last issue; see Devitt 2006 for arguments that there is no good reason to suppose that speakers use any representations of grammatical rules in their production and comprehension of language.) Relatedly, how does a speaker's cognizance of grammar (her ‘competence,’ in Chomskyan parlance) function in her linguistic ‘performance’ — i.e., in the actual production or comprehension of an utterance?

These issues bear on the argument from the poverty of the stimulus because that argument may appear more or less impressive depending onthe answers one gives to them. If, for instance, one held that grammars are belief-like entities, explicitly represented in our heads in some internal code (cf. Stich 1978), then the question of how those beliefs are acquired and justified is indeed a pressing one— as, for different reasons, is the question of how they function in performance (see Harman 1967, 1969). However, if one were to deny that grammar is represented at all in the heads of speakers, like Devitt 2006 and Soames 1984, then the issue of how language is learned and what role ‘evidence’ etc. might play in that process takes on a very different cast. Or if, to take athird possibility, one were to reject generative syntax altogether and adopt a different conception of what the content of speakers' grammatical knowledge is — along the lines of Tomasello (2003), say — then that again affects how one views the learning process. In other words, one's ideas aboutwhat is learned affect one's conception ofwhat is needed to learn it. Less ‘demanding’ conceptions of the outputs of language acquisition require less demanding conceptions of its input (whether experiential or inborn); this last approach to the problem of language learning is discussed further in §2.2.1 below.

2.2.1(b) Premiss 2: The learning algorithm

In the example of polar interrogatives, discussed above, we saw how children apparently require explicit falsifying evidence in order to rule out the plausible-seeming but false hypothesis, H1. Premiss 2 ofthe argument generalizes this claim: there are many instances in which learners need specific kinds of falsifying data to correct their mistakes (data that the argument goes on to assert are unavailable). These claims about the data learners would need in order to learn grammar are underpinned by certain assumptions about the learning algorithm they employ. For example, the idea that false hypotheses are rejected only when they are explicitly falsified in the data suggests that learners are incapable of taking any kind of probabilistic or holistic approach to confirmation and disconfirmation. Likewise, the idea that learners unequipped with inborn knowledge of UG are very likely indeed to entertain false hypotheses suggests that their method of generating hypotheses is insensitive to background information or past experience. (e.g., information about what sorts of generalizations have worked in other contexts

The non-nativist language learner as envisaged by Chomsky in the original version of the poverty of the stimulus argument, in other words, is limited to a kind of Popperian methodology — one thatinvolves the enumeration of all possible grammatical hypotheses, eachof which is tested against the data, and each of is rejected just in case it is explicitly falsified. As much work in philosophy of science over the last half century has indicated, though, nothing much of anything can be learned by this method: the world quite generally fails to supply falsifying evidence. Instead, hypothesis generation must be inductively based, and (dis)confirmation is a holistic matter.

Thus arise two problems for the Chomskyan argument. First, it is not all that surprising to discover that if language learners employed a method of conjecture and refutation, then language could not be learned from the data. In other words, the poverty of the stimulus argument doesn't tell us much we didn't know already. Secondly, and as a result, the argument is quite weak: it makes the negative point that language acquisition does not occur via a Popperian learning strategy, but it favors no specific alternative to this acquisition theory. In particular, the argument gives no more support to a nativist (UG-based) theory than to one that proposed (say) that learners formulate grammatical hypotheses based on their extraction of statistical information about thepld and that they may reject them for reasons other than outright falsification — because they lack explicit confirmation, or because they do not cohere with other parts of the grammar, for instance.

In reply, some Chomskyans (e.g., Matthews 2001) challenge non-nativists to produce these alternative theories and submit them to empirical test. It's pointless, they claim, for nativists to try to argue against theories that are mere gleams in the empiricist's eye, particularly when Chomsky's approach has been so fruitful and thus may be supported by a powerful inference to the best explanation. Others have argued explicitly against particular non-nativist theories — Marcus 1998, 2001, for instance, discusses the shortcomings of connectionist accounts of language acquisition.

A recent book by Michael Tomasello (Tomasello 2003) addresses the nativist's demand for an alternative theory directly. Tomasello argues that language learners acquire knowledge of syntax by using inductive, analogical and statistical learning methods, and by examining a broader range of data for the purposes of confirmation and disconfirmation. He argues that children formulate abstract syntactic generalizations rather late in the learning process (aroundthe age of 4 or 5) and that their earliest utterances are governed bymuch less general rules of thumb, or ‘constructions.’ More abstract constructions, framed in increasingly adult-like and ‘syntactic’ terms, are progressively formulated through the application of pattern-recognition skills (‘analogy’)and a kind of statistical analysis of both incoming data and previously acquired constructions, which Tomasello calls ‘functional distributional analysis.’[7]

Tomasello's theory differs from a Chomskyan approach in three important respects. First, and taking up a point mentioned in the previous section, it employs a different conception of linguistic competence, the end state of the learning process. Rather than thinking of competent speakers as representing the rules of grammar in the maximally abstract, simple and elegant format devised by generative linguists, Tomasello conceives of them as employing rules at a variety of different levels of abstraction, and, importantly, asemploying rules that are not formulated in purely syntactic terms. Headopts a different type of grammar, called ‘cognitive-functional grammar’ or ‘usage-based grammar,’ in which rules are stated partly in terms of syntactic categories, but also in semantic terms, that is, in terms of their patterns of use and communicative function. A second respectin which Tomasello's approach differs from that of most theorists in the Chomskyan tradition, is in employing a much richer conception of the ‘primary linguistic data,’ orpld. For generative linguists, thepld comprises a set of sentences, perhaps subject to some preliminary syntactic analysis, and the childlearning grammar is thought of as embodying a function which maps that set of sentences onto the generative grammar for her language. On Tomasello's conception, thepld includes not just a set of sentences, but also facts about how sentences are used by speakersto fulfill their communicative intentions. On his view, semantic and contextual information is also used by children for the purposes of acquiring grammatical knowledge.

Tomasello argues that by adopting a more ‘user-friendly’ conception of natural language grammars and by radically expanding one's conception of the language-relevant information available to children learning language, the ‘gap’ exploited by the argument from the poverty of the stimulus — that is, the gap between what we know about language and the data we learn it from — in large part disappears. This gives rise to a third important respect in which Tomasello's theory differs from that of the linguistic nativist. On his view, children learn language withoutthe aid of any inborn linguistic information: what children bring to the language learning task — their innate endowment — is not language-specific. Instead, it consists of ‘mind reading,’ together with perceptual and cognitive skills that are employed in other domains as well as language learning. These skills include: (i) the ability to share attention with others; (ii) the ability to discern others' intentions (including their communicative intentions); (iii) the perceptual ability to segment the speech stream into identifiable units at different levels of abstraction; and (iv) general reasoning skills, such as the ability to recognize patterns of various sorts in the world, the ability to make analogies between patterns that are similar in certain respects,and the ability to perform certain sorts of statistical analysis of these patterns. Thus, Tomasello's theory contrasts strongly with the nativist approach.

Although assessing Tomasello's theory of language acquisition is beyond the scope of this entry, this much can be said: the oft-repeated charge that empiricists have failed to provide comprehensive, testable alternatives to Chomskyanism is no longer sustainable, and if the what and how of language acquisition are along the lines that Tomasello describes, then the motivation for linguistic nativism largely disappears.

2.2.1(c) Premiss 3: What do thepld contain?

A third problem with the poverty of the stimulus argument is that there has been little systematic attempt to provide empirical evidence supporting its assertions about what thepld contain. This is an old complaint (cf. Sampson 1989) which has recently been renewed with some vigor by Pullum and Scholz 2002, Scholz and Pullum 2002, and Sampson 2002. Pullum and Scholz provideevidence that, contrary to what Chomsky asserts in his discussion of polar interrogatives, children can expect to encounter plenty of datathat would alert them to the falsity of H1. Sampson 2002, mines the‘British National Corpus/demographic,’ a 100 million wordcorpus of everyday British speech (available online at http://info.ox.ac.uk/bnc/), for evidence that contrary to Kimball's contention that complex auxiliaries are ‘vanishingly rare,’ they in fact occur quite frequently (somewhere from onceevery 10,000 words to once every 70,000 words, or once every couple of days to once a week).

Chomskyans respond in two main ways to findings like this. First, they argue, it is not enough to show thatsome children can be expected to hear sentences likeIs the girl in the jumping castle Kayley's daughter?All children learn the correct rule, so the claim must be thatall children are guaranteed to hear sentences of this form — and this claim is still implausible, data like those just discussed notwithstanding.[8] In order to take this question further, it would be necessary to determine when in fact children master the relevant structures, and vanishingly little work has been done on this topic. Sampson 2002:82ff. found no well-formed auxiliary fronted questions (likeIs the girl who is in the jumping castle Kayley's daughter?)in his sample of the British National Corpus. He notes that in addition to supporting Chomsky's claims about the poverty of thepld, such data simultaneously problematize his claims about children's knowledge of the auxiliary-fronting rule itself. Sampson found that speakers invariably made errors when apparently attemptingto produce complex auxiliary-fronted questions, and often emended their utterance to a tag form instead (e.g., The girl who's in the jumping castle is Kayley's daughter, isn't she?). Hespeculates that the construction is not idiomatic even in adult language, and that speakers learn to form and decode such questions much later in life, after encountering them in written English. If that were the case, then the lack of complex auxiliary fronted questions in thepld would be both unsurprising and unproblematic: young children don't hear the sentences, but nor do they learn the rule. To my knowledge, children's competence with the auxiliary fronting rule has not been addressed empirically.[9]

Secondly, Chomskyans may produce other versions of the poverty of thestimulus argument. For instance, Crain 1991 constructs a poverty of the stimulus argument concerning children's acquisition of knowledge of certain constraints on movement. However, while Crain's argument carefully documents children's conformity to the relevant grammaticalrules, its nativist conclusion still relies on unsubstantiated intuitions as to the non-occurrence of relevant forms or evidence in thepld. It is thus inconclusive. (Cf. Crain 1991; Crain's experiments and their implications are discussed in Cowie 1999 ; Cf. also Crain and Pietrowski 2001, 2002).

2.2.1(d) The validity of the argument

The argument from (1), (2), and (3) to (4) appears valid. However, asis implicit in my discussion of premiss (2), an equivocation between different senses of ‘learning’ threatens. What (1)-(3) show, if true, is that grammar G can't be learned from thepld by a learnerusing a ‘Popperian’ learningstrategy, that is, a strategy of ‘bold conjecture’and refutation. What (4) concludes, however, is that G is unlearnable,period, from thepld — a move that several authors,particularly connectionists, have objected to. (See especiallyElmanet al. 1996 and Elman 1998 for criticisms of Chomskyannativism along these lines; see Marcus 1998 and 2001 forresponses.)

Chomskyans typically take this point, conceding that the argument from the poverty of the stimulus is not apodeictic. Nonetheless, theyclaim, it's a very good argument, and the burden of proof belongs with their critics. After all, nativists have shown the falsity of the only non-nativist acquisition theories that are well-enough worked out to be empirically testable, namely, Skinnerian behaviorismand Popperian conjecture and refutation. In addition, they have proposed an alternative theory, Chomskyan nativism, which is more than adequate to account for the phenomena. In empirical science, this is all that they can reasonably be required to do. The fact thatthere might be otherpossible acquisition algorithms whichmight account for children's ability to learn language is neither here nor there; nativists are not required to argue against mere possibilities.

In response, some non-nativists have argued that UG-based theories are not in fact good theories of language acquisition. Tomasello (2003: 182ff.), for instance, identifies two major areas of difficulty for UG-based theories, such as the principles-and-parameters approach. First, there is the ‘linking’ problem, deriving from the fact of linguistic diversity: almost no UG-based accounts explain how children link the highly abstract categories of UG to their instantiations in the particular language they happen to be learning.[10] His example is the category ‘Head,’ In order to set the ‘Head parameter,’ a child needs to be able to identify which words in the stream of noise she is hearing are in fact clausalheads. But heads “do not come with identifying tags on them in particular languages; they share no perceptual features in common across languages, and so their means of identification cannot be specified in [UG]” (Tomasello 2003:183). Second, there is the problem of developmental change, also emphasized by Sokolov and Snow,1991. It is difficult to see how UG-based approaches can account for the fact that children's linguistic performance seems to emerge piecemeal over time, rather than emerging in adult-like form all at once, as the parameter-setting model suggests it should.[11] In response, generativists have appealed to such notions as ‘maturational factors’ or ‘performance factors.’ But, Tomasello argues, such measures are ad hoc in the absence of a detailed specification of what these maturational orperformance factors are, and how they give rise to children's actual performance.

At the very least, such objections serve to equalize the burden of proof: non-nativists certainly have work to do, but so too do nativists. Merely positing an innate UG and a ‘triggering’ mechanism by which it ‘grows’ into full-fledged language is insufficient. Nativists need to show how their theory can account for the known course of language acquisition. Merely pointing out that there is apossibilitythat such theories are true, and that they would, if true, explain how language learning occurs in the face of an allegedly impoverishedstimulus, is only part of the job.

2.2.1(e) Premiss 5: How general is the poverty of the stimulus?

Because they are defending the view that all of UG is inborn, Chomskyans must be credited with holding that the primary data are impoverished quite generally. That is, if the innateness of UGtout court is to be supported by poverty of the stimulus considerations, the idea must be that the cases that nativists discuss in detail (polar interrogatives, complex auxiliaries, etc.) are but the tip of the unlearnable iceberg. Nativists quite reasonably do not attempt to defend this claim by endless enumerationof cases. Rather, they turn to another kind of argument to support the ‘global impoverishment’ position. This argument is sometimes called the ‘Logical Problem of Language Acquisition’; here, we will call it ‘The Unlearning Problem.’ It will be discussed in section 3.

2.2.1(f) The validity of the argument (II): What is inborn?

Suppose that the primary linguistic data were impoverished in all theways that nativists claim and suppose, too, that children know a bunch of things for which there is no evidence available — suppose, as Hornstein and Lightfoot (1981:9) put it, that “[p]eople attain knowledge of the structure of their language for whichno evidence is available in the data to which theyare exposed as children.” What follows from this is that there must be constraints on the learning mechanism: children do not enumerate all possible grammatical hypotheses and test them against the data. Some possible hypotheses must be ruled outa priori. But, critics allege, what doesnot follow from this isany particular view about the nature of the requisite constraints. (Cowie 1999: ch.8.) A fortiori, what does not follow from this is the view that Universal Grammar (construed as a theory about the structural properties common to all natural languages, per Terminological Note 2 above) is inborn.

For all the poverty of the stimulus argument shows, the constraints in question might indeed be language-specific and innate, but with contents quite different from those proposed in current theories of UG. Or, the constraints might be innate, but not language-specific. For instance, as Tomasello 2003 argues, children's early linguistic theorizing appears to be constrained by their inborn abilities to share attention with others and to discern others' communicative intentions. On his view, a child's early linguistic hypotheses are based on the assumption that the person talking to him is attempting to convey information about the thing(s) that they are both currentlyattending to. (Another example of an innate but non-language specificconstraint on language learning derives from the structure of the mammalian auditory system; ‘categorical perception,’ and is relation to the acquisition of phonological knowledge is discussedbelow, §3.3.4.). Another alternative is that the constraints might be learned, that is, derived from past experiences. An example again comes from Tomasello (2003). He argues thatentrenchment, or the frequency with which a linguistic element has been used with a certain communicative function, is an important constraint on the development of children's later syntacticknowledge. For instance, it has been shown experimentally that the more often a child hears an element used for a particular communicative purpose, the less likely she is to extend that element to new contexts. (See Tomasello 2003:179).

In short, there are many ways to constrain learners' hypotheses abouthow their language works. Since the poverty of the stimulus argument merely indicates the need for constraints, it does not speak to the question of what sorts of constraints those might be.

In response to this kind of point, Chomskyans point out that the innateness of UG is an empirical hypothesis supported by a perfectly respectable inference to the best explanation. Of course there is alogical space between the conclusion that something constrains the acquisition mechanism and the Chomskyan view that these constraints are inborn representations of Binding Theory, Thetatheory, the ECP, the principle of Greed or Shortest Path and so on. But the mere fact that the argument from the poverty of the stimulus doesn'tprove that UG is innately known is hardly reason to complain. This is science, after all, and demonstrative proofs are neither possible nor required. What the argument from the poverty of the stimulus provides is good reason to think that there are strong constraints on the learning mechanism. UG is at hand to supply a theory of those constraints. Moreover, that theory has been highly productive of research in numerous areas (linguistics, psycholinguistics, developmental psychology, second language research, speech pathology etc. etc.) over the last 50 years. These successes far outstrip anything that non-nativist learning theorists have able to achieve even in their wildest dreams, and support a powerful inference to the best explanation in the Chomskyan's favor.

2.2.1(g) Who has the burden of proof?

As seen above (§2.2.1(d)), however, the strength of the Chomskyan's ability to explain the phenomena of language acquisition has been questioned, and with it, implicitly, the strength of her inference to the best explanation. In addition, there is a general debate within the philosophy of science as to the soundness of inferences to the best explanation: does an explanation's being the best available give any additional reason (over and above its abilityto account for the phenomena within its domain) to suppose it true? [Link to Encyclopedia Article ‘Abduction’ by Peter Achinstein for more on this topic.]

In the linguistic case, what sometimes seems to underpin people's positions on such issues is differing intuitions as to who has the burden of proof in this debate. Empiricists or non-nativists contend that Chomskyans have not presented enough data (or considered enough alternative hypotheses) to establish their case. Chomskyans reply that they have done more than enough, and that the onus is on their critics either to produce data disconfirming their view or to producea testable alternative to it.

That such burden-shifting is endemic to discussions of linguistic nativism (the exchange in Ritter 2002 is illustrative) suggests to me that neither side in this debate has as yet fulfilled its obligations. Empiricists about language acquisition have ably identified a number of points of weakness in the Chomskyan case, but have only just begun to take on the demanding task of developing develop non-nativist learning theories, whether for language or anything much else. Nativists have rested content with hypotheses about language acquisition and innate knowledge that are based on plausible-seeming but largely unsubstantiated claims about what thepld contain, and about what children do and do not know and say.

It is unclear how to settle such arguments. While some may disagree(especially some Chomskyans), it seems that much work still needs tobe done to understand how children learn language — and not justin the sense of working out the details of which parameters get setwhen, but in the sense of reconceiving both what linguistic competenceconsists in, and how it is acquired. In psychology, a new,non-nativist paradigm for thinking about language and learning hasbegun to emerge over the last 10 or so years, thanks to the work ofresearchers like Elizabeth Bates, Jeffrey Elman, Patricia Kuhl,Michael Tomasello and others. The reader is referred to Elmanetal. 1996, Tomasello 2003 and §3 below for an entréeinto this way of thinking.

For now, considerations of space demand a return to our topic, viz., linguistic nativism, rather than further discussion of alternatives to it.

2.3 The Argument from the ‘Unlearning Problem’

We saw in the previous section that in order to support the view thatall of UG is innately known, nativists about language need to hold not just that the data for language learning is impoverished in a fewisolated instances, but that it's impoverished across the board. Thatis, in order to support the view that the innate contribution to language acquisition is something as rich and detailed as knowledge of Universal Grammar, nativists must hold that the inputs to languageacquisition are defective in many and widespread cases. (After all, if the inputs were degenerate only in a few isolated instances, such as those discussed above, the learning problem could be solved simplyby positing innate knowledge of a few relevant linguistic hints, rather than all of UG.)

Pullum and Scholz (2002:13) helpfully survey a number of ways in which nativists have made this point, including:

  1. Finiteness: thepld (primary linguistic data) are finite,whereas languages contain infinitely many sentences.
  2. Underdetermination: thepld are always compatible withinfinitely many grammatical hypotheses.
  3. Degeneracy: thepld contain ungrammatical and incompletesentences.
  4. Idiosyncrasy: different children learning the same language areexposed to different samples of sentences.
  5. Positivity: thepld contain only positive instances (whatis a sentence of the language to be learned, a.k.a. the‘target language’).
  6. No Feedback: children are not told or rewarded when they getthings right, and are not corrected when they make mistakes.

In this section, I will set aside features (i) and (ii) as being characteristic ofany empirical domain: the data arealways finite, and theyalways underdetermine one'stheory. No doubt it's an important problem for epistemologists and philosophers of science to explain how general theories can nonetheless be confirmed and believed. No doubt, too, it's an important problem for psychologists to explain the mechanisms by which individuals acquire general knowledge about the world on the basis of their experience. But underdetermination and the finiteness of the data are everyone's problem: if these features of the languagelearning situation per se supported nativism, then we should accept that all learning, in every domain, requires inborn domain-specific knowledge. But while it's not impossible that everything we know thatgoes beyond the data is a result of our having domain-specific innateknowledge, this view is so implausible as to warrant no further discussion here.

I also set aside features (iii) and (iv). For one thing, it is unclear exactly how degenerate thepld are; according to oneearly estimate, an impressive 99.7% of utterances of mothers to theirchildren are grammatically impeccable (Newport, Gleitman and Gleitman 1977). And even if the data are messier than this figure suggests, it is not unreasonable to suppose that the vast weight of grammatically well-formed utterances would easily swamp any residual noise. As to the idiosyncrasy of different children's data sets, thisis not so much a matter of stimulus poverty as stimulus difference. As such, idiosyncrasy becomes a problem for a non-nativist only on the assumption that different children's states of linguistic knowledge differ from one another less than one would expect given the differences in their experiences. As far as I know, no serious case for this last claim has ever been made.[12]

In this section, we will focus on features (v) and (vi) of thepld. For it is consideration of the positivity of the data set, and the lack of feedback available to children, that has given rise to what I am calling the ‘Unlearning Problem,’ otherwise known (somewhat misleadingly) as the ‘Logical Problemof Language Acquisition.’ (For statements of the argument, see,e.g., Baker 1979; Lasnik; 1989:89-90; Pinker 1989.)

Diagram of possible relationships between H and L

Figure 2. Five possible relations between the language generatedby hypothesis (H) and the target grammar (L)

Take a child learning the grammar of her language,L. Figure 2 represents the 5 possible relations that might obtain between the language generated by her current hypothesis,H, and that generated by the target grammar,L. (v) represents the end point of the learning process: thelearner has figured out the correct grammar for her language. A learner in situation (i), (ii) or (iii) is in good shape, for she caneasily use thepld as a basis for correcting her hypothesis as follows: whenever she encounters a sentence in the data (i.e., a sentence ofL) that is not generated byH, she has to ‘expand’ her hypothesis so that it generates that sentence. In this way, H will keep moving, as desired, towardsL. However, suppose that the learner finds herself in situation (iv), where her hypothesis generates all of the target language,L, and more besides. (Children frequently find themselves in this position, for example, they invariably go through a phase in which they overgeneralize regular past tense verb endings to irregular verbs; their grammars generate the incorrect*I breaked it as well as the correctI broke it.) There, she is in deep trouble, for she cannot use thepld to discover her error. Every sentence ofL, after all, is already a sentence ofH. In order to ‘shrink’ her hypothesis — to ‘unlearn’ the rules that generate*I breaked it — she needs to know which sentences ofH are not sentences ofL — she needs to figure out that*I breaked it is not a sentence of English. But — and this is the problem — this kind of evidence, often called ‘negative evidence,’ is held to beunavailable to language learners.

For as we have seen, thepld is mostly just a sample of sentences, of positive instances of the target language. It contains little, if any, information about strings of words that are not sentences. For instance, children aren't given lists of ungrammaticalstrings. Nor are they typically corrected when they make mistakes. And nor can they simply assume that strings that haven't made their way into the sample are ungrammatical: there are infinitely many sentences that are absent from the data for the simple reason that no-one's had occasion to say them yet.

In sum: a child who is in situation (iv) — a child whose grammar ‘overgenerates’ — would need negative evidence in order to recover from her error. Negative evidence, however, does not appear to exist. Since children do manage to learn languages, they must never get themselves into situation (iv): they must never need to ‘unlearn’ any grammatical rules. Thereare two ways they could do this. One would be never to generalize beyond the data at all. But clearly, children do generalize, else they'd never succeed in learning a language. The other would be if there were something that ensured thatwhen they generalize beyond the data, they don'tovergeneralize, something, that is, that ensures that children don't make errors that they could onlycorrect on the basis of negative evidence. According to the linguistic nativist, this something is innate knowledge of UG.

2.3.1 Criticisms of the ‘Unlearning’ Argument

2.3.1 (a) What is lacking? Negative data vs. Negative Evidence

First, let's make a distinction between:

Negative Data: explicit information that a given string ofwords is not a sentence of the target language. (E.g., “No,that's not how you say it,” or “It'sI broke itnotI breaked it,” or “That string of words isungrammatical,” etc.)

and

Negative Evidence: information that would enable a learner totell that a given hypothesis is (very likely to be) incorrect. (Seebelow for examples.)

Second, let's abandon the idea, which reappears in many presentationsof the Argument from the Unlearning Problem; that learners' hypotheses must be explicitly falsified in the data in order to be rejected. Let's suppose instead that learners proceed more like actual scientists do — provisionally abandoning theories due tolack of confirmation, making theoretical inferences to link data withtheories, employing statistical information, and making defeasible, probabilistic (rather than decisive, all-or-nothing) judgments as to the truth or falsity of their theories.[13]

Intuitively, viewing the learner as employing more stochastic and probabilistic inductive techniques enables one to see how the unlearning problem might have been overblown. What the argument claims, rightly, is thatnegative data near enough do not exist in thepld. However, what learners need in order to recover from overgeneralizations, is not negative dataper se, butnegative evidence, and arguably, thepld do contain significant amounts of that. For example:

  1. Failures of understanding or communication: others'failures to understand children's linguistic productions (evidencedeither by requests for repetition or by communicative failure) areevidence to the learner that there is something wrong with the rule(s)she was using to generate her utterance. This evidence is not decisive(maybe Granny just couldn't hear her properly), but it is negativeevidence nonetheless.
  2. Non-occurrence of structural types as negative evidence: Suppose that a child's grammar predicted that a certain string is part of the target language. Suppose further that that string never appears in the data, even when the context seems appropriate. Proponents of the unlearning problem say that non-occurrence cannot constitute negative evidence — maybe Dad simply always chooses to sayThe girl who is in the jumping castle is Kayley's daughter, isn't she? rather than the auxiliary-fronted version,Is the girl who is in the jumping castle Kayley's daughter? If so, it would be a mistake for the child to conclude on the basis of this information that the latter string is ungrammatical.

    But suppose that the child is predicting not strings of words,simpliciter, but rather strings of wordsunder a certainsyntactic description (or, perhaps more plausibly, quasi-syntactic description — the categories employed need not be the same as those employed in adult grammars).[14] This would enable her to make much better use of non-occurrence as negative evidence. For non-occurring strings will divide into two broad kinds: those whose structureshave been encountered before in the data, and those whose structures havenot beenheard before. In the former case, the child has positive evidence that strings of that kind are grammatical, evidence that would enableher to suppose that the non-occurrence of that particular string was just an accident. (E.g., she could reason that since she's heardIs that mess that is on the floor in there yours? many times, and since that string has the same basic structure asIs that girl that's in the jumping castle Kayley's daughter?, the latter string is probably OK even though Dad chose not to say it.)

    In the case in which the relevant form hasnever been encountered before in the data, however, the child is better off: thefact that she has never heard any utterance with the structure of *Is that girl who in the jumping castle is Kayley's daughteror*Is that mess that that on the floor in there is yours? is evidence that strings of that type are not sentences. Again, the evidence is not decisive, and the child should be prepared to revise her grammar should strings of that kind start appearing. Nonetheless,the non-occurrence of a string, suitably interpreted in the light of other linguistic information, can constitute negative evidence and provide learners with reason to reject overgeneral grammars.

  3. Positive Evidence as Negative Evidence. Relatedly, learners can also exploit positive evidence as to which strings occur in thepld as a source of negative evidence — again in a tentative and revisable way.[15] Suppose that the child's grammar generated two strings as appropriatein a given kind of context, but that only one sort of string was everproduced by those around her. The fact that only strings of the firstkind occur is in this case negative evidence — defeasible, to be sure, but negative evidence nonetheless.

    In fact, the use of positive evidence to disconfirm hypotheses is endemic to science. For instance, Millikan used positive evidence to disconfirm the theory that electrical charge is a quantity that varies continuously. In his famous ‘Oil Drop’ experiment,he found that the amount of charge possessed by a charged oil drop was always a whole-number multiple of —(1.6 x 10-19)C. The finding that all observed charges were ‘quantized’ in this manner disconfirmed the competing ‘continuous charges’ hypothesis in the same way that positive evidence can disconfirm grammatical hypotheses.[16]

  4. Feedback The Argument from the ‘UnlearningProblem’ also points to the lack of feedback provided tochildren learning language. In a famous study often cited byproponents of the argument, Brown and Hanlon 1970 (see also Brown,1973 and Brown, Cazden, and Bellugi 1969) found no overt disapprovalby mothers of the syntactic errors of their children, and moreoverfound that caregivers had no trouble understanding their charges'ill-formed utterances. Only semantic errors were occasionallycorrected; grammatical mistakes went unremarked.

However, more recent findings have uncovered evidence indicating thatfailures of understanding occur with some regularity, and that there is a wealth of feedback about correct usage in the language-learning environment. For example:

  • Hirsh-Pasek, Trieman and Schneiderman (1984) studied interactionsbetween 2 year olds and their parents, and discovered that caregiversrepeated and corrected 20.8% of flawed sentences, whereas they onlyrepeated (without correction) 12.0% of well-formed utterances.
  • Demetras, Post and Snow (1986) found that in general, onlywell-formed sentences were repeated verbatim by parents, and thatill-formed sentences were not repeated verbatim, but were ratherfollowed by clarification questions (“What?” —indicating a lack of understanding) or expansions and/or recasts,correcting the error.
  • Bohannon and Stanowicz (1988) found that 34% of syntactic and 35%of phonological errors received some form of differential feedback(e.g., repetitions with corrections or explicit rejection of theutterance); that more than 90% of parents' exact repetitionsfollow well-formed sentences; and that more than 70% of recasts andexpansions follow ill-formed utterances.
  • Chouinard and Clark's (2003) longitudinal study of fivechildren learning language found that parents reformulate erroneousutterances more often than correct utterances, that they respondequally often to all error types (phonological, lexical, syntactic,semantic), and that they correct younger children, who make moreerrors, more frequently.
  • Perhaps most tellingly, Moerk (1991) performed a reanalysis ofBrown's “Eve” transcripts (among those on which the1970 ‘no feedback claim was based) and found many instances inwhich Eve's semantic and syntactic errors were explicitlycorrected, including: her use of noun labels; VPs (tense, modality,auxiliaries); determiners and prepositions; word order (these lastsorts of error were rare, but were invariably corrected).
  • Bohannon, MacWhinney and Snow (1990) review other results in thisvein, as well as responding to nativist criticisms of these findingsand their bearing on the unlearning problem.
2.3.1 (b) Children Can and Do Learn from ‘Noisy’ Data and Exploit Statistical Regularities

Chomsky has recognized the existence of such ‘indirect’ negative data in thepld. However, he concluded that they were too few and ambiguous to be of aid to the language learner. The sorts of findings reported above seem to show that negative evidence is pervasive in thepld. But can children learn from these sorts of statistical regularities?

Standard formulations[17] of the ‘Unlearning Problem’, assume that they cannot: theview seems to be that learning can only take place under idealized conditions where the world supplies unambiguous evidence pro or con the language learner's grammatical theories. Given such a conception of the learner, none of the examples of feedback just discussed will seem relevant to the problem. For only a learner employing fairly sophisticated data-analysis techniques and a confirmation measure that is sensitive to small changes in probabilities would be able to exploit the sorts of regularities in the linguistic environment that we have just discussed. However, there is increasing evidence that children are in fact remarkably sensitive to subtle feedback, in bothlinguistic and non-linguistic domains. For instance:

  • Bohannon and Stanowicz (1988) found that children pay particularattention to parental corrections of their mistakes: they imitate25.6% of adult expansions (saying the same thing as the child, butgiving more detail) and recasts (repetitions of the child'sutterance, correcting errors), whereas they only imitate 3.6% of exactor verbatim repetitions by parents of the child's utterance.
  • Relatedly, Farrer (1990 1992) found that children were morelikely to repeat a given morpheme if it were part of an adult recastof one of the child's own sentences, than if it were part of a anon-repetitious adult utterance (e.g., a change of subject or acontinuation of the conversation). She also found thatchildren's repetition of adult utterances facilitated thechild's acquisition of various grammatical morphemes.
  • Morgan and Travis (1989) and Morganet al. (1995)dispute the long-term efficacy of such corrective feedback; Bohannonnet al. 1996 respond.
  • However, in both longitudinal studies of children in naturalenvironments (Chouinard and Clark 2003) and in experimental studies(Saxtonet al. 1998, and Saxton 1997, Saxton, Backley and Galloway,2003) the long-term efficacy of feedback has been demonstrated.

In addition, it is becoming increasingly clear that babies, children,adults and many other mammals are highly sensitive not just to feedback, but to other non-obvious statistical regularities in their experience. For example:

  • Saffran, Aslin and Newport (1996) found that 8 month old babieswere able to learn where the word boundaries in an artificial languageoccurred after a mere 2 minutes' exposure to a stream ofartificial speech. The stream consisted of 3-syllable nonsense words(bidaku,padoti,golabu) repeatedcontinuously for 2 minutes (bidakupadotigolabubidakugolabi…etc.). The stream was constructed so that the‘transitional probability’ of two sounds X#Y was equal to1 when the sounds formed part of a word, and equal to 1/3 when thesounds spanned a word boundary. In two minutes, the infants hadlearned to discriminate the ‘words’ (likebidaku)from the ‘non-words’ (e.g.,kupado). See alsoChambers, Onishi and Fisher 2003.
  • Other studies have expanded upon these results, indicating thatchildren and babies are sensitive to patterns in a wide range ofverbal cues, such as linguistic rhythm (Nazzi and Ramus 2003);prosodic stress (Thiessen and Saffran 2003) and voicing and syllabicstructure (Saffran and Thiessen 2003).
  • Moreover, there is increasingly persuasive evidence thatstatistical or ‘distributional’ information may be usednot just for the extraction of word boundaries, but — contraryto an old argument of Chomsky's — to limn higher levels ofsyntactic structure as well — see, e.g., Redington and Chater,1998; Penaet al. 2002; Mintz 2002; Saffran 2002; Saffran andWilson 2003; Newport and Aslin 2004. Chater and Manning 2006 providea survey.)
  • Finally, and forestalling any response along the lines that whatwe are seeing here is just the nativist's ‘Language AcquisitionDevice’ in action, a number of studies have shown that similarmechanisms appear to be at work in learning in non-linguistic domains(Saffran 2002, studied learning of non-linguistic sounds and shapes);in adults (Penaet al. 2002); and in other animals, such ascotton-top tamarin monkeys (Hauser 2001; Hauser, Weiss and Marcus,2002).

Taken together, these kinds of results raise the possibility that some of the foundational learning mechanisms involved in language acquisition are not language specific. If it turns out that babies employ the sorts of distributional analysis studied by Saffran, Redington and Chater, Pena, and Mintz not only in learning artificiallanguages, but also in learning natural languages, then that is evidence against linguistic nativism. For this type of learning is employed by humans and other animals in other contexts as well: whatever is involved in language learning — be it innate or not— is not language-specific.

2.3.1 (c) The Generality of the Argument

The previous objections to the Unlearning Problem Argument made the points, first, that negative evidence does exist in thepld (in the form of regularities both in others' language use and in how others react to children's own productions), and second, that children (and other animals) seem very good at exploiting this kind of information for the purposes of learning about their world. This would seem to be rather a good thing, given that there is reason to think that learnersmust be able to learn in domains where explicit negative data do not exist, and in the absence of specialized innate knowledge of those domains. For the unlearning problem is a problem for learning from experience quite generally. That is, there are many domains in which learners lack explicit evidence as to what things arenot: trees are not cars, Irish stews are not curries, birds are not fish and MacDonald's is not a branch of the CIA. No-one ever told you any of these things, but it's crazy to think that you now know them because you possess analogs to the ‘Language Acquisition Device’ for each of these domains. Clearly, in at least some areas, people are able to learn an awful lot on the basis of largely positive data, and while this of course does nothing to show that language is one of those areas, it does indicate that the Unlearning problem argument by itself is no argument for linguistic nativism at all, let alone for the Chomskyan (UG-based) version of that position.

3. Other Research Bearing on the Innateness of Language: New Problems for the Nativist?

In this section, I will mention some other avenues of research that have been argued to have a bearing on the innateness of language. My goal is not to give an exhaustive survey of these matters, but ratherto provide the interested reader with a way into the relevant literatures. Still, I will try to give enough details so as to make acase that current empirical findings, together with the flaws identified in §§1 and 2 in the positive arguments for linguistic nativism, tend to militate against that position.

3.1 Linguistic Universals

Chomsky and others (e.g., Chomsky 1988:46-7; Pinker 1994:237-8) have pointed to the existence of ‘linguistic universals’ as supporting the idea that language is the product of a distinct faculty of mind. Universals are features thought to be common to all natural languages, such as the existence of constraints on the movement of elements during a derivation or, less controversially, the existence of a syntactic distinction between nouns and verbs. Butnot only is the existence of true universals a contested matter (see e.g., Maratsos 1989:111), it is unclear what the correct explanationof them — assuming they exist — is would be.

One explanation is certainly the Chomskyan one that they are consequences of speakers' innate knowledge of UG. Another is that they derive from other, non-linguistically-specific features of cognition, such as memory or processing constraints (e.g., Berwick and Weinberg 1983 trace certain constraints on movement to limitations on parsing imposed by the structure of human memory). Yetanother is that they derive from universal demands of the communication situation (e.g., Sapir 1921, argued that the distinction between nouns and verbs arises from the fact that language is used to communicate propositions, hence needs a way to bring an object subject to mind and a way to say something about it).Finally, as Putnam 1971 speculated, universals might be relics of anancestral Ur-language from which all other languages evolved. This last hypothesis has generally been rejected as lacking in empirical support. However, recent findings in genetics and historical linguistics are converging to suggest that all human populations evolved from a small group migrating from Africa in the fairly recentpast, and that all human languages have probably evolved from the language spoken by that group. (Cavalli-Sforza 1997.)

The Ur-language hypothesis is not, of course, inconsistent with linguistic nativism. However, if true, it does weaken anyargument from the existence of universals to the innateness of linguistic knowledge. For if languages have a common ancestor, then it is possible to explain universals — even ones that seemstrange from a functional point of view — as being the result of our ancestors' having adopted a certain solution to a linguistic coordination problem. Like driving on the right side of the road, a solution once established may become entrenched, because the benefitsof everyone's conforming to the same rule outweigh the costs of changing to a different rule, and this may be so even if the new rulewere in some sense more ‘reasonable.’ Thus, even arbitrary or odd features of language can be explained historically, without positing either compelling functional considerations or inborn linguistic constraints.[18]

If, by contrast, language emerged independently in a number of areas,the existence of universals would be a strong argument for nativism, For in that case, it would be implausible to maintain that each ancestral group ‘just happened’ to select the same solutions to the various coordination problems they encountered. Moreplausible would be the supposition that the different groups' choice of the rule was driven by somethinginternal to speakers, such as, perhaps, an innate representation of UG. In short: if languages have a common ancestor, then common descent from originallyarbitrary linguistic conventions is a possible explanation of linguistic universals, including the ‘odd’ or ‘arbitrary’ ones that don't seem to have any real functional significance. If they don't, then such universals seemingly could only be explained in terms of features internal to speakers.

3.2 Language Localization

Broca's and Wernicke's area diagram

Figure 3. Broca's area and Wernicke's area

Beginning with the work of Broca and Wernicke in the 19th century, a popular view has been that language is localized to certain areas of the brain (see Fig. 3), almost always the left hemisphere,[19] and that it is subject to characteristic patterns of breakdown, called ‘aphasias.’ (See Saffran 2000 for a survey of thevarious aphasias.) For example, Broca's area is strongly implicated in speech production, and damage to this area can result in a characteristic inability (‘Broca's aphasia’ or ‘agrammatism’) to produce fluent speech, especially complex grammatical structures and grammatical morphemes. The fact that syntax can apparently be selectively interfered with by lesions to Broca's area has been taken by some to indicate that grammatical knowledge is localized to that area, and this in turn has been taken to show support the view that there is a special biological inborn basis for that knowledge. (Lenneberg 1964, 1967 is the original proponent of this argument, which is echoed in more recent discussions, such as Pinker 1994:297-314.)

It is unclear, however, why this inference should seem compelling.First, as Elmanet al. 1996 argue, neural localization offunction can occur as a result of virtually any developmentaltrajectory: the localization of some function bears not at all on itsinnateness.

Secondly, it is now known that neural localization for language is very much a relative, rather than an all-or-nothing matter (Dronkerset al. 2000, Dicket al. 2001, Martin 2003). Notonly is language processing widely distributed over the brain (seeFig. 4), but traditionally language-specific areas of cortex areimplicated in a variety of non-linguistic tasks as well. Broca's area,for instance, ‘lights up’ on MEG scans(magnetoencephalography, a method for measuring changes in themagnetic properties of the brain due to electrical activity) whensubjects hear a discordant musical sequence in much the same way as itdoes when they hear an ungrammatical utterance. (Maessetal. 2001; a special issue ofNature Neuroscience, 6(7),July 2003, explores the implications of this finding.)

Finally, recent studies of cortical plasticity have shown that eventhe most plausible candidates for innate specification — such asthe use of visual cortex for vision or the use of auditory cortex forhearing — exhibit high degrees of experience-dependentplasticity. For example, in congenitally blind subjects, the areas ofthe brain normally used for seeing are taken over for the processingof Braille (Sadatoet al. 1996; Hamilton and Pascual-Leone,1998) and even in those with late-onset blindness, significant‘rewiring’ of visual cortex for other perceptual tasks isapparent (Kujalaet al. 1997). Likewise, in the congenitally deaf,auditory cortex is used for the processing of sign language (Nishimuraet al. 1999, von Melchner, Pallas and Sur 2000). (See Shimojo andShams 2001, for a review.)

pet scan images

Figure 4. Pet scan showing brain regions involved in various languagetasks. From Posner and Raichle (1997, 15). Used by permission of M. Raichle.

As Marcus (2004:40-45) points out in response to Elmanetal. 1996, the ability of the brain to ‘rewire’ itselfunder exceptional circumstances is consistent with its having been‘prewired,’ or set up, differently by the genes. However,these sorts of data indicate that complex functions, such as areinvolved in processing sign language, can be carried out in areas ofbrain that are ‘prewired’ (if they are) to do somethingquite different. This suggests that these abilities require little inthe way of task-specific pre-wiring, and are learned largely on thebasis of experience (together with whatever sort of 'prewiring' issupplied for the cortex as a whole). That is, if sign languageprocessing tasks can be carried out by areas of cortex that arepresumably innately predisposed (if they are) to do auditoryprocessing, then the former competence must be being learned in theabsence of inborn constraints or knowledge that are specific to thattask. Of course, these are pathological cases, and it is unclearwhether the subjects in these experiments had any special training inorder that their brains were ‘rewired’ in theseways. Nonetheless, examples like these provide an existence proof ofthe brain's ability to acquire complex processing capacities —indeed, processing capacities relevant to language — in thecomplete absence of inborn, domain-specific information. As such, theyraise the possibility that other aspects of language processing aresimilarly acquired in the absence of task-specific constraints.

In sum, the neuroscientific evidence currently available provides no support for linguistic nativism. The suggestion that localization of function is indicative of a substantial degree of innate prespecification is no longer tenable: localization can arise in manydifferent ways. In addition, linguistic functions do not seem to be particularly localized: language use and understanding are complex tasks, involving many different brain areas — areas that are inat least some cases implicated also in other tasks. It is hard to seehow to reconcile these facts with the Chomskyan postulation of a monolithic ‘language organ,’ the development or ‘growth’ of which is controlled largely by the genes. Finally, the fact that complex functions can be learned and carried out by areas of brain that are innately ‘prewired’ (if atall) to do quite different sorts of processing indicates that such competences can be and are acquired without any inborn, task-specificguidance. This is not, of course, to say that language is one of the competences that are acquired in this way. For all the current evidence shows, many areas of cortex in which language develops may indeed be ‘prewired’ for that task: linguistic nativism is still consistent with what is now known. It is, however, to suggest that although there may be other reasons to be a linguistic nativist, general considerations to do with brain organization or development as currently understood give no especial support to that position.

3.3 The Critical Period for Language Acquisition

Lenneberg (1964, 1967) also argued that although language acquisitionis remarkably robust, in the sense that all normal (and many abnormal) children do it, it can occur unproblematically only during a ‘critical period’ — roughly, up to late childhoodor early puberty. On analogy with other supposedly innately specifiedprocesses like imprinting or visual development, Lenneberg used the existence of a critical period as further evidence that language possesses a proprietary basis in biology.

In support of the critical period hypothesis about language, Lenneberg cited the facts (i) that retarded (e.g., Downs syndrome) children's language development stops around puberty; (ii) that whereas very young children are able to (re)learn language after aphasias produced by massive left-hemisphere trauma (including hemispherectomy), aphasias in older children and adults are typicallynot reversible; and (iii) that so-called ‘wild children,’viz., those who grow up with no or little exposure to human language,exhibit severely compromised language skills. (Lenneberg, 1957:142-55; see Curtiss 1977 for the (in)famous case of Genie, a modern-day ‘wild child’ from suburban Los Angeles, who was unable to acquire any but the most rudimentary grammatical competence after a miserable and wordless childhood spent locked alone in a room, tied to her potty chair or bed.)

As further support for the critical period hypothesis, others have added the observation that although children are able to learn a second language rapidly and to native speaker fluency, adult learnersof second languages typically are not: the capacity to learn a secondlanguage tapers off after puberty, no matter how much exposure to thelanguage one has. (Newport 1990). Thus, it was speculated, the innate knowledge base for language learning (e.g., knowledge of UG) becomes unavailable for normal acquisition at puberty, and adult learners must rely on less efficient learning methods. (Johnson and Newport 1989.)

As a preliminary to discussing these arguments (many of which are presented in more detailed in Stromswold 2000) it is worth distinguishing two notions that often get conflated under the name ‘critical period’:

Critical Period: a time during development which isliterally critical; the relevant competence either cannot develop or will be permanently lost unless certain inputs are received during that period.

Sensitive Period: a time during development in which a competence is acquired ‘normally,’ or ‘easily,’ or ‘naturally.’ The competence can be acquired outside the sensitive period, but perhaps less easily andnaturally, and or perhaps with less ultimate success.

The classic example of a critical period is due to the Nobel prize-winning work of Hubel and Wiesel. By suturing shut one of a kitten's eyes at various stages of development and for various periods of time, Hubel and Wiesel (1970) showed that certain corticaland thalamic areas supporting binocular vision (specifically, ocular dominance columns[20] and cells in the lateral geniculate body) will not develop normally unless kittens receive patterned visual stimulation during the 4th to12th weeks of life. They found that while the damage was sometimes reversible to some extent, depending on the exact duration and timingof the occlusion, occlusion for the entire first three months of lifeproduced irreversible blindness in the deprived eye.[21]

Language, however, is not like this. As we will see, there is little evidence for acritical period for language acquisition, although there is considerable evidence that there is asensitiveperiod during which language is acquired more easily. The implications of this for claims about the innateness of language willbe addressed in §3.3.4.

3.3.1 Language recovery after trauma

Lenneberg cited the superior ability of children to (re)learn language after left brain injury in support of the critical period hypothesis. But while there clearly is a difference between the abilities of young children, on the one hand, and older children and adults, on the other, to recover from left brain insults, the contrast in recovery course and outcome is not as stark as is often supposed.

First, older children — even those who have not succeeded in learning language previously — can substantially recover from left hemisphere trauma occurring well after the supposed closure of the ‘sensitive’ or ‘critical’ period; in effect, they learn language from scratch as adolescents. Vargha-Khademet al. 1997, for instance, report the case of Alex, who failed to speak at all during childhood and whose receptive language was at age 3-4 level at age 9. After his left cortex was removed at age 9, Alex suddenly began to learn language with gusto, and by age 15, his skills were those of an 8-10 year old.

Secondly, most adults suffering infarcts in the left hemisphere language areas do in fact recover at least some degree of language competence and many recover substantially normal competence, especially with treatment (Hollandet al. 1996). This is thought to be due both to the regeneration of damaged speech areas and to compensatory development in other areas, particularly in the right hemisphere (Karbeet al. 1998). Similar processes seem to be at workin young children with left hemisphere damage. Mulleret al. 1999, for instance, document significant relearning of language, together with increased right-hemisphere involvement in language tasks, after left-hemisphere lesions in both children (<10) years) and adults (>20 years).

Finally, not even very young children are guaranteed to recover language after serious insults, whether to the left or right hemisphere. As Bates and Roe (2001) argue in their survey of the childhood aphasia literature, outcomes differ wildly from case to case, and the reported studies exhibit numerous methodological confounds (e.g., inability to localize the lesion or to know its cause, different measures of linguistic competence, different time frames for testing, statistical irregularities, and failure to control for other factors known to affect language such as seizure history) that cast doubt on the degree of empirical support possessedby Lenneberg's claim in this instance.

3.3.2 ‘Wild children’

It has long been recognized that interpretation of the ‘wildchild’ literature — helpfully surveyed in Skuse 1993— is confounded by the fortunate rarity of these ‘naturalexperiments,’ the generally poor reporting of them, and theother environmental factors (abuse, malnutrition, neglect, etc.) thatoften go along with extreme linguistic deprivation. However, in workpioneered by Goldin Meadow and colleagues (e.g., Goldin Meadow andMylander 1983, 1990), a new population of individuals, who arelinguistically but not otherwise deprived, has begun to bestudied. Deaf but otherwise normal children of hearing parents who areneither educated in sign language nor sent to special schools for thedeaf do not acquire language, although they usually develop their ownrudimentary signing systems, called ‘homesign,’ to usewith their families. Studies of what happens to such children afterthey are exposed to natural languages (signed or verbal) at variousages promise to offer new insights into the critical and sensitiveperiod hypotheses.

At this time, however, there are still very few case reports in the literature, and the data so far obtained in these studies are equivocal with respect to the sensitive and critical period hypotheses. Some adolescents do seem to be able to acquire language despite early linguistic deprivation, and others do not. It is unclear what the explanation of these different outcomes is, but one important factor appears to be whether the new language is a signed language (e.g., ASL) or a spoken language. Perhaps because their childhood perceptual deficits prevented normal auditory and articulatory development, deaf children whose hearing is restored later in life do not seem to be able to acquire much in the way of spoken language. (Grimshawet al. 1998.)

3.3.3 Second language acquisition in children and adults

The issue of second language acquisition (“SLA”) has beenargued to bear on the innateness of language by supporting a critical(or sensitive) period hypothesis. For instance, Johnson and Newport (1989) found that among immigrants arriving in the U.S. before puberty, English performance as adults was better the earlier in lifethey arrived, but that there were no effects of arrival age on language performance for those arriving after puberty. The fact that the amount of exposure to the second language mattered for speakers if it occurredbefore puberty but notafter, was taken to confirm the critical period hypothesis.

However, these results have failed to be replicated (Birdsong andMolis 2001) and while it still has its supporters, the ‘criticalperiod’ hypothesis regarding second language acquisition isincreasingly being criticized (Hakuta, Bialystok and Wiley 2003;Nikolov and Djugunovich 2006). Newer studies have argued, forinstance, that the degree of proficiency in a second languagecorrelates better with, such factors as the learner's level ofeducational attainment in that language, her length of residence inthe new country,) and the grammatical similarities between the firstand second languages, and/or length of residence in the newcountry. (Flege, Yeni-Komshian and Liu 1999; Bialystok,1997)[22]

The fact that many adults and older children can learn both first andsecond languages to a high degree of proficiency makes clear that unlike the kitten visual system studied by Hubel and Wiesel, the language acquisition system in humans is not subject to a critical period in the strict sense. This finding is consistent with the emerging view that the cortex remains highly plastic throughout life,and that contrary to received wisdom, even old dogs can be quite goodat learning new tricks. (See Buonomano and Merzenich 1998; Cowen andGavazzi 1998; Quartz and Sejnowski 1997; and Stiles 2000.) It is also consistent with the idea, which seems more plausible than the critical period hypothesis, that there is asensitive period for language acquisition — a time, from roughly birth age 1 to age 6 or 7, in which language is acquiredmost easily and naturally, and when a native-like outcome is virtually guaranteed. (Cf. Mayberry and Eichen 1991.) The implications of this conclusion for linguistic nativism are examined in the next section.

3.3.4 Sensitive periods and innateness: phonological learning

What does the existence of a sensitive period for language masterytell us about the innateness of language? In this section, we willlook at a case, namely phonological learning, in which the existenceof a sensitive period has received much press, and in which theinference from sensitivity to the existence of language-specificinnate information has been made explicitly (see Eimas 1975). One canargue that even in this case, the inference to linguistic nativism isweak.

Much rarer than mastery of second language morphology and syntax is attainment of a native-like accent, something that first language learners acquire automatically in childhood.[23] A child's ability to perceive language-specific sounds begins in utero, as demonstrated, for instance, by newborns' preference for thesounds of their mother's voice and their parents' language, and by their ability to discriminate prose passages that they have heard during the final trimester from novel passages. In the first few months of life, babies reliably discriminate many different natural language phonemes, whether or not they occur in what is soon to become their language. By ages 6 months to 1 year, however, this sensitivity to unheard phonemes largely disappears, and by age 1, children tend to make only the phonological distinctions made in the language(s) they hear around them. For example, Japanese children lose the ability to discriminate English /r/ and /l/ (Kuhlet al., 1997b). As adults, people continue to be unable to perceive some phonetic contrasts not marked by their language, and many fail to learn how to produce even those second language sounds which they can distinguish.[24] For instance, many English speakers of French have great difficulty in producing the French /y/ (as intu) and back-of-the-throat /r/.

Thus, in the case of phonological learning, there does seem to be an inborn predisposition to segment vocal sounds into language-relevant units, or phonemes.[25] However, there is also evidence that learning plays a role in shapingphonological knowledge — and not just by ‘pruning away’ unwanted ‘phonological representations,’ as Eimas (1975) hypothesized, but also by shaping the precise boundariesof adult phonemic categories. For example, caregivers reliably speak a special ‘language’ (“Motherese” or “Parentese”) to young babies, raising pitch, shortening sentences, emphasizing stressed morphemes and word boundaries and — most relevant here — exaggerating the acoustical differences between certain crucial vowels (in English, /i/, /a/ and /u/) . This ‘stretching’ of the distance between vowels (demonstrated in Finnish and Russian as well as English by Kuhl et al. 1997a) facilitates the infant's representation of clearly distinguishable vowel prototypes. Kuhl 2000 argues that these prototypes subsequently function as ‘magnets’ around which subsequent linguistic experiences are organized, and form the set points of the language-specific phonological ‘map’ that emerges by the end of the first year.

If this is indeed how phonological learning works, it is clear that while experience clearly plays a role, the inborn contribution to that process is quite substantial. For discriminating phonemes — however those discriminations might be shaped by subsequent experience — is no simple matter. It involves what is called ‘categorical perception, that is, the segmenting of a signal that varies continuously along a number of physical dimensions (e.g.,voice onset time and formant frequency) into discrete categories, so that signals within the category are counted as the same, even thoughacoustically, they may differ from one another more than do two signals in different categories (see Fig. 5). (Harnad 1987 is a useful collection of work on categorical perception to the mid-1980s.)

But is this inborn contribution to phonological learninglanguagespecific, that is, does it support the conclusion that (this aspect of) language is innate? And to this question, the answer appears to be ‘No.’ First, the ‘chunking’ of continuously varying stimuli into discrete categories is a feature not just of speech perception, but of human perception generally. Forinstance, it has been demonstrated in the perception of non-linguistic sounds, like musical pitch, key and melody, and meaningless chirps and bleats (Pastore and Layer 1990). It has also been demonstrated in the processing of visual stimuli like faces (Beale and Keil 1995), facial expressions (Etcoff and Magee 1992; Kotsoni, de Haan and Johnson 2001); facial gender (Campanella, Chrysochoos and Bruyer 2001); and familiar physical objects (Newell and Bulthoff 2002). Secondly, it is known that other animals too perceive categorically. For instance, crickets segment consepecific songs in terms of frequency (Wyttenbach, May and Hoy 1996), swamp sparrows ‘chunk’ notes of differing durations (Nelson andMarler 1989), and rhesus monkeys can recognize melodies when transposed by one or two octaves, but not by 1.5 or 2.5 octaves, indicating a grasp of musical key (Wrightet al. 2000). Finally, other species respond categorically to human speech! Chinchillas (Kuhl and Miller 1975) and cotton-top tamarins (Ramuset al. 2000) make similar phonological distinctions to those made by human infants.

Together, as Kuhl 1994, 2000 argues, these findings cast doubt on the language-specificity of the inborn perceptual and categorization capacities that form the basis of human phonological learning. For given the fact that human (and animal) perception quite generally is categorical, it is arguable that languages have evolved so as to exploit the perceptual distinctions that humans are able to make, rather than humans' having evolved the abilities to make just the distinctions that are made in human languages, as a view like Eimas' would suggest.

Categorical Perception Graph

Figure 5. Note that the pair of sounds circled in blue differ in F2 starting frequency less than those circled in red, yet the former areboth reliably counted as instances of the sounds /b/ whereas the latter are reliably classified as different sounds, /d/ and /g/. Thispattern, together with the abrupt switch from one classification to another (e.g. /b/ to /g/), is characteristic of categorical perception.

The same may be true in non-phonological domains too. The notion thatat least some of the capacities responsible for syntactic learning are non-language specific is suggested by analogous results about thenon-species specificity of recursive rule learning and generalization— an ability that Chomsky has recently suggested forms the coreof the human language faculty. (Hauser, Chomsky and Fitch 2002; see below, 3.4 for further discussion.) Other species, notably cotton toptamarins, seem capable of learning simple recursive rules (Hauser, Weiss, and Marcus 2002). In addition, Hauser and McDermott 2003 argue that musical and syntactic processing involve similar competences, which are again seen in other species. Together, these findings suggest that there are aspects of the human ‘language faculty’ that are neither task-specific nor species-specific. Instead, language learning and linguistic processing make use of abilities that predate language phylogenetically, and that are used in humans and in animals for other sorts of tasks. (See e.g., Hauser,Weiss, and Marcus 2002 for an account of recent work on rule learning by cotton top tamarins; see Hauser and McDermott 2003 for the suggestion that aspects of musical and syntactic processing involve similar competences, which are again seen in other species.) Rather than viewing the human mind as being innately specialized for language language learning, it seems at least as reasonable to think of languages as being specialized so as to be learnable and usable bythe human mind; of this, more in §3.4 below.

3.4 Language Evolution

This brings us to the question of language evolution: if knowledge oflanguage (say, of the principles of UG) really is inborn in the humanlanguage faculty, how did such inborn knowledge evolve? For many years, Chomsky himself refused to speculate about this matter, stating that “[e]volutionary theory…has little to say, as of now, about questions of this nature” (1988:167). Other theorists have not been so reticent, and a large literature has grownup in which the selective advantages of having a language are adumbrated. It's good for communicating with, for instance, when trying to figure out what conspecifics are up to (Pinker and Bloom, 1990; Dunbar 1996). It's a mechanism of group cohesion, analogous toprimate grooming (Dunbar 1996). It's a non-genetic mechanism of phenotypical plasticity, allowing organisms to adapt to their environment in non-evolutionary time (Brandon and Hornstein 1986; Sterelny 2003). It's a mechanism by which we can bend others to our will (Dawkins and Krebs 1979; Catania 1990), or make social contracts (Skyrms 1996). Language makes us smarter, perhaps by beinginternalized and functioning as a ‘language of thought’ (Bickerton 1995, 2000). And so on.

The ability to speak and understand a language no doubt provided and continues to provide us with many of these benefits. Consequently (and assuming that the costs were not too great — as patently they weren't), one can be sure that whatever it is about human beingsthat enables them to learn and use language would have been subjectedto strong positive selection pressure once it began to emerge in our species.

But none of this speaks directly to the issue of linguistic nativism.The fact that Mother Nature would have favored individuals or groups possessing linguistic abilities tells us nothing about themeans she chose to get the linguistic phenotype built. That is, it tells us nothing about the sorts of psychological mechanisms that were recruited to enable human beings to learn, and subsequentlyuse, a natural language.

Nativism is, of course, one possibility. Natural selection might havebuilt a specialized language faculty, containing inborn knowledge about language (e.g., knowledge of UG), which subsequently was selected for because it helped human children to acquire linguistic competence, and having linguistic competence enhanced our ancestors' fitness. A problem with this hypothesis, however, is that it is unclear how a language faculty containing innate representations of UG might have arisen in the human mind. One view is that the languagefaculty was built up piecemeal by natural selection. This approach underlies Pinker and Bloom's (1990) and Jackendoff's (1999) proposalsas to the adaptive functions of various grammatical features and devices. Other nativists, however, reject the adaptationist framework. For instance, Berwick 1998, has argued that efforts to explain the piecemeal development of knowledge of linguistic universals in our species may be unnecessary in light of the new, Minimalist conception of syntax (see Chomsky 1995). On this view, all parametric constraints and rules of syntax are consequences of a fundamental syntactic process called Merge: once Merge was in place, Berwick argues, the rest of UG automatically followed. Chomsky, taking another tack, has suggested that language is a ‘spandrel,’ a byproduct of other non-linguistically directed selective processes, such as “the increase in brain size and complexity” (1982:23). And finally Bickerton 1998, on yet another tack, posits a massive saltative episode in which large chunks of syntax emerged all at once, although this posit is implicitly withdrawn in Calvin and Bickerton 2000.

The literature on language evolution is too large to survey in this article (but see Botha 2003 for an excellent overview and critique).Suffice it to note that as yet, no consensus has emerged as to how innate knowledge of UG might have evolved from whatever preadaptations existed in our ancestors. Of course, this is not in itself a problem for linguistic nativists: formulating and testing hypotheses about human cognitive evolution is a massively difficult enterprise, due largely to the difficulty of finding evidence bearing on one's hypothesis. (See Lewontin 1998 and Sterelny2003:95-116.)

It's worth noting, however, that linguistic nativism is just one possibility for how Nature got language up and running. Just as it may be that a language faculty embodying knowledge of UG was somehow encoded in the human genome, it's also possible that that our abilityto learn a language is based on a congeries of pre-existing competences, none of which is (or was initially — see below) specialized for language learning. Tomasello's theory of language acquisition, discussed above (§2.2.1.b), invites this alternative evolutionary perspective. On his view, the fundamental skills with which linguistic competence is acquired are skills that originally served, and still continue to serve, quite different, non-linguistic functions. For example, he argues that children's early word and phrase learning rest in part on their ability to shareattention with others, to discern others' communicative intentions, and to imitate aspects of their behavior. There is reason to think that these abilities evolved independently of language, at least initially: imitation learning enabled the fast and high-fidelity transfer of learned skills between generations (see Tomasello 1999, 2000) and the ability to form beliefs about the mental states of others (‘mind-reading’ or ‘theory of mind’) enabled highly intelligent animals, such as our hominid ancestors, tonegotiate a complex social environment made up of similarly intelligent conspecifics. (See, e.g., Sterelny 2003.) On this sort of view, the ability to learn language piggy-backed on other capacities, which originally evolved for other reasons and which continue to serve other functions in addition to their linguistic ones.

You might wonder, however, whether this latter kind of account reallydiffers substantively from that of a nativist. Assuming that she doesnot reject adaptationism altogether, the nativist will presumably be committed to the idea that the innate language organ, or faculty embodying knowledge of UG, was derived from pre-existing structures that were either functionless or had non-linguistic functions. These structures subsequently acquired linguistic functions through being selected for that reason: they became adaptations for language. But so too would the various capacities postulated by Tomasello. As soon as they started being used for language learning, that's to say, theywould have been selected for that function (in addition to any other functions they might serve, and always assuming that linguistic abilities were on balance beneficial). Hence they too will over time become adaptations for language. On both Tomasello's and the nativist's view, in other words, the inborn structures responsible for language acquisition will have acquired the biological function of enabling language acquisition: they will be specialized for that purpose. Is Tomasello, then, a nativist?

No. First, even though the psychological abilities and mechanisms that Tomasello posits have been selected for linguistic functions, these abilities and mechanisms have continued to be used (and, plausibly, selected) for non-linguistic purposes, such as face recognition, theory of mind, non-linguistic perception, etc. So, whereas a central tenet of linguistic nativism is its insistence thatthe structures responsible for language learning aretask-specific, Tomasello sees those structures as being muchmore general-purpose. In addition, and this is a second reason not tocount Tomasello as a nativist, the inborn structures he posits are not plausibly interpreted as containing any kind of language specificinformation or representations. Yet a commitment to the role of inborn, language-specific information (such as knowledge of UG) is another hallmark of linguistic nativism.

Several theorists (e.g., Clark 1996, Tomasello 1999, and Sterelny, 2003) have stressed that in addition to working on human linguistic abilities directly, via changes to the parts of the genome coding forthose abilities, natural selection can also bring about such changes indirectly, by making sure that our minds are embedded in certain kinds of environments. All sorts of animals create environments for themselves: this is called ‘niche construction.’ (The term is due to Odling-Smee, Laland and Feldman 1996.) Many animals also (or thereby) create environments for their offspring as well. And as Odling-Smeeet al. 1996, Avital and Jablonka 2000, and Sterelny 2003 stress, animals' dispositions to modify the environments of both themselves and their offspring in certain ways are just as much potential objects of selection as are other of theirtraits.

To see this, suppose that an organism O has a genetically encoded disposition N to build a special kind of nest; suppose further that being raised in this kind of nest causes O-type offspring to have characteristic C; and suppose finally, that Os with C enjoy greater reproductive success than those without. Then, assuming that there isvariation in N in the population, natural selection can operate so asto increase the proportion of Os with N — and hence also those with characteristic C — in the population. Down the track, Os will have C not by virtue of acquiring a special, genetically-encodeddisposition-for-C. Rather, they will have C because their parents have the genetically-encoded disposition N, and Os whose parents haveN ‘automatically’ develop C.

This toy example illustrates a further route by which language might have evolved in human beings. In addition to creating inborn language-learning mechanisms in individuals, natural selection may also have created dispositions to construct particular kinds of linguistic learning environmentsin theirparents. For example, as Clark (1996) and Sterelny (2003) both speculate, Mother Nature might have worked on our dispositions to use ‘Motherese’ to our children, and/or on our tendency to talk about things that are current objects of the child's perceptual attention, in order to create learning environments conducive to the acquisition of language.

In principle, the existence of this sort of ‘nicheconstruction’ can be accepted by all parties to the nativismcontroversy. That is, both Tomasello and Chomsky could agree thatdispositions to construct ‘linguistic niches’ —environments in which languages are easy for human offspring to learn— may have been selected for in our species. Nevertheless, thenotion of niche construction militates against the nativist,particularly when one takes into account the related notion of‘cumulative downstream niche construction.’

Cases of what Sterelny (2003: 149ff) calls ‘cumulative downstream niche construction’ occur when a generation of animals modifies an environment that has already been modified by earlier generations. A mountain thornbill's nest is an instance of downstream niche construction (since its offspring are affected by the thornbill's efforts). However, the construction is not cumulative, since the nest is built anew each year. By contrast, a rabbit warren extended and elaborated over several generations is an instance of cumulative construction: successive generations of offspring inherit an ever-more-complex niche and their other behaviors are tuned accordingly in ever-more-complex ways. Tomasello,1999 and Sterelny 2003 stress that niche construction, including downstream niche construction, is not limited to the physical world: animals make changes to their social and epistemic worlds as well. For instance, chimpanzees live in groups (= construction of a social niche) and dogs mark their territory (= a change in their epistemic niche, relieving them of the necessity of remembering where the boundaries of their territory are). Humans, says Sterelny, echoing a theme of Tomasello 1999, are niche constructors “with a vengeance” (2003:149) and many of the changes they make to their physical, social and epistemic environments accumulate over many generations (think of a city, a democracy, modern science, a natural language). Such cumulative modifications allow for what Tomasello calls a “ratchet effect”: a “cycle in which an improvement is made, becomes standard for the group, and then becomes a basis for further innovation.” (Sterelny 2003: 150-1)

The idea of cumulative niche construction has obvious application to the case of language. If parents shape the linguistic environment of their offspring, and if we all shape the linguistic environments of our conspecifics (merely by talking to them!) then the possibility ofa ‘linguistic ratchet effect’ is clearly open. Small changes made to the language of the group by one generation — changes which perhaps make it easier to learn, or easier to understand or produce — will be transmitted to later generations, who may in turn make further changes geared to increasing language learnability and ease of use. This scenario raises the possibility, already mentioned at the end of the last section, that language may have evolved so as to be learnable and usable by us, in addition to the converse scenario (stressed in much work on the evolution of language) thatwe had to change in many and complex ways in order to learn and use a language. Thus, we might speculate, languages' phonetic systems evolved so as to be congenial to our animal ears; their expressive resources (in particular, their vocabularies) evolved so as to fit our communicative needs; and perhaps, as Clark 1997 has suggested and as Tomasello 2003 implicitly takes for granted, natural language syntaxevolved so as to suit our pre-existing cognitive and processing capacities. To be sure, the languages we have coded in our heads lookcomplex and weird to linguists and psychologists and philosophers whoare trying to put together theories about them. But, if languages andhuman minds have evolved in tandem, as surely they have, then languages may not look weird at all from the point of view of the brains that implement and use them.

All of these processes have likely played a role in the evolution ofour capacities to learn and use a natural language. Pre-existingpsychological, perceptual and motor capacities would have beenrecruited for the task of language learning and use. These capacitieswould have been honed and specialized further by natural selection forthe performance of linguistic tasks. The functions of some of them,perhaps, would have become so specialized for language-related tasksthat they cease to perform any non-linguistic functions at all —and to this extent, perhaps, linguistic nativism would bevindicated. At the same time, however, language itself would have beenevolving so as the better to suit our cognitive and perceptualcapacities, and our communicative needs. Given the fact that manydifferent perceptual, motor and cognitive systems are implicated inlanguage use and learning, and given the co-evolution of our minds andour languages, the truth about language evolution, when it emerges, isunlikely to be a simple. For this reason, it is unlikely tovindicate the nativist's notion that a specialized and monolithic‘language organ’ or ‘faculty’ is at the rootof our linguistic capacities.

Before leaving the question of language evolution, it is necessary tomention a recent paper by Hauser, Chomsky and Fitch 2002 on this topic. First, they distinguish (2002:1571) what they call the ‘faculty of language in the narrow sense,’ or ‘FLN,’ from the ‘faculty of language in the broad sense,’ or ‘FLB.’ The FLN is the “abstract linguistic computational system alone…which generates internalrepresentations and maps them into the sensory-motor interface by thephonological system, and into the conceptual-intentional interface bythe (formal) semantic system.” (Ibid.) The FLB includes the FLN plus all the other systems (motor systems, conceptual systems, perceptual systems, and learning skills) which contribute to language acquisition and use.

Next, Hauseret al. speculate that the only thing that's really special about the human FLB is the FLN. That is, with the exception only of the FLN, FLB comprises systems that are shared with (or only slight modifications of) systems in other animals. Consequently, there is no mystery (or no more mystery than usual) about how these language-related abilities evolved. FLN, on the other hand, is distinctive to humans and what is special about it is its power ofrecursion, that is, its ability to categorize linguistic objects into hierarchically organized classes, and (on the behavioralside) for the generation of infinitely many sentences out of finitelymany words. According to Hauseret al., the only real evolutionary mystery about language is how this capacity for recursion evolved — and this question, argue Hauseret al, is eminently addressable by normal biological methods (e.g., comparative studies to determine possible precursor mechanisms, etc.).

However, there are two difficulties with this scenario. First, there is evidence that the power of recursion posited by Hauseret al. as being distinctive of the human FLN is in fact not distinctive to humans, because it is not species specific. (See Esser,et al. 1997and McGonigle, Chalmers and Dickinson, 2003.) Second, recursiveness is not language specific either, but is a feature of other domains of human cognition and endeavor as well. Our conceptualspace, for instance, appears to be hierarchically ordered (poodles are a kind of dog, which are a kind of quadruped, which are a kind ofanimal, etc.). Similarly, the planning and execution of non-linguistic actions seems often to involve the sequencing and combining of smaller behavioral units into larger wholes. Recursion might well be an important part of the human language faculty, but it's apparently not specific either to us or to that faculty. Or, to put the point more bluntly: if it's Chomsky's view that recursivenessis the pivotal feature of the language faculty, and if recursiveness is a feature of human cognition and action more generally, then it's not clear that Chomsky remains a linguistic nativist.[26]

3.5 Pidgins and Creoles

It has been argued (by, e.g., Bickerton 1981, and Pinker, 1994:32-9) that the process by which a pidgin turns into a creole provides direct evidence of the operation of an innate language faculty. Pidgins are rudimentary communication systems that are developed when people speaking different languages come together (often in a commercial setting or when one people has conquered and is exploiting another) and need to communicate about practical matters. Creoles arise when pidgins are elaborated both syntacticallyand semantically, and take on the characteristics ofbona fide natural languages.

Bickerton and, following him, Pinker, argue that creolization occurs when children take a pidgin as the input to their first language learning, and urge that the added complexity of the creole reflects the operation of the child's inborn language faculty. Moreover, they argue, since creole languages all tend to be elaborated in the same ways, and since they all respect the constraints of UG, the phenomenon of creolization also supports the idea that the inborn contribution to language acquisition is not just some general drive for an effective system of communication, but rather knowledge of linguistic universals.

There are two problems with this ‘language bioprogram hypothesis,’ as it is known in the creolization literature. Thefirst concerns the claim (e.g., Bickerton 1981:43-70) that even creoles that developed in quite different areas of the world, and in complete isolation from one another, bear “uncanny resemblances” (Pinker 1994:25) to each other, not just in respecting the constraints of UG, but — even more surprisingly — in using fundamentally the same means to elaborate their rootpidgins (e.g., in using the same syntactic devices to mark tense, aspect and modality). The stronger claim made by Bickerton — that Creoles use the same devices for the same grammatical purposes — is simply not true. For example, as Myhill (1991) argues, Jamaican Creole, Louisiana Creole, Mauritian Creole and Guyanese Creole mark tense, aspect and modality in ways that are quite different from those that Bickerton (1981) proposed as universal. (See, however, Mufwene 1999 for a case that confirms Bickerton's predictions.) The weaker claim — that creoles respect the constraints imposed by UG — has not, so far as I know, been contested. So we will assume, in what follows, that creoles, like other NLs, respect UG. The important question for our purposes is: how does this come about?

The bioprogram hypothesis claims that creolization occurs as a resultof the action of the language faculty: children who learn language from degraded (e.g., pidgin) inputs are compelled by their innate knowledge of grammar to produce a fully-fledged natural language (thecreole) as output. As an example of how children add UG-constrained structure to languages learned from degraded inputs, Pinker cites thecase of Simon, a deaf child studied by Newport and her colleagues, who learned American Sign Language (ASL) from parents who themselves were not exposed to ASL until their late teens. Although they used ASL as their primary language, Simon's parents were “in many ways…like pidgin speakers,” says Pinker (1994:38). For instance, they used inflectional markers in an inconsistent way and often failed to respect the structure-dependence of the rules governing topicalization in that language.[27] But “astoundingly,” says Pinker, “though Simon saw no ASL but his parents' defective version, his own signing was far better ASL then theirs…Simon must somehow have shut out his parents' ungrammatical ‘noise.’ He must have latched on to the inflections that his parents used inconsistently, and interpreted them as mandatory.” (1994:39) Pinker views this as a case of “creolization by a single living child” (ibid.) and explains Simon's conformity to ASL grammar in terms of the operation of his innate language faculty during the acquisition period.

In a recent overview of the Simon data from the last 10 or so years, however, Newport 2001, stresses a number of facts that Pinker's presentation obscures or downplays. First, Simon's performance was not that of a native signer, although he did develop “his own version of ASL whose structure was more like that of other natural languages [than that of his parents' ASL]” (Newport 2001:168).For instance, Simon's morphology stabilized at a level that was “not as complex as native ASL” and he didn't acquire standard classifier morphemes if they were not used by his parents (ibid.). Secondly, Simon's success in learning a given rule seemed to vary with how well or badly his parents signed. For instance, Simon's parents used the correct inflectional morphology 60-75% of the time for a large class of verbs of motion, and in this case, Simon's own use of such morphology was 90% correct. However, some members of the class of classifier morphemes were correctly usedby the parents only 40% of the time, and in this case, although Simon's performance was better than his parents', it was not at native signer level.

Newport argues that Simon appears to be ‘cleaning up’ hisparents' language, that is, “bas[ing] his learning heavily on his input, but reorganiz[ing] this input to form a cleaner, more rule-governed system than the one to which he was exposed.” (2001:168) She agrees that this result could be due to constraints imposed by an innate language faculty, but argues that it is also consistent with the existence of some more generalized propensity in children to generate systematic rules from noisy inputs, rightly pointing out that the latter hypothesis cannot be ruled out in advance of empirical test. (In this context, she notes (p.170) some preliminary studies suggesting that inferring systematic rules from messy data may indeed be a more general feature of learning in young children (though, interestingly, not in adults), for they can be seento exhibit this tendency in non-linguistic pattern-learning contexts too.) Newport concludes that “the contrasts between Simon and his parents are in certain ways less extreme, and more reorganizational, than might be suggested by the Language Bioprogram Hypothesis…[H]e does not appear to be creating an entirely newlanguage from his own innate specifications; rather, he appears to befollowing the predominant tendencies of his input, but he sharpens them, extends them, and forces them to be internally consistent.” (2001:173).

If Newportet al. are right, the case of Simon does not seem to give much support to the nativist hypothesis. Moreover, the argument from creolization suffers a number of additional flaws. First, the Bickerton-Pinker view, which assigns a dominant role to child language learners in the creation of creoles, is but one of three competing hypotheses currently being explored in the creolization literature. According to the ‘superstratist’ hypothesis, creolization occurs not when children acquire language from pidgins, but when successive waves of adult speakers try to learn the languageof the dominant culture as a second language. (Chaudenson 1992, for instance, defends this view about the origins of French creoles.) On this view, the additional devices seen in creoles are corruptions of devices seen in the dominant language. According to the ‘substratist’ hypothesis, creoles are again created by second language learners, rather than children, only the source of added structure is the first language of the learner. (Lumsden 1999 argues that numerous traces of a variety of African languages in Haitian creole support this hypothesis.) One need not take a stand onwhich of these views is correct in order to see that these competing explanations of creolization undermine Bickerton and Pinker's ‘bioprogram’ hypothesis. If creoles arise out of the attempts of adult learners to learn (and subsequently pass on to their children) another, non-native language, then what one might call ‘contamination of the stimulus,’ rather than the influence of an inborn UG in the learner, is what accounts for the UG-respecting ways in which creoles are elaborated.

However, there is a case of creolization in which these other hypotheses apparently fail to gain purchase, as Pinker (1994:37ff.) emphasizes. This is the case of the development of Idioma de Signos Nicaragüense (ISN, Nicaraguan Sign Language), a brand-new natural sign language which first emerged around 30 years ago in schools for the deaf in and around Managua. These schools were first set up in the 1970s, and ISN evolved from the hodge-podge of homesignsystems used by students who entered the schools at that time. ISN isan interesting test case of the bioprogram hypothesis for two reasons. First, homesign systems are idiosyncratic and possess littlesyntactic structure: the natural-languagelike syntax of ISN could therefore not derive from substrate influence. And Spanish, the only potential candidate for superstrate influence was allegedly inaccessible to signers because of its auditory modality. Pinker claims that ISN provides another example of creolization and the workings of the innate language faculty: it is “created…in one leap when the younger children were exposed to the pidgin singing of the older children.” (1994:36-7)

In their discussion of the development of ISN, however, Kegl, Senghasand Coppola (1999) show that things are not quite this straightforward. ISN did not develop ‘in one leap,’ from the very rudimentary homesigns or ‘Mimicas’ spoken by individual deaf students. Instead, its evolution was more gradual andwas preceded by the creation of what Keglet al. call “Lenguagede Signos Nicaragüense” (LSN), a kind of “pidgin or jargon” (181) that “developed from the point when these homesigners came together in the schools and began to share their homesigns with each other, quickly leading to more and more shared signs and grammatical devices” (180). In addition, the signers had access to Spanish language dictionaries, and their language was also influenced by the signing of Spanish-speaking, non-deaf teachersat the schools — signing which likely incorporated such grammatical devices of the teachers' language as were transferable toa non-vocal medium. (K. Stromswold, private communication.)

While Keglet al. endorse the language bioprogram hypothesis that ISNemerged ‘in one leap,’ in the minds of children exposed to degraded Mimica or LSN inputs, their data are equally consistent with the idea that ISN developed more gradually by means of successive elaborations and innovations among a community of highly-motivated (because language-starved) young users. Indeed, as Keglet al. themselves describe the history (p.187), this is precisely what happened. First, a group of signers, each with his or her own idiosyncratic form of Mimicas, entered the schools. Members of this group gained in expressive power as their individual Mimicas were enriched by borrowings from others' homesign systems. Then, a new cohort of Mimicas signers entered the schools. Their sign systemsbenefitted both from exposure to the Mimicas of their peers and from exposure to the richer system developed by the earlier group. Througha process of successive elaborations in this manner, LSN developed and then, by a similar series of steps, ISN developed. At present, all three sign systems are still being used in Nicaragua, presumably reflecting the different ages at which people are exposed to languageand the kinds of inputs (ISN or LSN vs. signed and written Spanish orlipreading in regular schools) they receive. In addition, ISN and to a lesser extent LSN are still constantly changing — as one would expect if ISN were a community-wide work in progress, not the finished product of an individual child's mind,

3.6 Developmental Language Disorders and the Search for ‘Language Genes’

Dissociations of language disorders acquired in adulthood (e.g., Broca's and Wernicke's aphasia) may tell us something about how language is organized in the mature brain, but cannot tell us much about how language is acquired or the role of innate knowledge in that process — a fact that nativists about language generally acknowledge. By contrast, language dissociations arising during childhood are sometimes held to bear strongly on the question of whether language is innate. Pinker (1994:297-314) articulates this latter line of thought, arguing that there is a double dissociation between ‘general intelligence’ and language in two developmental disorders called Williams Syndrome (WS) and Specific Language Impairment (SLI). People with WS have IQs well below the normal range (50-60), yet are able to speak fluently and engagingly about many topics. Those with SLI, by contrast, have normal (≈90) non-verbal intelligence but speak effortfully and slowly,frequently making errors in their production and comprehension of sentences and words. Pinker argues that there is a double dissociation here, and that it supports the view that there is a special ‘language acquisition device’ that is separable from any general learning abilities children might possess. In addition, following Gopnik 1990a,b, and Gopnik and Crago 1991, he urges that the fact that the dissociation appears to concern aspects of syntax in particular indicates that the language faculty in question is the grammar faculty. Finally, and again following Gopnik,he argues that since SLI appears to run in families and, in at least one case, displays a Mendelian inheritance pattern, what we have hereis evidence not just of a ‘grammar faculty,’ but of a ‘grammar gene.’

3.6.1 Williams Syndrome

WS is a rare genetic disorder with a complex phenotype. Physically, WS individuals display dismorphic facial features, abnormal growth patterns, gastrointestinal problems, early puberty, neurological abnormalities (including hypotonia, hyperreflexia, hyperacuisis and cerebellar dysfunction), defective vision and eye development, bad teeth, connective tissue abnormalities, and heart problems. Psychologically, in addition to their low non-verbal IQ and comparatively spared language abilities, they display relatively goodaudiovisual memory but very impaired visual-spatial abilities, leading to difficulties in daily life (e.g., getting dressed). They have outgoing personalities and are highly sociable to the point of overfriendliness, but also display numerous behavioral and emotional problems (especially hyperactivity and difficulty concentrating in childhood, and anxiety in later life). (Morris and Mervis 2000; Merviset al. 2000.)

As to their language, there is currently a debate in the literaturewith regard to its normalcy. According to one point of view, thelinguistic competence of WS individuals is remarkably normal,especially in comparison with that of similarly retarded individuals,such as those with Downs syndrome (Pinker 1994, 1997; Clahsen andAlmazan 1998; Belloet al. 2004; Bellugietal. 1998). While rather unusual in their choice of words (e.g.,producingchihuahua, ibis andcondor in addition tomore usual animal words in a word fluency test) and despitean excessive use of clichés and social stock-phrases, theirability to use language, especially in conversational contexts, ismore or less intact. For example, they may appear relatively normal insocial interactions, and their processing of conditional questions andability to repeat sentences with complex syntax is closer to that ofnormal controls than to matched Downs syndrome controls(Bellugiet al. 2000: 13, 15).

According to another school of thought, however, the language abilities of WS subjects might be more normal than those of Downs syndrome individuals, and might look remarkable in contrast to their own marked disabilities in other areas, but nonetheless display a number of abnormal characteristics across a variety of measures when investigated further. WS language shows “massively delayed” early acquisition, especially of vocabulary (Bellugiet al. 2000:11) and grammatical morphemes (Caprircietal. 1996); overregularization of regular plural and past tenseendings as well defective competence with regard to irregular nounsand verbs (Clahsen and Almazan 2001); “inordinate difficultywith morphosyntax” (Morris and Mervis 2000: 467; see alsoVolterraet al. 1996; Karmiloff-Smithet al. 1997;Levy and Hermon 2003); and impaired mastery of relative clauseconstructions (Grantet al., 2002), embedded sentences, and(in French) grammatical gender assignment (Karmiloff-Smithetal. 1997). Indeed, Bellugiet al., 2000, found that WSchildren's performance on a sentence-repetition task wasindistinguishable from that of matched controls diagnosed withSpecific Language Impariment, or SLI (see below, §3.6.2).Findings such as these lead experts such as Annette Karmilloff-Smithto urge “dethroning the myth” of WS' “intact”syntactic abilities (Karmiloff-Smithet al. 2003) and moveUrsula Bellugi — formerly a proponent of the ‘sparedlanguage’ viewpoint — to caution that “because theirlanguage abilities are often at a level that is higher than theiroverall cognitive abilities, individuals with WMS might be perceivedto be more capable than they really are.” (Bellugietal. 1999.)

In contrast to its cognitive profile, which is, as we have seen, a subject of debate, the genetic basis of WS is known. It results from a ≈1.5 Mb deletion encompassing the elastin gene ELN at chromosome 7q11.23; most cases appear to be due to new mutations. ELNis crucial in synthesizing elastin, a protein which holds cells together in the elastic fibers found in connective tissues throughoutthe body and in especially high concentrations in cartilege, ligaments and arterial walls. Failure to synthesize this protein disrupts development in numerous ways, from the first trimester onwards, and gives rise through processes that are not well understood to the raft of symptoms associated with the syndrome. (Morris and Mervis 2000; Merviset al. 2000.)

3.6.2 Specific Language Impairment

In contrast with Williams syndrome, in which one sees comparatively spared language in the face of mild to moderate mental retardation and numerous physical defects, specific language impairment (‘SLI’) is diagnosed when (i) non-verbal intelligence as measured by standard IQ tests is normal; (ii) verbal IQ is well belownormal; and (iii) obvious causes of language impairment (e.g., deafness, frank neurological damage) can be ruled out. As one might expect given these diagnostic criteria, a diagnosis of SLI embraces ahighly heterogeneous collection of language-related deficits, not allof which co-occur in every case of language impairment. (Bishop, 1994; Bishopet al,. 2000.) These include:

  • productive and receptive phonological deficits (e.g., difficultyproducing clusters of consonants, as inspectacle, andfailure to show categorical perception of phonemes differentiated byplace of articulation (/ba/ vs. /ga/) and voicing (/ba/ vs. /pa/);
  • morphological deficits (e.g., generation of past tenses or pluralsby using affixes);
  • productive and receptive syntactic deficits (e.g., analyzing‘reversible’ passives (Katie kissed Jacobvs.Jacob was kissed by Katie), complex dative constructions(e.g.,Katie gave Jacob the book) and anaphora (e.g.,Katie said that Sarah scratched her vs.Katie said thatSarah scratched herself).

As a consequence of this heterogeneity, SLI, researchers have introduced a number of subtypes of the disorder, including such things as ‘Verbal auditory agnosia,’ ‘Lexical-syntactic deficit syndrome’ and ‘Phonological programming syndrome’ (Bishop 1994). Also as a consequence, and in part because studies do not always distinguish between different subtypes, the etiology of SLI in general is not well understood (O'Brienet al. 2003), although recent research suggests at least two distinct genetic loci are involved in at least some subtypes of the disorder (Bishop 2006). Some posit an underlying defect in the ‘grammar module.’ For instance, Rice and Wexler (1996) attribute SLI individuals' morphological deficits to a missing UG principle, namely, the principle of inflection, and Van der Laly and Stollwerk 1997, attribute some SLI children's difficulty with anaphora to their failure to acquire Binding Theory. Others see non-linguistic defects,such as auditory, memory or processing deficits as the root problem. For instance, Tallal 1980, 1985 argue that many SLI cases result from deficits in the processing of rapid auditory stimuli, giving rise to a failure to learn to distinguish phonemes correctly, which in turn leads to a failure to acquire other aspects of grammar. Others, such as and Norbury, Bishop and Briscoe 2002 argue that such children's limited processing capacities are the culprit.

While the varied symptomatology of SLI suggests that no unified theory of its etiology might be forthcoming, the cause of the disorder is comparatively well understood in the case of one subtype,involving a severe disruption of morphosyntax (i.e., the rules governing the formation of words from smaller semantic units, or morphemes). This subtype, seen in about half the members of a large, three-generation English family, the KE's, and in another, unrelated individual, has been traced to a specific genetic mutation, the function of which is actively under investigation.

The KE family has received much press since the early 1990s, when Gopnik 1990a,b and Gopnik and Crago 1991 (see also Gopnik 1997) proposed that their morphosyntactic deficits were caused by a mutation in a single dominant gene normally responsible for the encoding of grammatical features, such as function words and the inflections used to mark number, tense, aspect, etc. According to Gopnik, the affected KE's are ‘feature blind’ as a consequence of this mutation. And according to Pinker (1994), their pedigree (Fig. 6) and specifically morphosyntactic deficits constitutes “suggestive evidence for grammar genes … genes whose effects seem most specific to the development of the circuits underlying parts of grammar” (Pinker 1994:325).

KE Diagram

Figure 6. The KE family pedigree
(Image used by permission of Simon E. Fisher)

Other intensive studies of the KE family, by Vargha-Khadem andcolleagues (e.g., Vargha-Khademet al 1995, 1998;Watkinset al., 2002) have vigorously disputed the hypothesisthat the root cause of the KE's language disorder is a syntacticdeficit. Instead, they argue, the KE phenotype is much broader thanGopnik's account suggests, and their ‘feature blindness’is merely one among the many effects of an underlying articulatoryproblem. As characterized by Vargha-Khadem's team, the affected KE'sspeech is effortful, “sometimes agrammatical and oftenunintelligible” (Watkinset al. 2002:453), and showsimpairments not just in morphosyntax (e.g., regular plural and pasttense endings) but also in the formation of irregular past tenses(where correct usage is lexically determined, rather thanrule-governed) and in sentence-level syntax, particularly word order.Comprehension, too, is impaired at the level of syntax as well aswords, and as is their reading of both words and non-words. Theseresults indicate that the KE's problems go beyond morphosyntax, andthe fact that affected KE's have significantly lower non-verbal IQs(by 18-19 points; Vargha-Khademet al. 1995) than unaffectedfamily members indicates that their deficits may be further reaching still.[28] Finally, affected KE's have trouble sequencing and executingnon-language-related face, mouth and tongue movements and showabnormal activation not just of speech but also of motor areas on fMRIscans (Liegeois,et al. 2003); this deficiency in‘orofacial praxis’ supports Vargha-Khadem's hypothesisthat the root problem for the KE's is articulatory.

As Gopnik noted, the pattern of inheritance in the KE family suggeststhat a single, dominant gene is responsible for the disorder. (Seefig. 6) In the early 1990's, Fisher and colleagues began working toisolate the gene. First, it was localized to a region on chromosome7q31 containing about 100 genes (Fisheret al. 1997, 1998;O'Brienet al. 2003). Later, it was identified (Laietal. 2001; Fisher et al. 2003) as the gene FOXP2, which encodes aregulatory protein or ‘transcription factor’ (i.e., aprotein that helps to regulate the rate of transcription of othergenes in the genome — in the case of FOXP2, the protein acts toinhibit transcription of the downstream gene(s)). In affected familymembers, a single base-pair substitution in the gene coding for thisregulatory protein leads to the insertion of the amino acid arginine(instead of the normal histamine) in an area of the protein (viz., the‘forkhead binding domain’) that is critical for itsability to modulate the transcription of the downstream DNA. As aconsequence, FOXP2 cannot perform its normal regulatory role inaffected KE family members.

The failure of FOXP2 to perform its normal role in turn leads toabnormal brain development in affected KE individuals. Studies ofother animals and humans (e.g., Laiet al. 2003;Takahashiet al., 2003; Ferlandet al. 2003;Teramitsuet al. 2004) show that FOXP2 is normally highlyexpressed in both development and adulthood in two distinct braincircuits, One is a corticostriatal circuit, in which inputs from theprefrontal and premotor cortex are modulated by the basal ganglia andthe thalamus, and then sent back to prefrontal and premotor coticalareas; the other is an olivocerebellar circuit, in which sensory inputis sent via the spinal cord for processing in the medulla, cerebellumand thalamus before being handed on to prefrontal cortex. (Seefig. 7.) The basal ganglia are known to be involved in the sequencingand reward-based learning of motor behaviors (Graybiel 1995, 1998),and the cerebellar circuit, while less well understood, is thought tobe a proprioceptive circuit involved in motor regulation andcoordination (Lieberman 2002). FOXP2 is expressed in homologues ofthese areas in other species (e.g., canaries, zebra finches, rats) andin all species studied, these areas are involved in motor sequencingand coordination (Sharff and White 2004).

Figure 7

Figure 7. Two circuits in which FOXP2 is expressed. (Based on figuresby Diana Weedman Molavi, The Washington University School of Medicine Neuroscience Tutorial).

So, what appears to be the case is that affected KE family members' language difficulties result from a mutation in the FOXP2 gene, whichresults in abnormal development of the striatal, cerebellar and cortical areas necessary for the sequencing and coordination of speech-related movements of the mouth, tongue and possibly larynx; MRI scans of affected family members showing reduced gray matter density in these areas support this hypothesis, as do fMRI scans showing abnormal striatal and cortical activation during receptive and active language processing (Beltonet al. 2003; Liegeois, et al. 2003.)

Vargha-Khadem speculates (cf. Watkinset al. 2002:463) that those ofthe KE's deficits that do not appear to be motor related (e.g., theircomprehension and reading problems, their difficulties with word order and syntax) are a result of impaired learning that itself results from their motor deficits. For instance, impaired articulation could lead to impoverished phonological representations,which would then impair the acquisition of morphological and morphosyntactic knowledge, which would then constitute a poor basis for further syntactic learning. Impaired representation at all of these levels would then express itself in receptive language and reading, as well as in the realm of spoken language. Another possibleexplanation of the KE's non-articulatory deficits, which is not necessarily in competition with the previous one, derives from the fact that the basal ganglia are also known to be implicated in working memory (Bosnan 2004) and reward-based learning (e.g., classical conditioning) that is mediated by dopaminergic circuits that interact with basilar structures (Lieberman 2002). If reward-based learning and working memory are impaired in the KE's, then this could explain not only their higher-level syntactic deficits, but also their overall lower IQ (Lieberman 2002).

Neither of these explanations of the KE's seems especially congenial to the linguistic nativist. For both tacitly assume that language learning, including syntactic learning, is not (or not entirely) subserved by special-purpose mechanisms. Rather, it is mediated by more general motor circuitry (according to the Vargha-Khadem hypothesis) or reward-based learning and working memory abilities (according to Lieberman) that are also involved in other learning tasks.

On the other hand, however, there is evidence that FOXP2 isparticularly implicated in vocal learning and expression. First, it ishighly expressed in songbirds that modify their innate vocalrepertoires: in canaries it is expressed seasonally, when adult birdsmodify their songs (Teramitsuet al. 2004) and in zebrafinches, it is expressed more at the time when young birds learn theirsongs (Haessleret al. 2004). In addition, there is evidencethat the variant of the FOXP2 gene that is present in humans hasundergone strong positive selection in the hominid line (Enardetal 2002; Zhanget al. 2002). The protein produced byhuman FOXP2 differs in just three out of its 715 constituent aminoacids from that of the mouse, and a recent analysis (Zhangetal. 2003) indicates that two of these differences are unique tothe hominid lineage. According to Enardet al 2002, the factthat these two differences are fixed in the human genome, whereas nofixed substitutions occurred in the lineage of our closest relatives,the chimpanzees, suggests that those changes were strongly selectedfor in our lineage; Enardet al. put the date of fixation ofthese changes in the human population at around 200,000 yearsago. This date accords well with at least some estimates of theemergence of modern human language, suggesting that the vocalcapacities underwritten by FOXP2 — and impaired in those lackingthe gene — are after all critical to language competence.

3.6.3 The grammar module and the genetics of language

At this point, two questions arise. First, is there a double dissociation between language capacities and general cognitive capacities to be found in a comparison of Williams syndrome and SLI? Second, what does our current knowledge of the role of FOXP2 in language development tell us about linguistic nativism?

As to the first question, there seems to be no double dissociation. First of all, WS individuals' language, while startling in contrast with their level of mental retardation, is not normal; indeed, as we have seen, it is indistinguishable on some tests from that of language impaired individuals. In addition, as Thomas and Karmiloff-Smith 2002, caution, it is not at all clear that one can assume, in the case of a pervasive developmental disorder like Williams syndrome, that apparently ‘intact’ competences are a result of normal development of the underlying neurological andpsychological structures. That is, given the known capacity of the brain to compensate for deficits in one area by cobbling together a solution in another, one cannot assume that there is a ‘language module’ in WS patients which develops more or less normally despite other cognitive systems' being massively disrupted. Thomas and Karmiloff-Smith argue that the numerous discrepancies between WS language development and that of normal children suggests that this ‘residual normality’ assumption is misguided in this case, thus undermining the claim thatwhat is spared in WS is ‘the language (or grammar) module.’

Moving to the other side of the dissociation, since it is hard to sayexactly what about language is disrupted in cases of SLI, it isdifficult to determine whether this disruption is specific tolanguage, let alone grammar. While researchers like Van der Lely andChristian 2000, and van der Lely and Ullman 2001 argue that there isa purely grammatical form of the deficit, which does support thehypothesis of a grammar module, this is controversial, as we have seenabove. Certainly consideration of the KE's does not support such ahypothesis. Their root deficit appears to concern orofacial praxis,rather than language specifically; and in addition, their‘general intelligence,’ as measured by tests of non-verbalintelligence, while “normal,” nonetheless appears to havebeen affected by their neurological and/or linguistic abnormalities— witness their scores 18-19 points lower than those of theirrelatives. It is, in other words, unclear that there is anydissociation of language and general intelligence in this case atall. One can conclude that as things stand now, SLI seems to be soheterogeneous a disorder as to defy neat characterization, and thatconsideration of this disorder does not support the view that there isa language or grammar module that functions independently of othercognitive processes.

The second question asked above was: what can be learned about the innate basis of language from a consideration of the KE's and FOXP2. In a recent article, Marcus and Fisher (2003) argue that the kinds ofresults discussed above offer valuable insights into the ways that language is implemented in the brain and controlled (to the extent that it is) by the genes. However, they refrain — rightly in myview — from drawing morals to the effect that FOXP2 is a “gene for language” or even “for articulation.” The effects of FOXP2 are wider than this (it is expressed in the developing heart and lungs, in addition to the brain-- REF) and the functions of the neural circuits in which it is active are as yet too poorly understood to do more than gesture at the ways in which FOXP2 is involved in constructing the human linguistic phenotype.

All the topics covered in §3 deserve books of their own. My aimhere has been to sketch the ways in which modern understanding of themind reveals the inadequacy and implausibility of the claim thathumans have innate representations of UG that are responsible fortheir acquisition of language. There are likely many, many processesimplicated in the attainment of linguistic competence, that many ofthem are likely specialized by natural selection for linguistic tasks,but that many of them also retain their other, and older,functions. The linguistic nativist's theory views our acquisition ofgrammatical competence as asimple matter — one thatcan be described at one level of explanation, and in terms of a singlekind of process. This is very unlikely to be the case. Multiplesystems and multiple processes are at work in the acquisition oflinguistic knowledge, and our understanding of language acquisition,when it comes, is likely to involve theories of many kinds and at manydifferent levels, and to resemble the theory of the Chomskyan nativistin few or no respects.[29]

Bibliography

  • Adolphs, R. (1999). Social Cognition and the HumanBrain.Trends in Cognitive Science, 3(12), 469-479.
  • Avital, E. and Jablonka, E. (2000).Animal Traditions:Behavioural Inheritance in Evolution, Cambridge: CambridgeUniversity Press,
  • Ayoun, D. (2003).Parameter Setting in LanguageAcquisition, London: Continuum International Publishing Group,Incorporated.
  • Baker, C.L. (1979). “Syntactic Theory and the ProjectionProblem”,Linguistic Inquiry, 10: 533-81.
  • Bates, E. and Roe, K. (2001). “Language development inchildren with unilateral brain injury”, inHandbook ofDevelopmental Cognitive Neuroscience, C.A. Nelson andM. Luciana,(eds.), Cambridge: MIT Press, pp. 281-307.
  • Beale, J.M. and Keil, F.C. (1995). “Categorical Effects inthe Perception of Faces”,Cognition, 57: 217-239.
  • Bello, A., Capirci, O., & Volterra, V. (2004). Lexicalproduction in children with Williams syndrome: spontaneous use ofgesture in a naming task.Neuropsychologia, 42(2),201-213.
  • Bellugi, U., & Lai, Z. (1998). Neuropathological and cognitivealterations in Williams syndrome and Down syndrome.Faseb Journal,12(4), A355-A355.
  • Bellugi, U., Lichtenberger, L., Mills, D., Galaburda, A., &Korenberg, J. R. (1999). Bridging cognition, the brain and moleculargenetics: evidence from Williams syndrome.Trends inNeurosciences, 22(5), 197-207.
  • Bellugi, U., L. Lichtenberger,et al. (2000). “Theneurocognitive profile of Williams syndrome: A complex pattern ofstrengths and weaknesses.”Journal of CognitiveNeuroscience 12: 7-29.
  • Belton, E., C. H. Salmond,et al. (2003). “Bilateral brainabnormalities associated with dominantly inherited verbal andorofacial dyspraxia.”Human Brain Mapping 18(3):194-200.
  • Berwick, R.C. and Weinberg, A.S. (1983).The Grammatical Basisof Linguistic Performance: Language Use and Acquisition,Cambridge, MA: MIT Press.
  • Berwick, R. C. (1998). Language evolution and the MinimalistProgram: the origins of syntax. In J. R. Hurford, M. Studdert-Kennedy& C. Knight (Eds.),Approaches to the Evolution ofLanguage (pp. 320-340). Cambridge: Cambridge UniversityPress.
  • Bickerton, D. (1981).Roots of Language: KaromerPublishers, Inc.
  • ––– (1998). Catastrophic evolution: the case for a single stepfrom protolanguage to full human language. In J. R. Hurford, M.Studdert-Kennedy & C. Knight (Eds.),Approaches to theEvolution of Language (pp. 341-358). Cambridge: CambridgeUniversity Press.
  • ––– (2000).Language and Species,Chicago : University of Chicago Press.
  • Birdsong, D. and Molis, M. (2001). “On the evidence formaturational constraints in second-language acquisition”,Journal of Memory and Language, 44: 235-49.
  • Bishop, D. V. M. (1994). “Is Specific Language Impairment AValid Diagnostic Category - Genetic And PsycholinguisticEvidence.”Philosophical Transactions Of The Royal SocietyOf London Series B-Biological Sciences 346(1315): 105-111.
  • –––, P. Bright,etal. (2000). “Grammatical SLI: A distinct subtype ofdevelopmental language impairment?”AppliedPsycholinguistics 21(2): 159-181.
  • Bohannon, J.N. and Stanowicz, L. (1988). “The issue ofnegative evidence: Adult responses to children's languageerrors”,Developmental Psychology, 24: 684-89.
  • Bohannon, J.N., MacWhinney, B. and Snow, C. (1990). “Nonegative evidence revisited: Beyond learnability or who has to provewhat to whom”,Developmental Psychology, 26:221-26.
  • Bohannon, J. N., Padgett, R. J., Nelson, K. E., & Mark, M.(1996). Useful evidence on negative evidence.DevelopmentalPsychology, 32(3), 551-555.
  • Bosman, C., R. Garcia,et al. (2004). “FOXP2 and thelanguage working-memory system.”Trends In CognitiveSciences 8(6): 251-252.
  • Botha, R. P. (2003).Unravelling the Evolution ofLanguage. U.S.: Elsevier Science.
  • Bradlow, A.R., Pisoni, D.B., Akahane-Yamada, R., and Tohkura, Y.(1997). “Training Japanese listeners to identify English /r/ and/l/: IV. Some effects of perceptual learning on speechproduction”,Journal of the Acoustical Society ofAmerica, 101: 2299-2310.
  • Brandon, R.N. and Hornstein, N. (1986a). “From Icons toSymbols: Some Speculations on the Origins of Language”,Biology and Philosophy, 1 :169-89.
  • Brown, R. (1973).A first language: The early stages,Cambridge, MA: Harvard University Press.
  • Brown, R., Cazden, C. and Bellugi, U. (1969). “The Child'sGrammar from I to III”, inMinnesota Symposium on ChildPsychology, Vol. II, J. Hill (ed.), Minneapolis: University ofMinnesota Press.
  • Brown, R. and Hanlon, C. (1970). “Derivational Complexity andOrder of Acquisition in Child Speech”, inCognition and theDevelopment of Language, J. R. Hayes (ed.), New York: John Wileyand Sons.
  • Buonomano, D.V. and Merzenich, M.M. (1998). “Corticalplasticity: From synapses to maps”,Annual Review ofNeuroscience, 21: 149-186.
  • Campanella,. S., Chrysochoos, A., Bruyer, R. (2001).“Categorical perception of facial gender information:Behavioural evidence and the face-space metaphor”,VisualCognition, 8 (2): 237-262.
  • Capirci, O., L. Sabbadini,et al. (1996). “Languagedevelopment in Williams syndrome: A case study.”CognitiveNeuropsychology 13(7): 1017-1039.
  • Catania, A.C. (1990). “What good is 5 percent of a languagecompetence”,Behavioral and Brain Sciences, 13 :729-730.
  • Cavalli-Sforza, L.L. (1997). “Genes, peoples, andlanguages”, PNAS: 94: 7719-772
  • Chambers, K.E., Onishi, K.H., and Fisher, C. (2003). “Infantslearn phonotactic regularities from brief auditory experience”,Cognition, 87: 69-77.
  • Chater, N., & Manning, C. D. (2006). Probabilistic models oflanguage processing and acquisition.Trends in Cognitive Sciences,10(7), 335-344.
  • Chaudenson, R. (1992).Des îles, des hommes, deslangues. Paris: L'Harmattan.
  • Chomsky, N. (1957).Syntactic Structures, The Hague :Mouton.
  • ––– (1959). Review of Skinner'sVerbalBehavior,Language, 35: 26-58.
  • ––– (1965).Aspects of the Theory ofSyntax, Cambridge, MA: MIT Press.
  • ––– (1975).Reflections on Language, London :Fontana.
  • ––– (1980).Rules and Representations. New York:Columbia University Press.
  • ––– (1981).Lectures on Government andBinding, Hawthorne, NY: Walter De Gruyter Incorporated.
  • ––– (1982).The Generative Enterprise,Dordrecht: Foris Publications.
  • ––– (1986).Knowledge of Language, ItsNature, Origin and Use, NY: Praeger.
  • ––– (1988).Language and Problems ofKnowledge, The Managua Lectures, Cambridge, MA: MIT Press.
  • ––– (1975a).The Logical Structure ofLinguistic Theory, NY: Plenum.
  • ––– (1980).Rules and Representations,NY: Columbia University Press.
  • ––– (1990). “On the nature, acquisitionand use of language”, inMind and Cognition: A Reader,W.G. Lycan (ed.), Cambridge MA and London UK: Blackwells,pp.627-45.
  • ––– (1995).The Minimalist Program,Cambridge: MIT Press.
  • Chouinard, M.M. and Clark, E.V. (2003) “Adult reformulationsof child errors as negative evidence,”Journal of ChildLanguage, 30:637-69.
  • Clahsen, H. and Muysken, P. (1986). “The availability ofuniversal grammar to adult and child learners”,SecondLanguage Research, 2: 93-119.
  • ––– and M. Almazan (1998). “Syntax andmorphology in Williams syndrome.”Cognition 68(3):167-198.
  • ––– (2001). Compounding and inflection in language impairment:Evidence from Williams Syndrome (and SLI).Lingua, 111(10),729-757.
  • Clark, A. (1997).Being There: Putting Brain, Body, and WorldTogether Again, Cambridge, MA: MIT Press.
  • Clark, A. (1997).Being There: Putting Brain, Body and WorldTogether Again. Cambridge, Ma: Bradford Books/MIT Press.
  • Cowen T. and Gavazzi, I. (1998). “Plasticity in adult andageing sympathetic neurons”,Progress in Neurobiology,54:249-88.
  • Cowie, F. (1997). The Logical Problem of Language Acquisition.Synthese, 111, 17-51.
  • ––– (1999).What's Within: NativismReconsidered, New York: Oxford University Press.
  • Crain, S. and Pietroski, P.M. (2001). “Nature, Nurture, andUniversal Grammar”,Linguistics and Philosophy, 24:139-86.
  • ––– (2002). “Why language acquisition is asnap”,Linguistic Review, 19 (1-2): 163-183.
  • Crain, S. (1991). “Language acquisition in the absence ofexperience”,The Behavioral and Brain Sciences, 4:597-650.
  • Curtiss, S. (1977).Genie: a Psycholinguistic Study of aModern-day “Wild Child”. New York: AcademicPress.
  • Dawkins, R. and Krebs, J.R. (1979). “Arms races between andwithin species”,Proceedings of the Royal Society of London,Series B — Biological Sciences, 205: 489-511.
  • DeCaspar, A.J. and Spence, M.J. (1986). “Prenatal maternalspeech influences newborns' perception of speech sounds”,Infant Behaviorand Development, 9:133-150.
  • DeGraff, D. (1999a). Creolization, Language Change, and LanguageAcquisition. In M. Degraff (Ed.),Language Creation and LanguageChange: Creolization, diachrony, and development. (pp. 1-46).Cambridge, MA: MIT Press.
  • DeGraff, D. (Ed.). (1999b).Language Creation and LanguageChange: Creolization, Diachrony and Development. Cambridge, Ma.:MIT Press.
  • Demetras, M.J., Post, K.N. and Snow, C.E. (1986). “Feedbackto first language learners: The role of repetitions and clarificationquestions”,Journal of Child Language, 13: 275-92.
  • Descartes, R. (1984), “Discourse on the Method”, inThe Philosophical Writings of Descartes, Vol. II,J. Cottingham, R. Stoothoff and D. Murdoch (eds.), and trans. (1984),Cambridge: Cambridge University Press.
  • Devitt M. (2006).Ignorance of Language, Oxford: OxfordUniversity Press.
  • Dick F., Bates E., Wulfeck B., Utman J.A., Dronkers N. andGernsbacher M.A. (2001). “Language deficits, localization, andgrammar: Evidence for a distributive model of language breakdown inaphasic patients and neurologically intact individuals”,Psychological Review, 108: 759-788.
  • Dronkers, N.F., Redfern, B.B. and Knight, R.T. (2000). “TheNeural Architecture of Language Disorders”, inThe NewCognitive Neurosciences, 2ndEd.,M.S. Gazzaniga (ed.), Cambridge: MIT Press, pp. 949-58.
  • Dunbar, R. (1996).Grooming, Gossip and the Evolution ofLanguage, Cambridge, Ma.: Harvard University Press.
  • Eimas, P.D. (1975). “Speech perception in earlyinfancy”, inInfant Perception, Vol.2. From sensation tocognition, L. Cohen and P. Aalapatel (Eds.), New York: AcademicPress. Pp.193-231.
  • Elbert, T., Heim, S. and Rockstroh, B. (2001). InHandbook ofDevelopmental Cognitive Neuroscience, C.A. Nelson and M. Luciana,Cambridge: MIT Press, pp. 191-202.
  • Elman, J.L., Bates, E.A., Johnson, M.H., Karmiloff-Smith, A.,Parisi, D. and Plunkett, K. (1996).Rethinking Innateness: AConnectionist Perspective on Development, Cambridge: BradfordBooks/MIT Press.
  • Elman, J.L. (1998). “Generalization, simple recurrentnetworks, and the emergence of structure”, inProceedings ofthe Twentieth Annual Conference of the Cognitive Science Society,M.A. Gernsbacher and S.J. Derry (eds.), Mahwah, NJ: Lawrence ErlbaumAssociates.
  • Enard, W., Przeworski, M., Fisher, S., Lai, C., Wiebe, V., Kitano,T.,et al. (2002). “Molecular evolution of FOXP2, a geneinvolved in speech and language”,.Nature, 418,869-872.
  • Epstein, S., Flynn, S., and Martohardjono,G. (1996). “Second language acquisition: theoretical andexperimental issues in contemporary research,” Brain andBehavioral Sciences, 19:677-758.
  • ––– (1998). “The strong continuityhypothesis: some evidence concerning functional categories in adult L2acquisition”. In The generative study of second languageacquisition, S. Flynn, G. Martohardjono and W. O'Neil (eds.), Mahwah,NJ: Lawrence Erlbaum, pp.61-77.
  • Esser, K-H., Condon, C.J., Suga, N., and Kanwal, J.S. (1997).“Syntax processing by auditory cortical neurons in theFM—FM area of the mustached bat Pteronotus parnellii”,Proc. Natl. Acad. Sci. USA. 94: 14019—14024.
  • Etcoff, N.L., Magee, J.J. (1992). “Categorical Perception ofFacial Expressions”,Cognition 44: 227-240.
  • Farrar M.J. (1990). “Discourse and the acquisition ofgrammatical morphemes”,Journal of Child Language, 17:607-24.
  • ––– (1992). “Negative Evidence andGrammatical Morpheme Acquisition”,DevelopmentalPsychology, 28: 90-98.
  • Feldman, J. (1972). “Some decidability results on grammaticalinference and complexity”,Information and Control, 20:244-62.
  • Ferland, R. J., T. J. Cherry,etal. (2003). “Characterization of Foxp2 and Foxp1 mRNA andprotein in the developing and mature brain.”Journal OfComparative Neurology 460(2): 266-279.
  • Fisher, S. E., F. Vargha-Khadem,etal. (1997). “Localisation of a gene implicated in a severespeech and language disorder.”American Journal Of HumanGenetics 61(4): A28-A28.
  • –––, F. Vargha-Khadem,etal. (1998). “Localisation of a gene implicated in a severespeech and language disorder,”Nature Genetics 18(3):298-298.
  • –––, C. S. L. Lai,etal. (2003). “Deciphering the genetic basis of speech andlanguage disorders.”Annual Review of Neuroscience 26:57-80.
  • Flege, J.E., Yeni-Komshian, G.H. and Liu, S. (1999). “Ageconstraints on second language acquisition”,Journal ofMemory and Language, 41: 78-104.
  • Fodor, J.A. (1981). “The Present Status of the InnatenessControversy”, inRePresentations: Philosophical Essays onthe Foundations of Cognitive Science, Cambridge: MITPress/Bradford Books, pp. 257-316.
  • Fodor, J.A. (1998).Concepts: Where Cognitive Science wentWrong, New York: Oxford University Press.
  • Gold, E.M. (1967). “Language Identification in theLimit”,Information and Control, 10: 447-74.
  • Goldin-Meadow, S., and Mylander, C. (1983). “Gesturalcommunication in deaf children: noneffect of parental input onlanguage development”,Science, 221: 372-4.
  • Goldin-Meadow, S., and Mylander, C. (1990). “The role ofparental input in the development of a morphological system”,Journal of Child Language, 17(3): 527-63.
  • Goldin-Meadow, S., and Mylander, C. (1998). “Spontaneous signsystems created by deaf children in two cultures”,Nature, 391:279-81.
  • Gopnik, M. (1997). “Language deficits and geneticfactors”,Trends in Cognitive Science, 1: 5-9
  • ––– (1990a) “Feature-blind grammar anddysphasia”,Nature, 344:715.
  • ––– (1990b) “Genetic basis of a grammardefect”,Nature, 347: 26.
  • –––, and Crago, M. (1991). “Familialaggregation of a developmental language disorder”,Cognition, 39: 1-50.
  • Grant, J., V. Valian,et al. (2002). “A study of relativeclauses in Williams syndrome.”Journal Of ChildLanguage 29(2): 403-416.
  • Graybiel, A. M. (1995). “Building action repertoires: Memoryand learning functions of the basal ganglia.”CurrentOpinion In Neurobiology 5(6): 733-741.
  • Graybiel, A. M. (1998). “The basal ganglia and chunking ofaction repertoires.” Neurobiology Of Learning And Memory70(1-2): 119-136.
  • Griffiths, P. (2002). “What is innateness?”,Monist, 85: 70-85.
  • Grimshaw, G.M., Adelstein, A., Bryden, M.P., and MacKinnon, G.E.,1998, “First language acquisition in adolescence: Evidence for acritical period for verbal language development”,Brain andLanguage, 63: 237-55.
  • Haesler, S., K. Wada,et al. (2004). “FoxP2 expression inavian vocal learners and non-learners.”Journal ofNeuroscience 24(13): 3164-3175.
  • Hakuta, K., Bialystok, E., and Wiley, E. (2003). “Criticalevidence: a test of the critical-period hypothesis for second-languageacquisition”,Psycholigical Science, 141: 31-8.
  • Hamilton, R.H. and Pascual-Leone, A. (1998). “Corticalplasticity associated with Braille learning”,Trends inCognitive Science, 2: 168-174.
  • Harman, G.H. (1967). “Psychological Aspects of the Theory ofSyntax”,Journal of Philosophy, 64: 75-87.
  • Harman, G. (1969). Comments on Linguistic Competence and Nativism.Synthese, 19, 410-424.
  • Harnad, S.R., ed. (1987).Categorical Perception : TheGroundwork of Cognition, Cambridge: Cambridge UniversityPress.
  • Harris, Z.S. (1951).Structural Linguistics, Chicago:University of Chicago Press.
  • Hauser, M.D., Chomsky, N., and Fitch, W.Y. (2002). “TheFaculty of Language: What Is It, Who Has It, and How Did ItEvolve?”,Science, 298: 1569-1579.
  • Hauser, M. D. and McDermott, J. (2003). “The evolution of themusic faculty: a comparative perspective”,NatureNeuroscience, 6: 663-668.
  • Hauser, M.D., Weiss, D., and Marcus, G. (2002). “Rulelearning by cotton-top tamarins”,Cognition, 86:15-22.
  • Hirsh-Pasek, K., Trieman, R. and Schneidermann, M. (1984).“Brown and Hanlon Revisited: Mothers' sensitivity toungrammatical forms”,Journal of Child Language, 11:81-88.
  • Holland, A.L., Fromm, D.S., DeRuyter, F. and Stein, M. (1996).“Treatment efficacy: aphasia”,Journal of Speech andHearing Research, 39: S27-36.
  • Hornstein, N. and Lightfoot, D., (eds.) (1981).Explanation inLinguistics: The Logical Problem of Language Acquisition, London:Longman.
  • Hubel, D.H. and Wiesel, T.N. (1970). “The period ofsusceptibility to the physiological effects of unilateral eye closurein kittens”,Journal of Physiology, 206: 419-436.
  • Jackendoff, R. (1999). “Possible stages in the evolution ofthe language capacity,”Trends in Cognitive Sciences,3:272-279.
  • Jain, S., Osherson, D., Royer, J.S. and Sharma, A. (1999).Systems that Learn, 2nd Ed., Cambridge, MA: MIT Press.
  • Johnson, J. S., & Newport, E. L. (1989). “Criticalperiod effects in 2nd language-learning - the influence ofmaturational state on the acquisition of english as a 2ndlanguage.”Cognitive Psychology, 21(1), 60-99.
  • Johnson, J.S. and Newport, E.L. (1991). “Critical PeriodEffects on universal properties of language: the status of subjacencyin the acquisition of a second language”,Cognition,39: 215-58.
  • Karbe, H., Thiel, A., Weber-Luxemberger, G., Herholz, K., Kessler,J. and Heiss, W. (1998). “Brain Plasticity in Poststroke Aphasia:What Is the Contribution of the Right Hemisphere?”,Brainand Language, 64: 215-230.
  • Karmiloff-Smith, A., Grant, J., Berthoud, I., Davies, M., Howlin,P., and Udwin, O. (1997). “Language and Williams Syndrome: Howintact is intact?”,Child Development, 68(2):246-62.
  • –––, J. H. Brown,etal. (2003). “Dethroning the myth: Cognitive dissociations andinnate modularity in Williams syndrome.”DevelopmentalNeuropsychology 23(1-2): 227-242.
  • Katz, L.C., Weliky, M., and Crowley, J.C. (2000). “Activityand the development of the visual cortex: new perpsectives”, inThe New Cognitive Neurosciences, 2ndEd.,M.S. Gazzaniga (ed.), Cambridge: MIT Press, pp. 199-212.
  • Kegl J., Senghas A., Coppola M. (1999). “Creation throughcontact: Sign language emergence and sign language change inNicaragua.” In M. DeGraff (ed),Comparative GrammaticalChange: The Intersection of Language Acquisition, Creole Genesis andDiachronic Syntax, Cambridge MA: MIT Press, pp.179-237.
  • Kellerman, E., van Ijzendorn, J., and Takashima, H. (1999).“Retesting a universal: the Empty Category Principle andlearners of (pseudo)Japanese”, inThe acquisition ofJapanese as a second language, K. Kanno (ed.), Amsterdam: JohnBenjamins, pp. 71-87.
  • Kellerman, E. and Yoshioka, K. (1999). “Inter- andintra-population consistency: a comment on Kanno (1998)”,Second Language Research, 15: 101-9.
  • Kotsoni E., de Haan, M. and Johnson, M.H. (2001).“Categorical perception of facial expressions by 7-month-oldinfants”,Perception, 30 (9): 1115-1125.
  • Kuhl, P.K. (1994). “Learning and Representation in Speech andLanguage”,Current Opinion in Neurobiology, 4:812-822.
  • ––– (2000). “A new view of language acquisition”,PNAS, 97 (22): 11850-11857.
  • –––, and Miller, J.D. (1975). “Speech Perception by theChinchilla: Voiced-Voiceless Distinction in Alveolar PlosiveConsonants”,Science, 190: 69-72.
  • –––, Andruski, J.E., Chistovich, I.A., Chistovich, L.A.,Kozhevnikova, E.V., Ryskina, V.L., Stolyarova, E.I., Sundberg, U.,Lacerda, F. (1997a). “Cross-Language Analysis of Phonetic Unitsin Language Addressed to Infants”,Science, 277:684-686.
  • –––, Kiritani, S., Degughi, T., Hayashi, A., Stevens, E.B.,Dugger, C.D. and Iverson, P. (1997b). “Effects of languageexperience on speech perception: American and Japanese infants'perception of /ra/ and /la/”,Journal of the AcousticalSociety of America, 100: 2425-38.
  • Kujala, T., Alho, K., Huotilainen, M., Ilmoniemi, R.J., Lehtoski,A., Leionen, A., Rinne, T., Salonen, O., Sinkkonen, J.,Standertskjold-Nordenstam, C.G., and Naatanen, R. (1997).“Electrophysiological evidence for cross-modal plasticity inhumans with early- and late-onset blindness”,Psychophysiology, 34: 213-6.
  • Jones, P. (1995). Contradictions and unanswered questions in theGenie case: A fresh look at the linguistic evidence.Language andCommunication, 15(3), 261-280.
  • Lai CSL, Fisher SE, Hurst JA, Vargha-Khadem F, Monaco AP (2001).“A forkhead-domain gene is mutated in a severe speech andlanguage disorder”,Nature, 413 (6855): 519-523.
  • –––, D. Gerrelli,et al. (2003). “FOXP2expression during brain development coincides with adult sites ofpathology in a severe speech and language disorder.”Brain 126: 2455-2462.
  • Lasnik, H. (1989). “On Certain Substitutes for NegativeData”, Demopoulos and Marras (1989:89-105).
  • Lasnik, H. and Uriagereka, J. (1986).A Course in GB Syntax:Lectures on Binding and Empty Categories (Current Studies inLinguistics), Cambridge: MIT Press.
  • Lenneberg, E.H. (1964). “A biological perspective oflanguage”, in E.H. Lenneberg, ed.,New Directions in theStudy of Language, Cambridge: MIT Press. Pp.65-88
  • ––– (1967).Biological Foundations ofLanguage, New York: Wiley.
  • Levy, Y. and Hermon (2003). “Morphological abilities ofHebrew-speaking adolescents with Williams syndrome,”Developmental Neuropsychology, 23(1-2):59-83
  • Lewontin, R. (1998). “The Evolution of Cognition: Questionswe will Never Answer,” In R. Sternberg and D. Scarborough(eds.),An Invitation to Cognitive Science, Vol. 4: Methods,Models and Conceptual Issues, Cambridge, MIT Press,pp.107-32.
  • Lenneberg, E.H. (1964). “A biological perspective oflanguage”, in E.H. Lenneberg, ed.,New Directions in theStudy of Language, Cambridge: MIT Press. Pp.65-88
  • ––– (1967).Biological Foundations of Language, NewYork: Wiley.
  • Lieberman, P. (2002). “On the nature and evolution of theneural bases of human language,”American Journal ofPhysical Anthropology, 119, pp. 36-62.Liegeois, F., T. Baldeweg,et al. (2003). “Language fMRI abnormalities associated withFOXP2 gene mutation.”Nature Neuroscience 6(11):1230-1237.
  • Lumsden, J.S. (1999). “Language acquisition andcreolization”,Language Creation and Language Change:Creolization, Diachrony and Development, M. DeGraff (ed.),Cambridge: MIT Press, pp.129-57.
  • Maratsos, M.P. (1989). “Innateness and plasticity in languageacquisition,” in M.L. Rice and R.L. Schiefelbusch, eds.,The Teachability of Language, Baltimore, Paul H. Brooks,105-25.
  • Markson, L. (2006). Core mechanisms of word learning. InProcesses of Change in Brain and Cognitive Development: Attentionand Performance Xxi (pp. 111-128). New York: Oxford UnivPress.
  • Martin, R.C. (2003). “Language Processing: FunctionalOrganization and Neuroanatomical Basis”,Annual Review ofPsychology, 54:55-89.
  • Mayberry R.I., Eichen, E.B. (1991). “The Long-LastingAdvantage of Learning Sign Language in Childhood - Another Look at theCritical Period for Language-Acquisition”,J Mem Lang,30 (4): 486-512.
  • Macdonald, C., (ed) (1995). “Tacit Knowledge” inPhilosophy of Psychology: Debates on PsychologicalExplanation, Cambridge: Blackwell.
  • McGonigle, B., Chalmers, M., and Dickinson, A. (2003).“Concurrent disjoint and reciprocal classification by Cebusapella in seriation tasks: evidence for hierarchicalorganization”,Animal Cognition Vol. 6, No. 3:185-197.
  • Maess, B., Koelsch, S., Gunter, T. C. & Friederici, A. D.,2001, “Musical Syntax is processed in Broca's area: an MEGstudy”,Nature Neuroscience, 4: 540-545.
  • Marcus, G. F. (1998). “Can connectionism saveconstructivism?”Cognition, 66: 153-182.
  • ––– (2001).The Algebraic Mind: IntegratingConnectionism and Cognitive Science, Cambridge, MA: MITPress.
  • ––– (2004).The Birth of the Mind: How a Tiny Number of GenesCreates the Complexity of Human Thought, Basic Books, NewYork.
  • Margolis, E. (1998). “How to Acquire a Concept”,Mind and Language, 13(3): 347-369.
  • Matthews, R.J. (2001). “Cowie's Anti-Nativism”,Mind and Language, 16: 215-230.
  • von Melchner, L., Pallas, S.L. and Sur, M. (2000). “Visualbehaviour mediated by retunal projections directed to the auditorypathway”,Nature, 404: 871-76.
  • Mervis, C. B., B. F. Robinson,et al. (2000). “The Williamssyndrome cognitive profile.”Brain and Cognition,44(3): 604-628.
  • Mintz T.H. (2002). “Category induction from distributionalcues in an artificial language”,Memory and Cognition,30: 678-686.
  • Moerk, E. (1991). “Positive evidence for negativeevidence”,First Language, 11: 219-51.
  • Mohr, J.P., Pessin, M.S., Finkelstein, S., Funkenstein, H.H.,Duncan, G.W. and Davis, K.R. (1978). “Broca aphasia: pathologicaland clinical”,Neurology, 28: 311-24.
  • Morgan, J. L., Meier, R. P., & Newport, E. L. (1987).“Structural packaging in the input to language-learning -contributions of prosodic and morphological marking of phrases to theacquisition of language.”Cognitive Psychology, 19(4),498-550.
  • Morgan, J.L. and Travis, L.L. (1989). “Limits on negativeinformation in language input”,Journal of ChildLanguage, 16: 531-552.
  • Morgan, J.L., Bonamo, K.M. and Travis, L.L. (1995). “NegativeEvidence on Negative Evidence”,DevelopmentalPsychology, 31: 180-197.
  • Morris, C. A. and C. B. Mervis (2000). “Williams syndromeand related disorders.”Annual Review of Genomics and HumanGenetics 1: 461-484.
  • Mufwene, S.S. (1999). “On the language bioprogram hypothesis:hints from Tarzie”,Language Creation and Language Change:Creolization, Diachrony and Development, M. DeGraff (ed.),Cambridge: MIT Press, pp.95-127.
  • Muller R.A., Rothermel R.D., Behen M.E., Muzik O., ChakrabortyP.K., and Chugani H.T. (1999). “Language organization in patientswith early and late left-hemisphere lesion: a PET study”,Neuropsychologia, 37: 545-557.
  • Myhill, J. (1991). “Typological test analysis: tense andaspecte in creoles and second languages”, inCrosscurrentsin Second Language Acquisition and Linguistic Theories,T. Huebner and C.A. Ferguson (eds.), Amsterdam: John Benjamins,pp. 93-121.
  • Nazzi, T. and Ramus, F. (2003). “Perception and acquisitionof linguistic rhythm by infants”,Journal of SpeechCommunication, 41: 233-43.
  • Nelson D.A., Marler P. (1989). “Categorical perception of anatural stimulus continuum: birdsong”,Science, 244:976-8.
  • Newell, F.N., Bulthoff, H.H. (2002). “Categorical perceptionof familiar objects”,Cognition 85 (2): 113-143
  • Newmeyer, F.J. (1986).Linguistic Theory in America, 2ndEd., New York: Academic Press.
  • ––– (1997).Generative Linguistics: A HistoricalPerspective, London and New York: Routledge.
  • Newport, E.L. (1990). “Maturational constraints on languagelearning”,Cognitive Science, 14: 11-28.
  • ––– (2001). “Reduced input in the acquisition of signedlanguages: contributions to the study of creolization”,Language Creation and Language Change: Creolization, Diachrony andDevelopment, M. DeGraff (ed.), Cambridge: MIT Press,pp.161-78.
  • Newport, E., Gleitman, H. and Gleitman, L. (1977). “Mother,please, I'd rather do it myself: some effects and non-effects ofmaternal speech style”, inTalking to Children: LanguageInput and Acquisition, Snow, C. and Ferguson, C. (eds.), NewYork: Cambridge University Press, pp.109-50.
  • Newport, E. L., & Aslin, R. N. (2004). Learning at a distanceI. Statistical learning of non-adjacent dependencies.CognitivePsychology, 48(2), 127-162.
  • Nishimura, H., Hashikawa, K., Doi, K., Iwaki, T., Watanabe, Y.,Kusuoka, H., Nishimura, T. and Kubo, T. (1999).Nature, 397:116.
  • Norbury, C. F., D. V. M. Bishop, and Briscoe,J. (2002). “Does impaired grammatical comprehension provideevidence for an innate grammar module?”AppliedPsycholinguistics 23(2): 247-268.
  • O'Brien E.K., Zhang X.Y., Nishimura C., Tomblin J.B., Murray J.C.(2003). “Association of specific language impairment (SLI) to theregion of 7q31”,American Journal Of Human Genetics 72(6): 1536-1543
  • Odling-Smee, J., Laland, K. and Feldman, M. 1996, “NicheConstruction,”American Naturalist, 147:641-8.
  • Pastore, R.E., Li, X.F. and Layer, J.K. (1990). “Categoricalperception of nonspeech chirps and bleats”,PerceptPsychophys, 48(2):151-6.
  • Pena, M., Bonatti, L.L., Nespor, M., and Mehler, J. (2002).“Signal-driven computations in speech processing”,Science, 298: 604-7.
  • Pinker, S. (1989).Learnability and Cognition: The Acquisitionof Argument Structure, Cambridge, MA: Bradford/MIT Press.
  • ––– (1991). “Rules of Language.”Science,253:530-35.
  • ––– (1994).The Language Instinct, New York: HarperCollins.
  • ––– (1997).How the Mind Works, New York, Norton.
  • ––– and Bloom, P. (1990). “NaturalLanguage and Natural Selection,”Behavioral and BrainSciences, 13, pp.707-84.
  • Posner, M.I. and Raichle, M.E. (1994).Images of Mind. 2ndEd., New York: W H Freeman & Co.
  • Prinz, J.J. (2002).Furnishing the Mind: Concepts and theirPerceptual Basis, Bradford Books/MIT Press.
  • Pullum G.K. and Scholz, B.C. (2002). “Empirical assessment ofstimulus poverty arguments”,The Linguistic Review, 199-50.(Special issue, nos. 1-2: ‘A Review of “The Povertyof Stimulus Argument”,’ ed. N. Ritter).
  • Putnam, H. (1971). “The ‘Innateness Hypothesis’and explanatory models in linguistics”, inThe Philosophy ofLanguage, J. Searle (ed.), London: Oxford University Press,pp.130-9.
  • Quartz, S.R. and Sejnowski, T.J. (1997). “The Neural Basis ofCognitive Development: A Constructivist Manifesto,”Brainand Behavioral Sciences, 20:537-596.
  • Ramus, F., Hauser, M.D., Miller, C., Morris, D., Mehler, J. (2000).“Language Discrimination by Human Newborns and by Cotton-TopTamarin Monkeys”,Science, 288: 349-351.
  • Redington, M. and Chater, N. (1998). “Connectionist andstatistical approaches to language acquisition: A distributionalperspective”,Language and Cognitive Processes, 13:129-191.
  • Rice, M. L. and K. Wexler (1996). “Toward tense as aclinical marker of specific language impairment in English-speakingchildren.”Journal of Speech and Hearing Research39(6): 1239-1257.
  • Ritter, N.A. (2002). ‘A Review of “The Poverty ofStimulus Argument”,’The Linguistic Review(Special Issue), 19:1-2.
  • Roeper, T. and Williams, E., (eds.) (1987).Parameter Setting:Studies in Theoretical Psycholinguistics, Vol. 4, Dordrecht:Kluwer.
  • Sadato, N. Pascual-Leone, A., Grafman, J., Ibanez, V., Deiber,M.P., Dold, G. and Hallett, M. (1996). “Activation of the primaryvisual cortex by Braille reading in blind subjects”,Nature, 380: 526-8.
  • Saffran, E.M. (2000). “Aphasias and the relationship oflanguage and brain”,Seminars in Neurology, 20:409-418.
  • Saffran, J.R., Aslin, R.N. and Newport, E.L. (1996).“Statistical Learning by 8-Month-Old Infants”,Science, 274: 1926-28.
  • Saffran, J.R. (2002). “Constraints on statistical languagelearning”,Journal of Memory and Language, 47:172-96.
  • Saffran, J.R. and Thiessen, E.D. (2003). “Pattern inductionby infant language learners”,Developmental Psychology,40: 484-94.
  • Saffran, J. R., & Wilson, D. P. (2003). From syllables tosyntax: Multilevel statistical learning by 12-month-old infants.Infancy, 4(2), 273-284.
  • Samet, J. (1986). “Troubles with Fodor's Nativism”,Midwest Studies in Philosophy, X. Boston:Blackwells. 575-594.
  • Sampson, G. (1989). “Language Acquisition: Growth orLearning?”,Philosophical Papers, XVIII: 203-40.
  • ––– (2002). “Exploring the richness of thestimulus”,The Linguistic Review, 19: 73-104. (Specialissue, nos. 1-2: ‘A Review of “The Poverty of StimulusArgument”’, N. Ritter (ed.)).
  • Sapir, E. (1921).Language: An Introduction to the Study ofSpeech, Hillsdale, NJ: Erlbaum.
  • Saxton M. (1997). “The contrast theory of negativeinput”,Journal of Child Language, 24:139-61.
  • –––, Kulcsar B., Marshall G., and Rupra M.J.,1988, “Longer-term effects of corrective input: an experimentalapproach”,Child Language, 25: 701-21.
  • Scharff, C. and S. A. White (2004). Genetic components of vocallearning.Behavioral Neurobiology of Birdsong. 1016:325-347.
  • Scholz, B.C. and Pullum, G.K. (2002). “Searching forarguments to support linguistic nativism”,The LinguisticReview, 19: 185-224. (Special issue, nos. 1-2: ‘A Review of“The Poverty of Stimulus Argument”,’N. Ritter(ed.))
  • Schulte, O. (2006). “Formal Learning Theory”, TheStanford Encyclopedia of Philosophy (Spring 2006 Edition), EdwardN. Zalta (ed.), URL =<https://plato.stanford.edu/archives/spr2006/entries/learning-formal/>.
  • Sharwood Smith, M. (1994).Second Language Learning:Theoretical Foundations, London and New York: Longman.
  • Shimojo, S. and Shams, L. (2001). “Sensory modalities are notseparate modalities: plasticity and interactions”,CurrentOpinion in Neurobiology, 11: 505-9.
  • Shinohara, T. (1994). “Rich classes inferable from positivedata: length-bounded elementary formal systems”,Informationand Computation, 108: 175-186.
  • Singleton, J. L., & Newport, E. L. (2004). When learnerssurpass their models: The acquisition of American Sign Language frominconsistent input.Cognitive Psychology, 49(4), 370-407.
  • Skinner, B.F. (1957).Verbal Behavior, NY: PrenticeHall.
  • Skuse, D.H. (1993). “Extreme deprivation in earlychildhood”, inLanguage development in exceptionalcircumstances, D. Bishop and K. Mogford (eds.), Hillsdale, NJ:Erlbaum, pp. 29-46.
  • Skyrms, B. (1996).The Evolution of the Social Contract,Cambridge, Cambridge University Press.
  • ––– (2002).The Stag Hunt and the Evolutionof Social Structure, Cambridge, Cambridge University Press.
  • Smith, E.E. and Medin, D.L. (1981).Categories andConcepts, Cambridge, MA: Harvard University Press.
  • Smolensky, P. (1991). “Connectionism, Constituency and theLanguage of Thought”, inMeaning in Mind: Fodor and HisCritics, B. Loewer and G. Rey, (eds.), Cambridge MA: Blackwells,201-227.
  • Soames, S. (1984). “Linguistics and Psychology”,Linguistics and Philosophy, 7: 155-79.
  • Sokolov, J.L. and Snow, C.E. (1991). “The Premature Retreatto Nativism,”Behavioral and Brain Sciences,14:635-6.
  • Sterelny, K. (1989). “Fodor's Nativism”,Philosophical Studies, 55: 119-141.
  • ––– 2003,Thought in a Hostile World: The Evolution of HumanCognition, London, Blackwells.
  • Stich, S.P. (1971). “What every speaker knows”,Philosophical Review, 80: 476-96.
  • ––– (1978). “Beliefs and SubdoxasticStates,”Philosophy of Science, 45, pp.499-518.
  • Stiles, J. (2000). “Neural plasticity and cognitivedevelopment”,Developmental Neuroscience, 18 (2):237-272.
  • Stromswold, K. (2000). “The cognitive neuroscience oflanguage acquisition”, inThe New Cognitive Neurosciences,2ndEd M.S. Gazzaniga (ed.), Cambridge: MITPress, pp.269-80.
  • Takahashi K., Liu, F.C., Hirokawa, K., and Takahashi, H. (2003).“Expression of Foxp2, a gene involved in speech and language, inthe developing and adult striatum”,Journal of NeuroscienceResearch, 73 (1): 61-72.
  • Tallal, P. (1980). “Auditory Temporal Perception, Phonics,And Reading Disabilities In Children.”Brain AndLanguage 9(2): 182-198.
  • Tallal, P., R. E. Stark,et al. (1985). “Identification OfLanguage-Impaired Children On The Basis Of Rapid Perception AndProduction Skills.”Brain and Language 25(2):314-322.
  • Teramitsu, I., L. C. Kudo,et al. (2004). “Parallel FoxP1and FoxP2 expression in songbird and human brain predicts functionalinteraction.”Journal of Neuroscience 24(13):3152-3163.
  • Thiessen, E.D. and Saffran, J.R. (2003). “When cues collide:Use of stress and statistical cues to word boundaries by 7-to9-month-old infants”,Developmental Psychology, 39:706-16.
  • Thomas, M. and Karmiloff-Smith, A. (2002): “Aredevelopmental disorders like cases of adult brain damage? Implicationsfrom connectionist modeling.”Behavioral and BrainSciences, 25: 727—788
  • Tomasello, M. (1999).The Cultural Origins of HumanCognition, Cambridge, MA, Harvard University Press.
  • ––– (2000). “Two Hypotheses about PrimateCognition,” in C. Heyes and L. Huber (Eds.),Evolution ofCognition, Cambridge, MIT Press, pp.165-84.
  • ––– (2003).Constructing a Language: AUsage-Based Theory of Language Acquisition. Harvard UniversityPress.
  • vanderLely, H. K. J. and L. Stollwerck (1997). “Bindingtheory and grammatical specific language impairment inchildren.”Cognition 62(3): 245-290.
  • van der Lely, H. K. J. and V. Christian (2000). “Lexicalword formation in children with grammatical SLI: a grammar-specificversus an input-processing deficit?”Cognition 75(1):33-63.
  • van der Lely, H. K. J. and M. T. Ullman (2001). “Past tensemorphology in specifically language impaired and normally developingchildren.”Language and Cognitive Processes 16(2-3):177-217.
  • Vargha-Khadem, F., Carr, L.J., Isaacs, E., Brett, E., Adams,C. and Mishkin, M. (1997). “Onset of speech after lefthemipsherectomy in a nine-year-old boy”,Brain, 120:159-82.
  • –––, Watkins, K., Alcock, K., Fletcher, P. and Passingham, R.,1995, “Praxic and nonverbal cognitive deficits in a large familywith a genetically transmitted speech and language disorder,”Proc. Natl. Acad. Sci. USA, 92:930-33.
  • Volterra, V., O. Capirci,et al. (1996). “Linguisticabilities in Italian children with Williams syndrome.”Cortex 32(4): 663-677.
  • Watkins, Dronkers & Vargha-Khadem, ‘Behavioral analysisof an inherited speech and language disorder: comparison with acquiredaphasia.’Brain (2002), 125, 452-464.
  • White, L. (2003).Second Language Acquisition and UniversalGrammar, Cambridge Textbooks in Linguistics, Cambridge: CambridgeUniversity Press.
  • Wimsatt, W.C. (1999). “Generativity, Entrenchment, Evolutionand Innateness: Philosophy, evolutionary biology, and conceptualfoundations of science”, inWhere Biology Meets Psychology:Philosophical Essays, V. Hardcastle (ed.), Cambridge: BradfordBooks/MIT Press, pp.139-179.
  • Wright, A.A, Rivera, J.J., Hulse, S.H., Shyan, M. and Neiworth,J.J. (2000). “Music perception and octave generalization inrhesus monkeys”,Journal of Experimental Psychology:General, 129 (3): 291-307.
  • van Gelder, T. (1995). “What Might Cognition Be, If NotComputation?”,Journal of Philosophy, 92: 345-381.
  • Vargha-Khadem F., Watkins, K.E., Price, C.J., Ashburner, J.,Alcock, K.J., Connelly, A., Frackowiak, R.S., Friston, K.J., Pembrey,M.E., Mishkin, M., Gadian, D.G., and Passingham, R.E. (1998).“Neural basis of an inherited speech and languagedisorder”,Proceedings of the National Academy of Sciences,USA, 95 (21): 12695-12700.
  • Watkins, K.E., Dronkers, N.F., and Vargha-Khadem, F. (2002).“Behavioural analysis of an inherited speech and languagedisorder: comparison with acquired aphasia”,Brain,125: 452-464.
  • Zhang J.Z., Webb, D.M., and Podlaha, O. (2002). “Acceleratedprotein evolution and origins of human-specific features: FOXP2 as anexample”,Genetics, 162 (4): 1825-1835.

Acknowledgments

First, I thank two anonymous reviewers for theStanford Encyclopedia of Philosophy, whose pungent criticisms contributed to the clarity, breadth and depth (and also the length) of this article. I thank my friends, adversaries and colleagues (you know who you are) for their criticism and encouragement. Finally, and most fervently, I thank theeditor of theEncyclopedia, Ed Zalta, for patience and understanding that go way beyond the saintly.

Copyright © 2008 by
Fiona Cowie

Open access to the SEP is made possible by a world-wide funding initiative.
The Encyclopedia Now Needs Your Support
Please Read How You Can Help Keep the Encyclopedia Free

Browse

About

Support SEP

Mirror Sites

View this site from another server:

USA (Main Site)Philosophy, Stanford University

The Stanford Encyclopedia of Philosophy iscopyright © 2023 byThe Metaphysics Research Lab, Department of Philosophy, Stanford University

Library of Congress Catalog Data: ISSN 1095-5054


[8]ページ先頭

©2009-2025 Movatter.jp