The relation between mind and language: The Innateness Hypothesis and the Poverty of the Stimulus EDOARDO LOMBARDI VALLAURI Abstract This article tries to show some specifically linguistic weak points in the Povertyof-the-Stimulus Argument (PSA). Besides some quantitative considerations, from a qualitative point of view it is shown that the innatist tradition underestimates analogy as a resource for children to build their own grammars from the incomplete stimuli they receive from the environment; that knowledge and consciousness of reality surrounding the speech acts are also underestimated, and in fact play a major role in allowing children to build their internal grammars; that the role of “negative information”, conceived as the fact that some structures simply (but systematically) do not occur in the stimulus, is also underestimated. It is also suggested that the high degree of convergence of all known grammars does not need to be explained by means of one grammar in the brain, but simply results from a series of constraints that are pragmatic in nature, or directly derive from the definition of a system designed for communication. 1. Foreword This contribution may be viewed as belonging to what has been called the “biolinguistic approach”,1 in that it will consider language as a natural object, more precisely, a component and a product of the human mind/brain. By adopting what may be termed a “functionalist” perspective, we will choose one of the many possible levels at which the working of the brain may be analyzed, namely the “language” level, remaining relatively autonomous from the study of the physical mechanisms involved. In other words, we will concentrate on a major output of the brain, i.e. the linguistic output. To use a metaphor, we 1. See Jenkins (2000). The Linguistic Review 21 (2004), 345–387 0167–6318/04/021-0345 c Walter de Gruyter 346 Edoardo Lombardi Vallauri will deal mainly with the software, not with the computing machine. I believe that, as linguists, this is a legitimate perspective which may also turn out to be very useful because it is precisely on the strictly linguistic level that more or less recent literature has established some “truths”, at times highly debatable, at others needing further examination. More specifically, I think it is still necessary to clarify the consistency of an argument which, however, still represents one of the mainstays of the innatist conception of language, namely the so-called Poverty-of-the-Stimulus Argument (PSA). 2. The logical problem of language acquisition Cecchetto and Rizzi (2000: 119) summarize the problem of the poverty of the stimulus as follows: Humans acquire a natural language early in life, without specific instruction, apparently in a non-intentional manner, with limited individual variation in spite of the fragmentary and individually variable courses of experience which ground individual knowledge of language. More importantly, the precise understanding of fragments of the adult knowledge of language reveals the massive presence of “poverty of stimulus” situations: our adult knowledge of language is largely underdetermined by the data available in childhood, which would be consistent with innumerable generalizations over and above the ones that speakers seem to unerringly converge to. This empirical observation is of great importance, as it grounds the necessity of postulating a structured system of predetermined linguistic principles which guide language acquisition; it also leads to the expectation of a fundamental cross-linguistic uniformity of human languages. Literature on the matter is overwhelming, and clearly Noam Chomsky must be credited with the merit of having raised a question of such importance. As is well known, Chomsky’s hypothesis is that the brain somehow “contains” an inborn Universal Grammar (UG), also working as a Language Acquisition Device (LAD), which makes it possible for any child to acquire any language within the first three/five years of life. Crucially, this UG is conceived of strictly as a grammar, that is to say, a set of specifically linguistic rules or properties, and not as an operating mental/cerebral module for more general activities of problem solving, analogical extension, abstraction, recursive use of structures and similar activities, which may be devoted, beside language, to many other kinds of tasks related to life. According to Chomsky, and to generative linguistics, the biological bases of the language faculty are not limited to such features of the brain and the mind that can serve language among other human faculties. They also include some features so specifically and uniquely linguistic in nature, that they really do have the appearance of a grammar. The relation between mind and language 347 As correctly stated by Putnam (1967: 14): The Innateness Hypothesis is supposed [by Chomsky] to justify the claim that what the linguist provides is “a hypothesis about the innate intellectual equipment that a child brings to bear in language learning” (Chomsky 1962: 530). Of course, even if language is wholly learned, it is still true that linguistics “characterizes the linguistic abilities of the mature speaker” (Chomsky 1962: 530), and that a grammar “could properly be called an explanatory model of the linguistic intuition of the native speaker” (Chomsky 1962: 533). However, one could with equal truth say that a driver’s manual “characterizes the car-driving abilities of the mature driver” and that a calculus text provides “an explanatory model of the calculus-intuitions of the mathematician”. Clearly, it is the idea that these abilities and these intuitions are close to the human essence, so to speak, that gives linguistics its ‘sex appeal’, for Chomsky at least. In practice, the innateness hypothesis takes generative linguistics and, as far as it proposes itself as a candidate for the “office” of UG, promotes it from a simple description of language(s) to a theory of how an important part of the brain is structured, or at least of how it works. From now on, terms such as “innatism” or “innatist thesis” will be used here to refer to the position which, beside the specific structure of speech organs (larynx, pharynx, mouth etc.), beside the ability to deal with deduction, abstraction, recursive processes and similar, includes among the biological bases of language a true grammar, considered to be resident in the brain prior to any linguistic experience and to be perfectly identical in all individuals belonging to the homo sapiens (sapiens) species. This hypothesis is challenged, obviously and also authoritatively, from many sides. As stated above, here we will stay on a linguistic-functional level, and we will not dwell extensively on a variety of different critical approaches that have been proposed in the last four decades. We will only recall some of them, almost telegraphically. 3. A short overview of non-linguistic criticisms It would be interesting, for instance, to consider in detail the neurological arguments developed by Philip Lieberman (1991: 130–133), regarding the complete absence of cases where UG is disturbed as such: there seem to be no disturbed subjects in the Structure Dependency Principle, or in the Head Parameter, as a daltonian is disturbed in color recognition, or the like.2 This does not raise 2. Actually, theses such as this, held by Lieberman and others, are still under discussion, and attempts are under way to demonstrate the existence of so-called specific language impairment, which we cannot dwell upon here. 348 Edoardo Lombardi Vallauri any problem if we imagine that grammar is run by some general multi-purpose modules of the brain, so that if one of them breaks down, a similar one can replace it; but it is suspect if we hypothesize that specific portions of the brain are designed for specifically grammatical functions. When one of them (say the one in charge of the Head Parameter) goes wrong, nothing should be able to take its place. Neither can we discuss here the fact that language acquisition passes through successive stages of one-word, then two-word, then three-word utterances, which hardly seems compatible with the idea of a ready made UG in the child’s brain from the beginning. The innatist “defense” says that this is compatible with the full pre-existence of grammar, because the acquisition of grammar has already taken place, so to say, at birth, but its development is progressive, having to wait for the maturation of non specifically linguistic faculties such as short term memory and, more generally, the “channel capacity.”3 In any case, these claims still lack full experimental support, so that at present they are neither verifiable nor falsifiable. We do not consider here the more specific possibility, proposed by some innatists, that UG as such may mature after a while (for instance, there may first be a semantic stage, followed by a syntactic one).4 What motivates this hypothesis (which is sometimes extended only to parts of UG; for instance by supposing different development times for Principle A and Principle B of Binding Theory5 ) is the fact that the linguistic behaviour of children at early stages does not follow the proposed UG, and systematically violates it. But of course this is a very dangerous admission: if “internal” grammar grows over time, this growth may still be entirely ruled by innate instructions like the growth of a milk-tooth; but it becomes much easier to think that it happens only through exposure to language, being the product of interaction between general problem solving abilities and the environment. In this case, grammar would not be innate. We must also leave out of the present work the consideration of the paleontological, genetic and evolutionary evidence on the origins of language. We can only observe what follows, developing an idea of Putnam (1967): if linguistic ability presupposes innate UG and LAD, within an evolutionary framework it can only have imposed itself by way of selection in a human individual where a happy mutation has taken place (or reached completeness); then it should have 3. See, for instance, Chomsky (1980: 53) and (1981b); Hyams (1986); Cook and Newson (1996: 120–121). And for critical discussion, Tomasello (2000b: 229–232). 4. See Cook and Newson (1996: 147 and 331–333); Gleitman (1982); Chomsky (1986: 54). 5. See Cook and Newson (1996: 337–338 and 340); Montalbetti and Wexler (1985); Wexler and Chien (1985): Principle B would be already present in a very initial phase, while Principle A would develop between 21/2 and 61/2 years of age. The relation between mind and language 349 been handed down from that individual to his/her descendants, up to the complete colonization of the world. In particular, it cannot have been transmitted by learning to any individual who is not a direct descendant of the first one where the useful mutation appeared. Consequently, all speakers are descendants of a linguistic Adam or Eve, and all languages are necessarily offsprings of the protolanguage he/she spoke. In other words, the whole of mankind must descend from one single man or woman in whose brain UG developed by mutation. We are forced to believe that all those who, on the contrary, did not have the mutation died without descendants, leaving no trace. Even the children of the linguistic Adam/Eve who had not inherited UG (taking more from their other parent) must have died without descendants. Otherwise, there would still be some human beings around, absolutely normal but unable to acquire any language. In short, the monogenesis of languages hypothesis is a necessary consequence of the innatist idea, at least in a classic evolutionary framework. This raises no great difficulty, because the hypothesis is increasingly confirmed by genetic research à la Cavalli Sforza, and more generally by what is emerging about the precisely located and numerically restricted origins of the human species. However, if we accept the monogenesis hypothesis, there is no further need for an innate UG to explain the existence of linguistic universals. Those which are not explained on pragmatic grounds6 can always be seen as remnants of the protolanguage, in which case, nothing prevents it and its offspring from being acquired and not innate. Besides, as observed by Putnam, the pragmatic factor might have favored the preservation of certain original features as being particularly useful. This might be the case, for instance, for the existence of nouns to designate classes of concrete and abstract entities, verbs to designate states and processes, proper nouns to designate single well-known entities without having recourse to complex identifying descriptions, etc. From an evolutionary point of view, one more objection should perhaps be raised, briefly: if UG is an anatomic, physiological configuration of the human brain that arose by mutation, why shouldn’t it admit different “versions”? Human beings are not all genetically identical in what is determined genetically: there are specimens with light- and dark-colored skin, with fair, black, brown, red hair, with hook and pug noses, with blue, brown, hazel eyes, not to mention the other innumerable differences between races and between individuals. Now, if UG is genetically determined, why do we not have individuals or populations with obligatory subjects, or even, within the same linguistic community, groups of speakers with basic SVO order and groups with basic SOV 6. In fact, the explanations one can invoke for linguistic universals are so numerous that the existence of a UG does not appear necessary in this respect. On the matter, see, for instance, the articles in Hawkins (1992), and Lombardi Vallauri (1999). 350 Edoardo Lombardi Vallauri order? Why should all observable variation be external to what is genetically determined? Why should UG, unique among the genetic features of man, be exactly the same in all human specimens, without exception and without any synchronic or diachronic variation? In this sense I think it useful to read an important statement by Medawar (1983: 109), also cited by Moro (2002): One of the gravest and most widespread aberrations of geneticism is embodied in the belief that if any characteristic is enjoyed by all individuals of the community, it must be genetically underwritten. Thus, if it should turn out that a certain basic linguistic form such as the Aristotelian subject/predicate form is an element of all languages of the world, then its usage must be genetically programmed. [. . .] It may be well to repeat in this context the reason why this supreme canon of geneticism is not satisfactory: if any trait is to be judged “inborn” or genetically programmed, then there must be some people who lack it. The ability to taste phenylthiocarbamide, for instance, is known to be genetically programmed because there are those who lack it. Moro (2002) correctly concludes that, in order to study language as a genetically programmed phenomenon, it is necessary that between the grammars of the speakers of a language one can find differences that are dichotomic and systematic in nature, and that show a statistical distribution on family basis that is typical of bona fide genetically determined phenomena. I think that this prerequisite is necessary to even hypothesize that grammar is a genetically programmed faculty. Finally, does the generative idea that principles are innate and what we acquire is the setting of the parameters (see Section 4.1) have some analogy with the ontogenesis of human or animal individuals? Is there, in the growth of arms or hair, something that is left to the setting of a parameter? Does environmental learning intervene to make a biological parameter prevail over another?7 And even if this is the case, why does the setting of parameters, if it is present in any aspect of individual growth, always happen on a genetic basis (so that an individual can, on genetic grounds, grow differently from another close by), while, on the contrary, only in language acquisition the parameters should be set by the environment, thus determining that perfect sameness between individuals from the same environment, typical of what is acquired and not biologically predetermined? We will not develop here the extra-linguistic considerations mentioned above, but even on a strictly linguistic level there might be some perplexity over the facts commonly mentioned in support of the innateness of grammar, 7. In fact, this opinion is put forward by Boncinelli (1999: 221–224). The relation between mind and language 351 and particularly those usually referred to under the label of the Poverty-of-theStimulus Argument. This is what we will deal with in the following sections. 4. Linguistic evidence against the poverty of the stimulus My considerations here will not be based on new empirical data. I will rather focus on previously published data and, indeed, especially on those (kinds of) data most frequently mentioned by advocates of the PSA, from which I think quite different conclusions may be reached. Of course it will not be possible to discuss here the overwhelming quantity of arguments proposed within the framework of the PSA, but it may be useful to analyze some of them, since the same kind of confutation may possibly be extended to further examples. Therefore, I will propose some “central” examples from the propaganda on the matter, such as those presented in the works of Chomsky and in the periodical handbooks of the generative tradition. Chomsky (1987), referring to language acquisition, and giving the question the evocative name of “Plato’s Problem”, poses the problem in the same way as Bertrand Russell: “How do we come to have such rich and specific knowledge, or such intricate systems of belief and understanding, when the evidence available to us is so meagre?”. In other words, the time and linguistic material available to children during language acquisition are regarded as insufficient to explain their rapid mastering of such a complex instrument. The argument can be divided into two parts, that we may call “quantitative” and “qualitative”. There are, actually, two reasons why language should be “unlearnable”: (1) the amount of information that makes up language may be too big, as compared to the time it needs to be acquired; (2) the kind of information a language is made of is neither to be found in the stimulus (i.e. the linguistic and metalinguistic material the child is exposed to), nor deducible from it. In Cook’s (1991: 103) formulation: “if someone knows an aspect of language that is unlearnable from the environment, then it must be built-in to the mind.” This would hold even for a minimal amount of information: if children know something about language, even a very small thing, that they cannot have learned from the environment, then that thing, at least, was already in their brain at birth. We will deal with the two perspectives separately. 4.1. The amount of information which makes up language We cannot ascertain exhaustively whether the time available to the child for L1 acquisition is too little, but I agree with Putnam (1967) that it is not as little as usually claimed. 352 Edoardo Lombardi Vallauri A high school student who learns a language in, say, five years8 acquires it less successfully than a three-year-old child as regards pronunciation, but this can be explained by previously acquired “bad habits”. In fact, the pronunciation of L2-sounds that are the same as L1-sounds is usually perfect, and imperfections essentially arise where different habits interfere. Then, one must take into account the particular “plasticity” of the brain in children,9 and the progressive reduction of its learning capacity, that are completely independent from the innateness hypothesis on language, and that can be observed also in other fields, such as piano playing or downhill skiing. As for the lexicon, the L2-vocabulary of a graduating student is usually more extensive than the L1-vocabulary of a three-year-old child (also because the adult’s L2 lexicon “relies upon” that of the L1). Finally, as for grammar (morphology and syntax), the frequency of errors is not dramatically different. Now, how many hours of exposure to L2 can a graduating student rely on? With an average of two hours a week (taking into account flu, ski week and simple truancy) for 30 weeks, we have about 60 hours per year, that is to say 300 hours in five years. It must be pointed out that these hours are not consecutive, are characterized by little or no contact with native speakers, and are disturbed by an “existential climate” that makes emotional and intellectual engagement much stronger in other directions. On the contrary, children who acquire L1, at a rate of (about) 12 hours a day for 52 weeks per year, reach an equal number of hours in 25 days – less than a month. Is it so incredible that in a period 40 times longer (about three years, comparable to 200 school years) they acquire their language?10 I wouldn’t say so, especially if we consider that learning to talk is probably, at the time, the main aim of the child’s life, in contrast with the adult student, who has other preoccupations. Facing more specifically the problem of the great amount of things to be learned that language represents, the question arises: What is innate? Syntax? Or, rather, phonology, morphology, syntax and semantics? Chomsky (1988: 26 and 32) answers that beside syntax, UG includes “fixed principles governing possible sound systems”, and “a rich and invariant conceptual system, which is prior to any experience”. Now, one can wonder whether these language modules are all equally extensive and difficult to acquire, or whether some of them, so to say, “take less space in memory”. From the viewpoint of the amount of information to be acquired, phonology seems to be of small import, and there 8. This has been the time devoted to foreign languages in Italian schools, for many years. 9. A remarkable argument in this perspective comes from so-called “neural darwinism”, for which see, for instance, Edelman (1992). 10. If we consider a period of five years, required by the average child to acquire the language without errors, the corresponding school years would be more that 350. The relation between mind and language 353 is definitely no need for it to be innate in order to be acquired. Morphology is slightly more cumbersome. Even syntax, if conceived in the most restrictive sense, i.e., completely excluding any contribution of semantics, as the bare inventory of possible (recursively combinable) constructions, is not too great a load; but it certainly becomes much more bulky if we conceive it as the task of learning which meanings can be expressed by each construction, and which constructions can express each meaning. This is due to the virtually infinite number of meanings, and to the very high number of lexical entries, each of which is associated in quite idiosyncratic ways to certain syntactic patterns and not others. One is tempted to say that UG, if it must settle the problem of language acquisition, should be an Innate Universal Lexicon and Innate Universal Semantics, rather than an Innate Grammar stricto sensu. Only on these conditions would it help to explain the very rapid development of the complex human understanding of reality, much more complex than that of all other animals, even the most evolved. But it must be noted that we learn this complex semantic organization of reality throughout life, or, at least, unlike syntax, its acquistion is no way near complete after the famous first three (or five) years of life that allow the maturation of phonology, almost all of morphology and major parts of syntax in its restricted sense. From its beginnings (when the idea of innate UG was first proposed) to the present formulations, generative linguistics has undergone significant changes that are absolutely far from irrelevant to the problem under discussion. In particular, there has been a shift of the “seat” of grammatical rules, which, from an initial autonomous condition, are viewed increasingly as associated with lexical entries. In some present formulations of the theory, the majority of syntactic rules “link up” with the lexicon. Chomsky (1982: 8) states that “a large part of ‘language learning’ is a matter of determining from presented data the elements of the lexicon and their properties.” Chomsky (1991: 419) confirms that “language acquisition is in essence a matter of determining lexical idiosyncrasies”. Wexler and Manzini (1987) hypothesize that all parameters belong to the lexicon, rather than to the principles. Chomsky (1986: 150) is even more explicit: What we “know innately” are the principles of the various subsystems of S0 11 and the manner of their interaction, and the parameters associated with these principles. What we learn are the values of the parameters and the elements of the periphery (along with the lexicon to which similar considerations apply). The “elements of the periphery” also include things of a certain mnemonic and acquisitional “weight”, such as verb conjugations, and similar properties. If 11. S0 is the “initial state”, when the mind already contains UG but no linguistic experience. See Cook and Newson (1996: 78). 354 Edoardo Lombardi Vallauri one adds to this the specific values adopted by the parameters, and the whole lexicon, the bulk of things to be learned is so great that UG seems of little account in the comparison. This corresponds to the intuitive notion we all have of different languages, whereby what they all have in common is “much less” than what differentiates them. And obviously, UG cannot extend itself beyond what languages have in common. In short, by recognizing that there are few principles, that the parametric values they can assume are more numerous, and that in any case (whether we call them principles and parameters or not) the majority of the rules are incorporated in the lexicon, generative linguistics in its recent versions risks undermining the idea of an innate UG. What is innate would represent such a small amount of information that its influence in facilitating language acquisition would be little more than nil: The PSA would become irrelevant, or even totally obsolete. Independently from the problem of where grammatical rules reside (i.e., of their possibly being associated to parts of the language that are acquired “in the long run” like the lexicon) the last great turn of generative linguistics, minimalism, also moves towards a drastic simplification of the grammatical rules. This is obviously bound to produce a more elegant theory of UG, characterized by lesser complexity and encumbrance, and certainly welcome after the “orgy” of proliferating explanations and theories that has characterized the recent past. However, one side-effect of this reduction is the parallel scaling down of the role UG will play as a LAD and as an answer to the problem of the poverty of the stimulus. If UG comprises very few essential rules that are declined in very different ways across the different languages, there is no need to postulate its innate preexistence because acquisition will present more or less the same difficulties with or without an innate UG. Even without going as far as minimalism, it may be useful to examine an example put forward in an important generative grammar handbook based on the “Principles and Parameters” stage of the theory.12 The authors try to make explicit what, in a linguistic utterance, is interpreted by the child as a result of innate knowledge, and what is understood owing to acquired knowledge: Native speakers of English know that (7) His father plays tennis with him in the summer. 12. Cook and Newson (1996: 88). By the way, the Principles and Parameters stage of the theory is, by far, the most suitable to present generative grammar as an innate UG. The relation between mind and language 355 is a possible sentence. What have they acquired? First they have set the head parameter to head-first, so that play appears to the left of tennis, in to the left of summer. This setting was presumably acquired by noticing that English phrases have a head-first tendency. They also know that the Verb play can be followed by an object NP. This comes from the built-in Projection Principle and from the lexical entry for play, which is derived from observation of sentences. They also know that an actual lexical subject – his father – must be included; they have set the value of the pro-drop parameter to non-pro-drop. The fact that him must refer to someone other than his father derives from their built-in knowledge of Binding Principles, together with the specific lexical knowledge that him is a pronominal. Considering that the specific setting of the parameters is still something that must be learned, if one examines the knowledge necessary to process a single utterance from this perspective, the impression is that about half may be innate, and half acquired. It is not important to be much more precise; even if it were two thirds against one third, things would not change significantly from our point of view because what is lacking here is the extension to the whole language activity of the individual. In fact, if we added another thousand examples to (7), we would see that the innate things remain almost identical; there is some principle whose parameter must be set. On the contrary, we would be forced to add thousands of acquired notions, related mainly to the syntactic behavior of the single lexical entries, their autonomous (paradigmatic) semantic values and the (syntagmatic) meanings they assume when they establish relations with other elements of the lexicon. (Imagine adding sentences like his father reads the newspaper with him on the sofa, or my girl-friend played the devil with me in the past, and so on.) In other words, according to the criteria proposed in the quotation above, the innate knowledge required to process a thousand sentences from the language may be very small, while the language specific knowledge (different from language to language), to be acquired may be very large. The more we increase the number of utterances under consideration, the more we may expect the disproportion to grow. This can be regarded as an effect of that highly praised feature of human language(s) which is recursivity and is often cited as the very feature that differenciates us from other animals. Through it, we can produce an infinite number of utterances, and even an infinite number of slightly different syntactic constructions, continuously re-using a much smaller number of basic universal structures. Thus, we do not need to know all universal syntactic constructions (if there are any) innately, but just the basic ones that are combined to obtain the others. Then, of course, thousands of idiosyncratic rules associated to the single lexical entries and to the different languages must be added, and these we must obviously learn. Here I do not try to answer the question whether an innate UG actually exists or not; this would be beyond anyone. Still, it is worth underlining that among 356 Edoardo Lombardi Vallauri the arguments in favor of its existence, its absolute necessity in order to explain language acquisition in children is still open to debate. An innate UG may exist after all, but if it is the one outlined, it would be of very little help in language acquisition. Its existence is by no means proved by the PSA and must be proved in some other way, which will probably be more difficult. The way to prove it, of course, would be to demonstrate that it is truly impossible to learn the principles of UG through social interaction, i.e., by direct exposure to samples of utterances (no matter how numerous), so that, if these principles really do exist in the mind(s) of the speaker(s), they must necessarily be innate. But this is also difficult because it must first be demonstrated that the principles of UG really exist and are the same in all speakers of all languages. That is to say that all languages really work like generative grammar or, in other words, that there exist some (and even many) universals of language that are grammatical and ultimately not pragmatic in nature. To take one of the most favorable examples, this would mean demonstrating that the consistency observable in the head-complement order in all languages is not only preferred on general grounds of processing economy, but is rigidly determined by the structure of some module of the brain or, at least, of the mind. This demonstration is rendered extremely difficult (if not impossible) by the fact that many languages actually violate this consistency in many respects.13 It remains absolutely true that the amount of things children learn about their language is large, and language acquisition exceeds, in involvement and volume, the acquisition of any other thing in the first years of life, being, probably, the most inexplicable performance of the human mind. Still, that part of it which is language specific, and therefore acquired, is probably superior (and perhaps hugely so) in quantity to what is universal and therefore potentially innate. In this case the UG hypothesis does not explain very much, and the great merit of Chomsky remains that of having pointed out a problem, not that of having supplied its solution. It must also be considered that, when dealing with mind and brain, the simple fact that something appears extremely difficult or even inexplicable neither requires nor justifies a description that reduces it to something interpretable in terms of a modest computing machine. For instance, the way in which human consciousness can arise from the brain is absolutely mysterious and probably 13. In the light of these facts, what Cook and Newson (1996: 101) say lends itself to conclusions opposed to those intended: “Principles of UG are incapable of being learnt by social interaction. Whatever degree of importance one assigns to principles such as structure-dependency or Binding, it is clear they are not learnable through routines, however elaborate.” So, if they do not even serve to explain acquisition, these principles risk having no justification at all and not existing in the mind of every speaker, but only in the mind of some linguists. The relation between mind and language 357 does not admit a simple explanation (such as an Innate Universal Consciousness made up of a set of rules); rather, in the end it may well reveal itself as the result of an unimaginable complexity. We are presently aware of the existence of this complexity, but we ignore the way it works, and when it is known, it will probably not present itself in a simple form. A similar situation may hold for language (which is inseparably bound to consciousness14 ), and there is no reason for thinking that we must expect a simple solution, or that a simple solution will be more true than a complex one. In particular, it must be observed that what we are able to do unconsciously and automatically usually exceeds by far in complexity and difficulty what we can do consciously. For example, walking on an uneven ground is much more difficult than doing three-digit divisions, in that it requires control of a frighteningly greater number of subjective and objective elements in the time unit. Confirmation of this is available if one tries to build a robot that does three-digit divisions, and a robot that walks on uneven ground. The fact that we learn to walk long before learning to do divisions (and before learning to talk!) is no authorization for believing that the mechanisms involved in foot movement on irregular ground must be described in a simple way, elevating a reassuringly small set of rules to the standing of a theory of how the brain manages deambulation. Similar considerations may be made for the capacity to recognize the meaning of facial expressions in other humans, which allows us to say whether someone (even someone whose face we have never seen) is happy, angry, scared and so on, with thousands of nuances, starting from proxemic features that are quite complex and difficult to translate into precise rules. In other words, if there is something we do customarily and somewhat unconsciously, having learned it from childhood, if this something is extremely complex and does not appear to be immediately describable in terms of “simple rules or principles” understandable by our capacity for conscious elaboration, this does not mean that such simple principles actually exist somewhere and are simply waiting to be discovered (for instance, an innate UG of foot movements). On the contrary, it may mean that no shortcuts will be found, because behind such performances there are unconscious mechanisms whose complexity is immense or, at least, great enough to exceed the elaboration capacities of our consciousness. In many respects, language activity may belong to this kind of automatism. Why then, in principle, should it be an exception? We can admit that there is an innate predisposition to language as there is a predisposition to walking, although they are different in that walking develops itself, even without environmental stimulus. For this reason, language seems more comparable 14. See Carruthers (2002) for an overview of opinions on the matter. 358 Edoardo Lombardi Vallauri to those activities one acquires when surrounded by people who carry them out (like football in Italy or skiing in Switzerland). In any case, it does not appear necessary to see it as a simple set of rules and a set of specifically linguistic rules.15 As we have seen, since UG must be suitable for all languages, in all its formulations it necessarily tends to specify very few essential things, leaving the bulk of the matter to interlinguistic variation (i.e., to the acquired, non-innate “portion” of the language faculty). Still, sometimes the opposite effect arises: In order to have a UG rule describe the behavior of any language, theorists are obliged to imagine that each child’s head contains (innate) formalisms that are much more complex than those required to learn each single language. This is the case, for example, of the rule elaborated by Wexler and Manzini (1987)16 to describe the reference of reflexive elements: The Governing Category Parameter β is the governing category of α iff β is the minimal category which contains α and a governor for α and: (a) (b) (c) (d) (e) a subject; or an inflection; or a tense; or an indicative tense; or a root tense. Of this macro-rule, English speakers use (a), Italians (a) and (b), Norwegians up to (c), Icelanders up to (d) and Japanese up to (e). In other words, a Japanese 15. As pointed out to me by Daniele Portolan, the analogy between linguistic activity and other automatisms would need more space. I will just point out here that throughout life we learn with difficulty many tasks that we can repeat later without reanalyzing them every time. Once a series of correlated tasks has become automatic, we reproduce them by simply “setting off” a routine now perceived as a blurred whole, and which always “produces itself” (with no need of consciousness) in the same coherent way, though admitting concomitant and, so to say, “suprasegmental” variations. For example, the same series of notes on a piano admits variations of length, intensity, pressure, etc. A backhand stroke in tennis can be varied each time in intensity, rotation, angle, although always involving the same synchronization of legs, shoulder, waist, wrist, head etc. Much the same holds for tactical movements made by players in team sports, operations carried out by assembly-line workers, etc. I cannot play piano, and I typewrite very badly. Now, there are some very frequent sequences of letters in Italian (such as z-i-o-n-e) that I have learned to type in a conscious way but now execute automatically, having no idea of how I move my fingers. When, for example, I write the word situazione, I “think” through the first part of the word letter by letter, but, as soon as I have typed the a, I “set off” an automatic and unanalyzable routine, exclusively motory in nature, by which my fingers, completely unaware of letters and alphabet, quickly type the rest. Crucially, unlike my typewriting, such automatisms can be extremely complex in amount of information one has to process in order to control them. 16. See Cook and Newson (1996: 304). The relation between mind and language 359 child uses it all, but if it is innate, this rule ends up being more or less useless in the heads of children who acquire different languages. In their case, acquisition without UG appears to be simpler than with it. If the rule is built-in and works, non Japanese, say English, children must not only induce, from the language they hear, that the minimal category containing α must be a subject; but they must also go against the rule and exclude that any of the other restrictions from (b) to (e) is valid. For this reason, it is necessary to imagine another mechanism, called the Subset Principle,17 through which children can set the parameter to the most restrictive value allowed by the language samples they experience. We will return to this question when dealing with the qualitative approach to the problem of the poverty of the stimulus, but here it is necessary to make an important remark. The example above is particularly manifest; but, at least in the Principles and Parameters Theory, UG always poses itself as an uneconomical explanation for language acquisition, in that it implies that every mind contains a lot of information that must be discarded. To be able to learn any language, the brain would need to contain the values for the parameters required by all languages, although in the end it only selects one for each principle. This situation can be considered as facilitating rather than hindering acquisition, only if one starts from a hypothesis which is not to be dismissed, but must be made explicit. The hypothesis is that the brain has unlimited capacity if conceived as a hard disk where the data are stored, while being quite weak if conceived as a processor that has to manage, and particularly learn them. If this were true, there would be no problem with the fact that all the parameters are stored in the brain, despite the majority being useless, and their preexistence on the “hard disk” as preset variables could actually facilitate a weak processor that would hardly be able to retrieve them from experience alone. The fact that the innate UG hypothesis is based on this idea of the mind as a machine with an unlimited hard disk (storage capacity) and a weak CPU (processing efficiency) is important because it renders the theory vulnerable through possible alternative hypotheses, as we will see later on. In this section we have dealt with the quantitative aspect of the povertyof-the-stimulus problem, and we finally suggested that, for many reasons, the amount of information in UG may be quite reduced when compared to the overall amount of what a child learns. It should be emphasized that if one adopts this perspective, we have an interesting inversion of a very important (at least in linguistics) onus probandi. On the basis of the PSA, supporters of UG, 17. The so-called Subset Principle, initially proposed by Dell (1981) and then formalized by Berwick and Weinberg (1984) and Berwick (1985), is supposed to oblige children to always generate the most restrictive grammar (the one which generates the lowest number of structures) compatible with the data they are exposed to, and extend it only if compelled by new data. For a short but effective criticism of such assumptions, see Bowerman (1988: 77–78). 360 Edoardo Lombardi Vallauri and more particularly of generative grammar as a hypothesis for UG, can say, “Language acquisition poses a problem. Innate UG is a solution. Generative grammar proposes itself as a model of UG, therefore it is up to those who don’t agree with that model to find a better one in order to explain language acquisition.” But if we do not admit that UG is a good explanation for the problem of the poverty of the stimulus, or even, more radically, if we think that such a problem does not pose the necessity (in a philosophical sense) of an explanation in simple and formal terms, then the burden of proof is inverted and those who want to maintain that an innate UG exists must prove it by means of independent evidence. In other words, it must be proved (a) that the proposed grammar is actually universal and common to all languages (a very difficult task, not significantly addressed up to now, in spite of the many decades of effort within a great part of formal linguistics); (b) that Principles and Parameters (or, in any case, the rules of language) that the speaker is viewed as being endowed with cannot be learned by induction. More precisely, therefore, the arguments for non-inducibility of the rules from the real utterances which the child is exposed to must be reviewed with renewed severity. In the next section, we will move in this direction, in order to show how examples of linguistic stimuli usually cited to illustrate that the child could never induce the rules from them are in fact open to a different interpretation. 4.2. The kind of information which makes up language 4.2.1. Analogy and knowledge of reality. Bloomfield (1933) maintained that a regular analogy permits the speaker to utter speech-forms which he has not heard; we say that he utters them on the analogy of similar forms which he has heard. Analogy allows speakers not only to create new forms by adopting the rules that hold for those already heard, but also to correctly understand new forms that have not yet been heard. In the Chomskyan tradition, confidence in the faculty of analogical extension has been dramatically reduced, and many linguistic utterances are reported in order to state that, since they are new and never heard, children would neither understand nor produce them if they had no innate linguistic knowledge of the matter. However, analogy (or, more generally, the various kinds of conceptual extensions we can gather under this label) is actually at work in all mental processes. If one only tries to suppose that it applies to the examples adduced by the supporters of the LAD, it is often possible to see that analogy may play a major role in their production and interpretation. The relation between mind and language 361 Chomsky (1988: 17–22)18 examines some sentences of a certain complexity, whose treatment by children and speakers in general he regards as inexplicable in terms of analogical extension from less complex sentences. For example, in (1) and (2) one would expect speakers to interpret the referent and the syntactic function of the clitic reflexive pronoun si by analogy from the simple utterance (3) (i.e., interpret the clitic as a direct object coreferent with the subject Gianni): (1) Gianni si fece radere i ragazzi. Gianni REFL made shave the boys ‘Gianni had the boys shave *themselves/*him(Gianni) /(themselves or by someone else) for him(Gianni) .’ (2) Chi si fece radere Gianni? Who REFL made shave Gianni ‘Whoi made Gianni shave *him(Gianni) /himselfi ?’ ‘Who made Gianni shave (himself or by someone else) for him(Gianni) ?’ (3) Giannii sii fece radere. Giannii REFLi made shave ‘Gianni had someone shave him(Gianni) .’ Indeed, (3) is the reflexive version of (4), that is to say, the one adopted when Gianni himself takes Pietro’s place, i.e. when the referent of the direct object coincides with that of the subject (Gianni fece radere Gianni, ‘Gianni had Gianni shaved’). (4) Gianni fece radere Pietro. Gianni made shave Pietro ‘Gianni had Pietro shaved.’ Chomsky himself agrees that this happens by analogy with the simpler, noncausative series (5a–c). The language acquirer can observe that when Gianni replaces Pietro in (5a) the result (5b) is systematically converted into (5c), where the reflexive pronoun is coreferent with the subject: (5) a. Gianni rade Pietro. Gianni shaves Pietro ‘Gianni shaves Pietro.’ 18. I refer to the examples given in the Italian edition cited in the References, as I will derive some further examples from these later, which will be more reliable if I use my mother tongue. Chomsky himself (p. 19 and 20) suggests that Italian and Spanish have similar behavior in this field, and (since UG is supposed to be valid for both languages) considerations similar to those we will make for Italian should be possible for the Spanish examples given by Chomsky in the original version. 362 Edoardo Lombardi Vallauri b. c. Gianni rade Gianni. Gianni shaves Gianni ↓ Giannii sii rade. Giannii REFLi shaves ‘Gianni shaves himself.’ Consequently, he can operate analogically on (4) as soon as its subject also takes on the function of direct object, as in (6): (6) Gianni fece radere Gianni. Gianni made shave Gianni ‘Gianni had Gianni shaved.’ ↓ (3) Giannii sii fece radere. Giannii REFLi made shave ‘Gianni had himself shaved.’ But Chomsky crucially claims that analogy is not extended further, being impossible to interpret the reflexive pronoun in (1) and (2) as a direct object coreferent with the subject Gianni. In fact, if the pronoun si in (1) has to be coreferent with Gianni, as in (3), then it will be a dative and not a direct object. As for (2), if ever the sentence were acceptable with si as a direct object, the referent of the reflexive object would not be Gianni, but the boys (Chii sii fece radere Giannij ?). Speakers know this and discard (1) and (2), at least in the interpretation where si is a direct object coreferent with Gianni. Thus, the usual question arises: “How does the child learning Spanish or Italian know such facts as these?” (1988: 20). Chomsky claims that “the examples reveal once again the hopelessness of an attempt to account for [. . .] the use of language in terms of analogy” (1988: 20), so that “analogy seems to be a useless concept, invoked simply as an expression of ignorance as to what the operative principles and processes really are” (1988: 24). Since the acquirer receives no direct information on the matter, such as, “You should not (analogically) extend the interpretation of (1) and (2) from that of (3),” then, Chomsky argues, we are obliged to hypothesize that the mechanisms that rule the reference of the reflexive pronoun pre-exist in the brain. However, it is possible to reason differently. First, it must be considered that the material available to children acquiring Italian is richer. Beside those cited by Chomsky, they also hear many utterances like the following:19 19. Utterance (11) is reported by Chomsky, but its potential usefulness for the analogical interpretation of (1) is not mentioned. The relation between mind and language (7) Gianni si fece radere il collo. Gianni REFL made shave the neck ‘Gianni had someone shave his(Gianni’s) neck.’ (8) Gianni si fece portare i ragazzi. Gianni REFL made bring the boys ‘Gianni had the boys brought to him.’ (9) Gianni si fece radere dai ragazzi. Gianni REFL made shave by-the boys ‘Gianni had the boys shave him.’ (10) Gianni si fece radere per i ragazzi. Gianni REFL made shave for the boys ‘Gianni had someone shave him for the boys.’ (11) Gianni fece radersi i ragazzi. Gianni made shave-REFL the boys ‘Gianni had the boys shave themselves.’ (12) I ragazzi si fecero radere. the boys REFL made shave ‘The boys had someone shave them.’ 363 From each of these, children learn something they can use analogically. From utterances like (7) and (8) they learn that when a nominal is added after the causative construction, the nominal itself becomes a direct object, while the clitic si can still refer to the subject, but only as an oblique case. This is more than sufficient to provide, by analogy, the only possible interpretation of (1). But even if this were not enough, children still have many available examples similar to (9) and (10), from which they learn that in order for the clitic si to remain a direct object when another nominal is added, the nominal itself must be marked in a different way (e.g., as an Agent, a Beneficiary, etc.). In order to exclude the clitic in (1) as coreferent with the object (i ragazzi), children can take utterances like (11) into account, which make it evident that in such cases the clitic is placed after the verb. Finally, even if it were not evident that in (1) i ragazzi is the direct object (the subject being Gianni), children can deduce this information from utterances like (12), which show that when something like i ragazzi is the subject, it requires morphological agreement on the verb. This analogical framework is reinforced by utterances with complex verbal forms, where the alternatives between clitic objects or obliques, between the clitic’s appearing before or after the verb, and between its coreferring with the causer or the causee are kept clearly separate and cannot lead to the same doubts as the simple verbal forms because they select different auxiliaries: 364 Edoardo Lombardi Vallauri (13) Gianni ha fatto radersii i ragazzii / *è fatto Gianni has made shave-REFLi the boysi *is made radersi i ragazzi. shave-REFL the boys ‘Gianni has had the boys shave themselves.’ (14) è fatto radere i ragazzi / *si ha Giannii sii Giannii REFLi is made shave the boys *REFL has fatto radere i ragazzi. made shave the boys ‘Gianni has had someone shave the boys for him.’ As for utterance (2), Chomsky claims that analogy should lead speakers to accept it as “for which person X, Gianni managed that X shaved Gianni?” based on utterances like (3) where the reflexive pronoun is coreferent with the subject (Gianni). But he ignores the fact that the interrogative version (transformation?) of (3), that speakers certainly have wide experience of, is actually (15): (3) fece radere. Giannii sii Giannii REFLi made shave ‘Gianni had someone shave him.’ ↓ (15) fece radere? Chii sii Whoi REFLi made shave ‘Who had someone shave him?’ On the contrary, (2) appears to be something quite different, with one more participant, and it is not plain to see why it should be interpreted analogically with (3). For the cases where a participant is added, speakers can actually draw a comparison with utterances like (16), where that participant is morphosyntactically marked in a different way, so that the clitic remains free to have direct object function and corefer with the subject: (16) fece radere Giannii ? Da chi sii by who REFLi made shave Giannii ‘Who had Gianni shave him(Gianni) ?’ It would be useful to render completely explicit the role played, through analogy, by the existence of many other types of utterances that belong to the linguistic material available to the child, such as the following, which we can only rapidly mention here. The relation between mind and language (17) Gianni fece cantare i ragazzi. Gianni made sing the boys ‘Gianni made the boys sing.’ (18) Chii fece cantare Giannij ? Giannij whoi made sing ‘Whoi made Giannij sing?’ (19) Chij fece cantare i ragazzii ? whoj made sing the boysi ‘Whoi made the boysj sing?’ 365 The correspondence between (17) and (18)–(19) makes it clear that the interrogative pronoun never refers to the nominal constituent of the assertive clause that appears also in the interrogative clause, but always to the constituent which no longer appears in full after the interrogative transformation. This is also evident in causative utterances like (20): (20) Chi fece radersi Gianni? who made shave-REFL Gianni ‘Who made Gianni shave himself(not Gianni) ?’ In order to interpret (20), utterances like the following, which we cannot dwell upon, may also be very useful: (21) Chi si è fatto portare Gianni? who REFL is made bring Gianni ‘Who(OBJ) did Gianni(SUBJ) have brought to him (Gianni) ?’/ ‘Who(SUBJ) did Gianni(OBJ) have brought to him (not Gianni) ?’ (22) Cosa si è fatto radere Gianni? what REFL is made shave Gianni ‘What did Gianni have someone shave for him(Gianni) ?’ (23) Cosa si è fatto cantare Gianni? what REFL is made sing Gianni ‘What did Gianni have someone sing for him(Gianni) ?’ (24) Chi ha fatto radersi Gianni? who has made shave-REFL Gianni ‘Who(SUBJ) made Gianni(OBJ) shave himself(Gianni) ?’/ ‘Who(OBJ) made Gianni(SUBJ) shave himself(not Gianni) ?’ In short, children and speakers in general can operate perfectly by analogy because analogy can take into consideration a far greater quantity and variety of information than that imagined by Chomsky. Moreover, analogy constantly works with significant support from the interpretation of reality. When they 366 Edoardo Lombardi Vallauri hear utterances like the ones above, children always have contexts at their disposal, which allow them to understand what the meanings of the utterances may be. For example, when someone produces an utterance like (1), the situation usually makes it clear that Gianni was interested in the boys’ shaving themselves. An utterance like (16), on the other hand, always appears in contexts where someone has shaved Gianni; and an utterance like (21) appears in contexts where someone had Gianni brought to him(someone) , or contexts where Gianni had someone brought to him(Gianni) ; and so on. Now, this reality that surrounds linguistic activity, is indeed extremely complex and, so far, cannot be formalized. Effectively it is Reality tout court, reality in general. It cannot be expressed by means of simple rules, despite fully existing; we have evidence of this and no onus probandi that it exists. The interpretation of the surrounding reality leads the child to understand something, in grammatical terms, along these lines: “In the utterance Gianni si fece radere i ragazzi, the clitic si does not refer to Gianni as Patient and Object, because there is already i ragazzi as Patient and Object; if the boys were not Patient but Agent, one would say dai ragazzi (= by the boys)”. The experience of every child supports such knowledge through thousands of examples. Reasoning as if language were acquired in total isolation from reality is very frequent in the literature, and lies at the very foundation of beliefs about innate UG. For example, according to Morgan (1986) English sentences like (25) The dog bites the cat. do not help the child to know whether the language is SVO or OVS because they contain no information about which of the two NPs is the Subject and which the Object. Morgan ventures as far as claiming that this information would be available to children only if such utterances were presented to them with some form of “bracketing” of the VP, as in (26): (26) The dog [bites the cat]. For this reason, he underlines the importance of pauses and intonation to indicate syntactic boundaries. It is significant that Vivian Cook and Mark Newson, in preparing a well-known synthesis on the matter we are discussing, have taken his argument so seriously as to use it as a step of their own reasoning.20 This can only happen if one completely ignores that the child’s experience is full of utterances like the teacher punishes the pupils or mum prepares a pudding, where knowledge of the world and the semantics of the utterance (when not morphological agreement itself) are more than enough to make it clear who 20. See Cook and Newson (1996: 117). The relation between mind and language 367 is the subject and who/what is the object, ultimately rendering completely obvious the syntactic datum which is not signaled by (25) – in other words, that the subject normally precedes the verb, while the object follows it. Summarizing, there are at least two components at work in the comprehension of linguistic utterances, namely comprehension of syntax and comprehension of reality. Of the two, reality is by far the most complex. Consequently, as we have said in the previous section, perhaps we should not wonder how children manage to master the syntax of a language, but rather how they manage (conceptually) to master reality.21 Moreover, it is worth underlining that the knowledge of syntax is probably less important than the knowledge of reality in the process of interpretation of utterances. If this is true, and if we want to explain how children understand linguistic utterances, it may be insufficient, and consequently misleading, to concentrate on syntax as if reality did not exist. It is worth quoting what Slobin (1975: 30) writes on the matter: Most studies of child language comprehension put the child into a situation where there are no contextual cues to the meaning of utterances, but in real life, there is little reason for a preschool child to rely heavily on syntactic factors to determine the basic propositional and referential meaning of sentences which he hears. Judith Johnston and I have gone through transcripts of adult speech to children between the ages of two and five, in Turkish and English, looking for sentences which could be open to misinterpretation if the child lacked basic syntactic knowledge, such as the roles of word order and inflections. We found almost no instances of an adult utterance which could possibly be misinterpreted. That is, the overwhelming majority of utterances were clearly interpretable in context, requiring only knowledge of word meanings and the normal relations between actors, actions, and objects in the world. So, besides analogy, there is another very important thing that those who try to demonstrate the existence of an innate UG tend to ignore, or at least to underestimate, namely, the capacity of the child, and the speaker in general, to have recourse to the knowledge of reality. I think one of the many examples of this attitude can be seen in the argumentation proposed by Hoekstra and Kooij (1988: 38),22 that we resume here. Let us consider utterances (27) and (28): (27) Where did John say that we had to get off the bus? 21. See Arbib and Hill (1988: 60). 22. Arbib and Hill (1988), besides providing an alternative proposal to the innatist one to explain the acquisition of negative information (which is the topic of our Section 4.2.2.), attack this example in a radical way, lamenting the complete absence of evidence that small children are really able to decide on the ambiguity or otherwise of (27) and (28). Although we agree with their caution, here we keep on a strictly linguistic level, that they do not directly consider. (See also Lombardi Vallauri (1999: 718)). 368 (28) Edoardo Lombardi Vallauri Where did John ask whether we had to get off the bus? Every speaker is aware that (27) is ambiguous, while (28) is not, because in (27) the interrogative adverb where can belong either to the main or to the subordinate clause, while in (28) it can only belong to the main clause. Consequently, in (27) the question can be either about the place where John has spoken to say that we had to get off, or about the place where, in John’s opinion, we had to get off. On the contrary, in (28) the question is only about the place where John has spoken. According to our authors, the knowledge of these facts is not accessible to the speakers through induction, and consequently must depend on some innate instruction. If we consider the utterances from a purely syntactic point of view, we may agree. But as soon as one takes into account some semantics and some of the extralinguistic reality to which the semantics of the utterances refers, one can see that they contain the reason why the speaker/acquirer knows that (27) can be ambiguous and (28) cannot. The difference resides in the different meanings of the verbs say and ask, therefore in their being suitable for different states of affairs. The two interpretations of (27) may be paraphrased as follows: (27) a. b. Where did John speak, to say that we have to get off the bus? According to what John said, where do we have to get off the bus? Both alternatives make sense. The same doesn’t hold for the corresponding interpretations of (28): (28) a. Where did John speak, to ask whether we had to get off the bus? b. *According to what John asked, where did we have to get off the bus? In fact, since it doesn’t have the sense of “asserting,” the verb to ask cannot mean a speech act from which information arises (in assertive modality) about where it is better to get off. It is possible to wonder where John has asked a question, but it is not possible to wonder what instructions his question contained about where to get off; this is because a question, in virtue of its illocutionary status, does not contain any instructions. This is why (28) has only one interpretation. World knowledge, and in particular that of the difference between a question and a statement, allows the speaker to exclude the alternative meaning of (28) despite “syntactic possibility.” In order to do this, there is really no need to imagine that the brain hosts a grammatical module pre-existent to all experience. The relation between mind and language 369 4.2.2. Negative information. Another element usually underestimated by those who support the innateness of grammar23 is the possibility of infering the ungrammaticality of a structure from its systematic absence. In addition to the connection between reality, meaning, and syntax, it is reasonable to think that analogy a negativo is also at work, that is to say, the capacity to infer that some elements do not belong to the language, simply because such elements never appear, even if there are no explicit warnings of their being unacceptable in the stimulus. This is a sore point in the innatist tradition because the argument recurs everywhere, and always naïvely, to say the least. Chomsky (1965: 25) sets out the question by dividing the information received by children into two types: “positive” and “negative.” The former consists of utterances produced in their presence without warning of their acceptability, and the latter consists of utterances for which they receive explicit warning of unacceptability, by way of some kind of correction. Even Wexler and Culicover (1980: 63), who devote a ponderous study to the formal conditions constraining the acquisition of grammars, make the following statement in a paragraph titled Negative information: If the learner hears a sentence, he can assume that it is in his language. But if he does not hear a sentence, it does not imply that the sentence is ungrammatical. Possibly he has simply not yet heard the sentence. And, more importantly, all the following discussion addresses the problem (much studied elsewhere24 ) of if, and in what way, parents present children with elements of negative evaluation about unacceptable language structures, mainly (but not only) by correcting them explicitly. In such arguments it is never hypothesized that the repeated, systematic and infallible absence of a structure, in spite of thousands of potential opportunities for occurrence, may at length lead the acquirer to the certitude that such a structure does not belong to the language.25 For example, one may see some naïvety in the argument of Hoekstra and Kooij (1988: 37), where innate UG is credited the fact that an English speaker, aware of the acceptability of (29a–c), is also aware of the unacceptability of structures like (29d), although nobody ever signals this explicitly: 23. A promising comment by Chomsky (1981a: 8–9), mentioning indirect negative evidence, has remained poorly developed, perhaps because it fails to grasp the full relevance of the question, only mentioning the idea that the presence of certain structures with certain functions may induce the child to exclude the possibility of the same function being expressed by different structures. 24. Scholars usually focus on the idea that such negative information, based on corrections or failure of comprehension on the part of parents, is almost absent in the child’s experience. See, for example, Pinker (1984: 29). 25. This is the “entrenchment” pattern described by Braine and Brooks (1995). 370 (29) Edoardo Lombardi Vallauri a. He proved the theorem. b. his proof of the theorem c. He proved the theorem wrong. d. *his proof of the theorem wrong Over time English speaking children are exposed to thousands of sentences with a transitive verb (e.g., prove) where the object is followed by a predicative complement (e.g., wrong), but they never hear an example where the same happens after a nominalization (e.g., proof ). So, why should they take the initiative to consider acceptable a structure that never appears in their experience, and in whose place (i.e., with the same function), they can observe other structures, such as (29e), (29f) or (29g)?26 (29) e. f. g. his proof/proving that the theorem was wrong his proof/proving of the theorem’s wrongness/falsity his having proved the theorem wrong Another example of the underestimation of how significant the constant repetition of a pattern can be is the opinion that the fixing of the pro-drop parameter cannot happen without a pre-existing, innate principle.27 It is well known that there are (“non-pro-drop”) languages where the subject must always be expressed and (“pro-drop”) languages where it can be omitted. Children orientate themselves in their language on the basis of what they hear. According to Cook and Newson (1996), who cite Chomsky a number of times, the only way a child can understand how things are arranged in his/her language is a parametric predisposition with two values: pro-drop and non-pro-drop: Children must be learning either from positive evidence alone or from indirect negative evidence, such as the lack of null-subject sentences in English. This is possible only if their choice is circumscribed; if they know there are a few possibilities, say pro-drop and non-pro-drop, they only require evidence to tell them which one they have encountered.28 Hyams (1986) goes as far as suggesting that, in acquirers of English, the fixing of the innate non-pro-drop parameter may be triggered by the existence of expletive subjects. Sentences like once upon a time there were three bears or it’s time for bed would be necessary for the child to understand that the expression 26. This is the “preemption” pattern described by Braine and Brooks (1995). For an accurate discussion of this kind of induction on the part of children, see also Bowerman (1988), Tomasello (2000a, 2000b, 2003), Brooks and Tomasello (1999), Brooks, Tomasello et al. (1999). 27. See Hyams (1986) and Cook and Newson (1988: 110–111). 28. Cook and Newson (1988: 110–111). Italics mine. The relation between mind and language 371 of subjects is obligatory in English. I think that the very reason why Englishspeaking children tend to always produce sentences with subjects is simply the fact that the thousands of sentences they hear every day all contain overt subjects. Similarly, Italian or Spanish children feel that they do not always need to use the subject because the subject is not always present in the sentences they hear. I do not see why an innate switch in the brain is necessary for that. In the same way, because all the cars they see have wheels, children are led to put wheels on all the cars they draw; while since only some houses have a chimney, they learn that they can draw a house with or without one. Moreover, it seems to me that Italian or Spanish children acquire (it not being innate) something that is much more complex than the simple notion that the subject is optional: They learn when and when not to make it explicit, according to the familiarity of the referent established by the context, and to the informational status they want to give to the subject. One of the most frequent examples in support of the innateness of grammar is the principle of structure-dependency, whereby language knowledge is based on the structural relations within utterances, and not on simple word sequentiality. Cook and Newson (1996: 3–13, 30, etc.) give it great importance.29 For example (pp. 82–83, 293), an English utterance like (30) would be recognized as unacceptable precisely because it violates the innate principle of structuredependency: (30) *Is Sam is the cat that brown? We can accept calling what is violated by an utterance like (30) the “structuredependency principle”, but this does not imply that it constitutes innate linguistic knowledge. On the contrary, it may be observed that, if it is innate, the principle of structure-dependency is not specifically linguistic because a great quantity of what we do is structure-dependent: We continuously manage reality in terms of complex objects, material or abstract, considering them as structured units and, so to say, further “grouping” them according to the structural relations we see among them. For example, when setting a table, we do not treat objects one at a time and based on the order they originally occupy in space (= in drawers and cupboards); on the contrary if, say, we want to change or eliminate someone’s place at table we move the whole setting, which we interpret as a complex and structured entity. Many other similar examples may be given.30 When we see 29. See also Cook (1991: 104–105). Many recent works by some authoritative representatives of the theory use this argument in a central role to support the innatist thesis; see also, for instance, Moro (2002: 43–44). 30. See on the matter some observations by Jean Piaget in Piattelli Palmarini (ed.)(1979), and Anderson (1983). 372 Edoardo Lombardi Vallauri all the parts of a bicycle laid out, do we perhaps not realize that they make up something other than a bicycle assembled according to its correct structure? Of course those who produce bicycles do it in a way which depends on structure, and we all recognize a bicycle that can work from one whose parts are assembled higgledy-piggledy; but there is no need to postulate the existence of an innate Bicycle Organization Device. However, the innatist position underlines the fact that children hear lots of sentences respecting structure-dependency, but they are never exposed to the explicit censure of sentences like (30), which violate it. Thus, they have no direct information on the prohibition to violate structure-dependency, and consequently this knowledge could only be innate. This is a very weak argument. Children see that objects always fall downwards; nobody ever explains to them that objects do not fall upwards, but they know it all the same, and they are fascinated by colored balloons filled with helium, precisely because they do not go downwards. The fact that something always happens and something else never does is very instructive and allows for generalizations. The very often-cited argument that children have no elements to exclude wrong structures because negative stimuli (parents quoting a wrong sentence to censure it) are extremely rare in their experience is only valid if we imagine the child as being completely unconscious and incapable of generalizations such as Gold’s (1967) algorithm,31 which compares input data with the grammars and accepts all grammars that do not violate the input. But there must be a difference between a grammar that does not violate any data while producing innumerable structures that are not to be found in the input and another grammar that, besides not violating the data, does not produce any structure unforeseen by the input either! If something never happens in a corpus of thousands and thousands of utterances where it could theoretically happen, the child (unlike Gold’s algorithm) ends by inducing that such a thing is excluded. However, let us see in detail an example of alleged effectiveness of the innate principle of structure-dependency in the linguistic faculty of the child. Crain and Nakayama (1987)32 observe that, in English, it is possible to derive (32) from (31). 31. See Gleitman and Wanner (1982: 5–7). Sampson (2002: 96–97) suggests an interesting comparison between language acquisition and “the adult scientist seeking to understand a previously unexplored domain of natural phenomena,” who obviously cannot “observe violations of laws of nature, accompanied by indicators that these are negative rather than positive examples [. . .]”, and still can formulate a theory excluding all irrelevant facts. See also Pullum and Scholz (2002: 31). 32. Cited in Akmajian et al. (1984). The example and the argument actually derive from Chomsky (1980); see Pullum and Scholz (2002: 39). The relation between mind and language (31) The man is tall. (32) Is the man tall? 373 From this, the child may deduce that the rule for making a question from an assertive clause is: Take the first verb and put it in first position. The result would be that one would derive (34) from (33). (33) The man who is tall is in the room. (34) *Is the man who tall is in the room? Now, “children do not produce questions like the ill-formed (34). Therefore, it appears that children know that structure, and not just the more salient linear order property of sentences, is relevant in the formation of yes/no questions.”33 It is surely true that a small child, in order to build the well-formed interrogative clause with verb fronting (Is the man who is tall in the room?), as well as having a remarkably precocious or warped mind, must somehow know that (33) incorporates some structure and is not a mere list of words. In particular, he/she must know that who is tall is something that has unity and cannot be damaged at will. But, crucially, why should we think that such knowledge is innate? Isn’t it true that the speech to which the child is exposed is full of examples of structure-dependency? Isn’t the child often exposed to a sort of commutation test, such as: (35) – Beware of the man with the gun. – Which man? The tall one? – The one who has a gun. Commutation tests like (35) may well make the child understand that the relative clause (the thing introduced by who) constitutes a unit. Besides, English speaking children are exposed to thousands of occurrences of the relative pronoun, among which not even one is followed by a predicative adjective. After the relative pronoun comes the verb (or a Subject, if the relative is not a subject). Consequently, there is no need for innate knowledge to judge that who tall in (34) is not acceptable. Moreover, the semantics of utterances can very well lead to understanding that the function of relative clauses is to specify or restrict the meaning of their heads, while the main predication is effected by the main clause. Therefore, after hearing a number of relative constructions, it is possible to work out that an interrogative sentence must be derived from them without touching the relative clause. This is because what one wants to turn into a question is not the 33. Akmajian et al. (1984: 470). 374 Edoardo Lombardi Vallauri content expressed by the verb in the relative clause, but rather the content of the main verb. In fact, all the questions containing a relative clause that a child may encounter will be questions where the relative clause is left untouched and fronting (or the do/does construction) takes place in the main clause. This is already apparent in sentences that, as Crain and Nakayama’s example illustrates, contain the verbal form is in both clauses. But there are also many main-relative clause constructions in the language where this is even more evident because the two verbs are different, thus having different semantic content. Children may well be exposed to examples containing elements that can prevent them from making mistakes. There will be no discussion here of Chomsky’s probably exaggerated assertion (in Piattelli Palmarini 1980: 40) that “a person might go through much or all of his life without ever having been exposed to relevant evidence” (such as ex. 38 or 42 below) that may lead him/her to know that the verb to be fronted in order to make a question out of a relative construction is that of the main clause. For its confutation, see Pullum and Scholz (2002), Sampson (2002). Still, the issue is controversial (see Legate and Yang 2002: 155). The points we make here are obviously relevant if Chomsky’s claim is wrong, but may still be significant if it is right. Quite evidently, if the content of the main verb in (36) has to be turned into a question, its interrogative version cannot be (37), where the verb of the relative clause undergoes the interrogative transformation: (36) (37) The man who owns a dog sleeps in the room. *Does the man who own a dog sleeps in the room? Rather it must be: (38) Does the man who owns a dog sleep in the room? One may object that neither of the verbs here admits fronting and that the example is not significant because, containing no verb movement, it does not allow us to act by analogy when forming the interrogative version of utterance (33). So, let us examine a “pitfall” case, where the verb in the relative clause admits fronting, while the main verb does not, thus encouraging the mistake represented by (34): (39) The man who is in the room owns a dog. If we applied the “rule” that disregards the structure, and moved the first verb to the first position we would obtain: (40) *Is the man who in the room owns a dog? The relation between mind and language 375 Such an error is excluded quite simply because no child has reason to think that when you move the verb is to first position what becomes interrogative is the content of the verb owns. The rule followed by the child is actually something like this: when allowed by the verb, moving it to the first position has the effect of interrogating ITS content (and not the content of another verb). In attaining this conclusion, children may be helped by the fact that the interrogative version of an utterance such as (41) is (42) because it is the idea of being old that has to be questioned, and not that of having won the battle. (41) The man who won the battle is old. (42) Is the man who won the battle old? But even if such sentences are rare or absent in the stimulus, children confronted with this problem for the first time (as perhaps those tested by Crain and Nakayama) can easily imagine that if one wants to turn the main clause into an interrogative clause, one has to move to first position the verb of the main clause, and not that of the relative clause. The same happens, for instance, with negation: If you want to negate the content of a verb, you put the negation close to it, and not close to another verb in the same sentence. This explains why children know that one has to respect relative clauses when building interrogative utterances. So, we can certainly refer to a principle and call it structure-dependency principle, but there is no reason to believe it is innate, since the information required for its acquisition is present in the stimulus. Let us finally consider an authoritative example of alleged phonologic innateness. Chomsky (1988: 25) claims: the person who has acquired knowledge of a language has quite specific knowledge about facts that transcend his or her experience, for example, about which nonexistent forms are possible words and which are not. Consider the forms strid and bnid. Speakers of English have not heard either of these forms, but they know that strid is a possible word, perhaps the name of some exotic fruit they have not seen before, but bnid, though pronounceable, is not a possible word of the language. Speakers of Arabic, in contrast, know that bnid is a possible word and strid is not; speakers of Spanish know that neither strid nor bnid is a possible word of their language. The facts can be explained in terms of rules of sound structure that the language learner comes to know in the course of acquiring the language. Acquisition of the rules of sound structure, in turn, depends on fixed principles governing possible sound systems for human languages, the elements of which they are constituted, the manner of their combination and the modifications they may undergo in various contexts. These principles are common to English, Arabic, Spanish, and all other human languages and are used unconsciously by a 376 Edoardo Lombardi Vallauri person acquiring any of these languages; the principles belong to the innate language faculty, a component of the mind/brain. Although this argument is proposed in favor of the innateness hypothesis, it can obviously be seen as supporting the hypothesis of acquisition from the environment. Chomsky is aware that different languages have different sound systems, and must elaborate some explanation to reconcile this fact with the existence of an innate universal phonology. Such an explanation should be based on the idea that if the speakers can exclude sequences they have never heard without anybody having ever told them that such sequences are unacceptable, the only reason why they can do this is some innate knowledge. But it must be observed that every speaker listens very many times a day to all the possible sound sequences of their language. This makes them able to recognize them and to develop the awareness that they know them all. In other words, they can distinguish what is part of the language and what is not, exactly as the director of a school recognizes all the pupils and, upon running into some new face in the corridor, understands it is probably a newcomer or an occasional visitor. The same holds for the commander of a battalion, a prison governor, etc. An English speaker, having never met an initial bn- sequence, is simply able to say, “I’ve never heard this one, and if it’s new to me who listens to English all the time, then it can’t be English.” And they really don’t need anyone to tell them, “initial bn- is unacceptable.” I think this is the most economical explanation, far more so than imagining that the brain contains deep, innate rules stipulating the unacceptability of initial bn- in a phonological system that presents the “deep characters” of English, while at the same time admitting it in a phonological system such as that of Arabic. 4.2.3. Cumbersome innate information. Some rules put forward for UG are more complex than those of any single language because they are the result of trying to prove that black is white and white black in order to account for all languages at the same time, something a child can happily ignore. A good example is the rule for the reference of reflexive elements (the “Governing Category Parameter”, see Section 4.1) proposed by Wexler and Manzini (1987). They wonder: How do children learning English set the parameter on the most restrictive value, since all the examples they hear are also compatible with the most extensive value? We can answer that, because of the absolute nonoccurrence of cases b, c, d, and e, a child can obviously deduce that they are not grammatical. This is, in fact, what any person who does not feel the need to postulate an innate UG will conclude: The systematic absence of an option can be interpreted by the child as evidence of its unavailability. In order to avoid this explanation, Wexler and Manzini are forced to accept (and they imagine that it resides in the brain) the so-called subset principle, which would lead children The relation between mind and language 377 to test certain values of the parameter before others, so that English children are content with the most restrictive one as soon as they have experimented with it, and raise no further questions. Children acquiring another language, on the contrary, change their minds and set the further value of the parameter because they hear positive examples that oblige them to do so. In short, children do not immediately make hypothesis (e), rather they first check (a), and only extend the range of possibilities if they find violations. This is very reasonable, but it is not clear why we should make the hypothesis that children have all these pre-set and so to say built-in options, when it is sufficient for them to produce all the structures they hear by imitation, and not to produce those they do not hear. We have already observed that such a hypothesis presupposes that the brain has a high capacity for containing data, and a poor capacity for processing them. We will explore some possible consequences of this assumption in the next section. 4.3. More general arguments 4.3.1. A greater stock. As we have observed, the innate-UG thesis, being based (although not explicitly) on the hypothesis that the mind/brain is a machine with an unlimited hard disk and a weak CPU, a hypothesis logically bound to the idea of Principles and Parameters, ends up justifying further conjectures. We briefly return to the question here, simply hinting at one of the possible alternative scenarios. For example, what if the hard disk were able to store an even greater amount of information? What if children acquiring language were able to (and actually did) store not only elements of the lexicon, but also most of the sentences they heard – if, at least in a certain stage, instead of the rules of the grammar, they had many sentences stored “whole” and used them as patterns for the production of similar sentences by simply substituting the lexical elements? It would be possible to hypothesize that when producing a sentence like dad is bad, a child does not apply any rule (i.e., the child doesn’t “dress” an abstract syntactic structure in the mind with appropriate lexical items), but simply substitutes dad and bad for the lexical entries of some sample sentence (such as mom is good) that is in memory and that works as a concrete pattern for countless other sentences.34 Then dad is bad may also 34. A similar hypothesis is often exposed in M. Tomasello’s works and corroborated experimentally for the production of children’s first complex sentences by Diessel and Tomasello (2000, 2001). It is not excluded that such a process can also take place in foreign language acquisition. I probably learned French in this way. My father used to sing French boy-scouts songs with us children (and translate them for us). We soon learned them by heart. Thirty songs represent a considerable inventory of sentences. Long before opening our first French grammar book, my brother and I were able to speak French by analogy, simply producing innumerable 378 Edoardo Lombardi Vallauri join the set of sample sentences, which might become extremely extensive and varied, thus rendering the task of creating new sentences most straightforward. The growth of language would resemble that of an oil stain expanding partly from its centre through the addition of sentences with new structures that the child hears from adults, and partly from the periphery, through the continuous creation (by the child), and immediately subsequent annexation, of sentences built on the pattern of already existing ones. After a while, of course, there would be no more need to expand the number of sample sentences. A certain number would be enough to set off all necessary creative automatisms. It must be stressed that, in order to replace mom with dad and good with bad, there is no need for the children to have a theory of lexical categories like nouns and adjectives. Rather, it is sufficient that they have an idea of the meaning of each word, conceived as its usefulness in actual situations. In other words, it is sufficient to be eager to say that dad is bad. Now, it is reasonable to think that in the child the attribution of meaning to a certain number of words precedes any beginning of syntactic organization, this being probably reflected in the stage (that all children go through) of one-word utterances, and partly also in the stages of two- and three-word utterances with poor syntactic organization.35 4.3.2. Brain compulsion, or optional respect of legality? A problem for UG also arises from the fact that in most cases, speakers are perfectly able to understand the meanings of grammatically incorrect utterances. This happens frequently for utterances produced by non-native speakers, where, although they violate the grammar rules, it is clear what they are “getting at”. More generally, producing unacceptable utterances is very easy, and we are all able to do it. In short, grammar- (and UG-) violating utterances are both producible and processable; the brain is perfectly able to produce them and, in innumerable cases, to understand much or even all of their meaning. Their supposed “incompatibility” with the brain seems limited to the fact that the latter can recognize their not belonging to a certain language. This seems to undermine the idea of innate grammar. If it were innate, it would be very difficult or even impossible for us to violate it, exactly as we cannot hear acoustic frequencies that are higher or lower than a certain range, or as we cannot prevent our nervous systems from upturning the images that form on our retinas, as we cannot prevent it from secreting adrenalin in case of danger, or other hormones in the presence of an attractive person, etc. These are processes determined by innate characsentences whose structure was identical to that of at least one of the sentences we knew by heart from one of the songs. Later on, occasional contact with French speakers allowed us to extend our data base slightly, and when at the age of eleven we started studying that language at school, the teacher couldn’t believe we had never been taught French before. 35. See Anderson (1983: 270–271). The relation between mind and language 379 ters, which consequently constrain us in an unavoidable way. We are able to do certain things, and we cannot do them in a way other than the way we are “programmed” to. Grammar, on the contrary, seems to be a matter of legality, rather than of capacity. We recognize that certain structures (i.e., certain patterns of behaviour) are “illegal”, that is to say, different from the conventions ruling that aspect of our behaviour, but nothing within us prevents us from adopting them. At most it is a preference, but surely not a biological obligation. Respecting the rules of UG is far more similar to respecting the rules of basketball than to the fact that, to move on the basketball court, we adopt an upright position and activate our five senses. In short, if grammar is as constraining and unavoidable as the inversion of the visual image, at the most it is so for the community that adopts it, but not for the individual. However, the seat of innatism has to be the individual, otherwise it is no longer innatism, but environmental learning. The innatist thesis confuses the level of the community with the level of the individual. The inviolability of grammar rules only holds within the community (and even there, in a variable and relative way), by way of convention; the individual conforms to them for the sake of the community and in order to avoid friction with it, but the brain would be perfectly able to produce any odd-sounding utterance prohibited by any putative UG.36 We can wonder what would happen if a community of children were to grow linguistically isolated from birth, for instance brought up by a group of mute or aphasic nurses. How would they set the parameters? Would they be able to develop a language, in spite of the absence of necessary information? They would all walk (even if the nurses weren’t able to) and carry out other innate human functions, but linguistically would they falter, because innate grammar is only potential? The idea of principles and parameters, conceived to reconcile innateness with language diversity, ends by describing quite a strange innate capacity, which, after all, only allows us to do what we . . . learn. One can guess that those children would probably develop some communication system, not necessarily verbal and articulated, and even less necessarily conforming to the latest version of UG.37 A very primitive communication system would evolve, 36. Fabio Rossi points out to me that if we change the analogy from the examples above to others such as the use of our legs, it becomes conceivable that UG is as innate as walking, and like walking it may admit some dispreferred options, such as hopping on one foot. The objection is good, and the whole problem obviously deserves further in-depth analysis; but Daniele Portolan’s (personal communication) reponse to Rossi is also true: Hopping on one foot is not a disorderly and imperfect use of our legs, rather a “further” use, organized and highly voluntary, to be compared with speaking with gestures, rather than with “grammarless” speaking. 37. The study of pidgins and creole languages obviously provides hypotheses on this kind of problem. However, even when the grammar to be acquired is so “rarefied”, it cannot be said to be completely lacking. 380 Edoardo Lombardi Vallauri becoming more powerful and sophisticated through the successive history of that community, down the generations. But this is actually nothing more than an opinion, derived from the persuasion that language is a product of history, belonging to the phylogenetic development of communities, not a product of individual, ontogenetic development. 4.3.3. A weaker UG is superfluous to language justification. UG may also be conceived more soberly in a “weak” and extremely generic form, and therefore less specifically linguistic than in the Chomskyan tradition: as the very general fact that a language cannot do without a syntax characterized by regularity and recursive patterns. And it may be argued that this is already specifically linguistic. But even such an extremely general feature is not necessarily to be regarded as determined by some innate configuration of the brain. In fact, even without postulating the innateness of grammar, many reasons why language tends to assume some universal characteristics, independent of single languages, are already apparent, not only on pragmatic grounds and as a consequence of speech organ physiology, but also and above all for structural reasons, deriving from the very definition of a system designed for communication. In order for languages to be confined within quite narrow margins of variation, there is no need to hypothesize an innate grammar in the brain. If we admit the existence of some ten constraints like the ones listed below (but many other, more detailed ones may be called upon38 ), we will already have narrowed the field considerably: Constraint 1. Monogenesis: All languages still carry within them the remnants of the first one, which prevented them from varying in a completely free way. Constraint 2. Structure of reality and social interaction: A verbal language is far more efficient than a proxemic/gestural one. This is both because of the variety of distinctions it allows us to manage, and because it allows communication in discontinuous environments, keeping eyes and hands free for other tasks.39 Constraint 3. a. Structure of the articulatory and auditory organs, that do not allow an infinite number of distinctions to be in operation, and b. Cognition of reality as composed by many entities: 38. See Lombardi Vallauri (1999) for an overview of the constraints that bind any form of articulate language, thus determining linguistic universals. 39. See Lombardi Vallauri (1999: 726). The relation between mind and language 381 In order to produce as many words as requested by our conceptualization of reality, phonemes are not sufficient, and something like double articulation is required. Constraint 4. Cognition of reality as made up of at least entities, states and processes: Necessity for any language to be equipped with (if not precisely nouns and verbs) instruments for designation and predication. Constraint 5. Cognition of reality as composed of entities that interact with each other in relations and happenings: Since the relations between entities are to be expressed, instruments are needed to express the relations between words denoting those entities. The nature of such instruments is determined by Constraint 6. Constraint 6. Development of the signal in time, and its linearity: Relations between words can only be expressed through linearity – that is, either morphologically: bound relational morphemes agglutinated to the lexical ones; or syntactically: free relational morphemes, grammatical exploitation of constituent order (= “configurationality”). Constraint 7. Existence of events and change in time, consequently the distinction between what is known because it has already appeared and what appears for the first time and is therefore new: Every language needs means to express given and new, topic and focus, presupposed and asserted information. Constraint 8. Limited short-term memory and limited computing capacity: A set of rules is needed. It is not feasible (in terms of time and energy) to invent a new method on the spot every time one has to translate a thought into an utterance. It is better to have routines one can rely upon, in order to “work automatically”. The demand for a structured system already arises in ensuring the simple (and purely interior) expressive function of language: Grammar is needed in so far as it is an automatized structure one can rely upon in order not to be obliged to invent every time a means by which to express one’s thoughts, like, for example, the painter who has to choose the lines and colors for a picture. The reasons for the establishment of regularity, no matter of what kind, are therefore, above all, economic in nature. It evidently allows us to reduce the mental load during coding. If speakers were obliged, for every utterance, to decide ex novo where to put every element of the sentence and how to signal their functions, they would consume far more energy and would probably speak far more slowly 382 Edoardo Lombardi Vallauri than they do, relying on the habit of conforming to the already established rules of grammar, not to mention how difficult the task of interpretation would be for the addressee (Constraint 9). Constraint 9. Communicative function: A set of rules is needed. In order to be understood, it is not enough to throw a mass of words/concepts, whose reciprocal relations we perceive somewhat idiosyncratically, at our addressee. We must do this in a conventionalized way where intended relations can be understood intersubjectivally. Therefore, in order to communicate, there must be agreement not only on the meaning of lexical items, but also on the interpretation to be given to the relations among them when they appear in utterances: this is to say, to a set of grammatical signals. In short, the presence of grammar rules (morphology and syntax) is rendered unavoidable by the function of sharing language. Language is an intersubjective activity which, in order to work, presupposes the agreement of two or more subjects on the modalities of its working. If such an agreement were lacking, the language would no longer work, just as any legal system (the laws of a State or the rules of baseball) wouldn’t, because for the purposes of such systems, working differently for each single participant means not working at all. This means language has to be a system rather than a series of spontaneous behavioral acts that may vary freely on the initiative of single individuals, and therefore be unforeseeable and unmanageable by the others. In other words, its nature as an intersubjective and shared device requires the presence of regularity. This is a good reason why language tends to regularize itself more than many other aspects of human behavior. Therefore, such regularity should not necessarily be interpreted as being genetically motivated. Of course the conformation of the brain allows it to manage regularity and recursion, but it does not require language to use them. Such a requirement is, on the contrary, imposed from outside.40 If we refer to the last two constraints, we can say that in any intersubjective communication system there is the need for a fair amount of regularity. This suffices to ground the universality of grammar qua talis, but not of the same grammar in all languages. Some concrete aspects of grammars may be 40. Something similar is implied by Searle (1972: 34–35) when he claims that “functional phenomena [. . .] function the way they do because we assign them a certain status,” and that this status requires “a system of rules,” which he calls “constitutive rules” because “such rules do not merely regulate existing activities, [. . .] but they create the very possibility of such activities.” The relation between mind and language 383 universal because of constraints like 1–7 above, but for the rest every language chooses which regularities to adopt as a consequance of the interaction of many internal and external factors, which we do not dwell upon here. 5. Conclusion The appeal of biological orientation in generative grammar, as presented for instance by Moro (1988: IX – X), is the assumption that (our translation) “grammar must not only account in a unitary way for all facts of the different languages (descriptive adequacy), but it must above all provide a model of language acquisition capability exclusive to our species (explanatory adequacy).” In view of what we have been saying here, it seems to me that the presupposition contained in this kind of statement, namely the existence of a specifically grammatical capability at the basis of language acquisition, still needs appropriate foundation. Moreover, it seems to me that in spite of extensive research and formulation efforts, such an assumption is as weakly founded today as it was at its early stages, particularly with regards to the PSA, which has not made any progress since its first formulations; at the same time, it is scarcely confirmed a posteriori by a satisfactory unitary model of UG. All this makes it increasingly hazardous, decades on, to go on proposing it as a self-evident and, so to say, “starting point” hypothesis. Besides, since the first formulation of the problem, there have been attempts to propose language acquisition models that do not presuppose an innate grammar, some of which are extremely interesting. We cannot go into the theses, upholding Piaget’s position, and which include those of Martin Braine, John Anderson, Michael Arbib and Jane Hill, Melissa Bowerman, Michael Tomasello and many others who deserve great attention41 . Clearly, if language is not conceived as the product of specifically and exclusively devoted faculties, but as the product of general faculties of the mind, then it can be regarded not only as a “window on the Language Faculty” but also as a “window on the Mind”, and this surely doesn’t make it less interesting. The aim of this work was, however, simply to specify a possible destruens and strictly linguistic argument against the very central innatist argument of the poverty of the stimulus which, in spite of its frequent general formulations, is far less demonstrated than is usually presumed, and is perhaps accepted by many with excessive confidence42 . It is perhaps not completely wrong to call 41. See for example Piattelli Palmarini (ed.) (1979), Anderson (1983), Braine (1971a, 1971b, 1992); Arbib and Hill (1988); Bowerman (1988), Tomasello (2000a, 2000b, 2003). Some of them, like Braine’s proposal on the so-called discovery procedures, seem to be surprisingly corroborated by recent developments of neurological sciences, such as the so-called “neural darwinism” by G. M. Edelman, and others. 384 Edoardo Lombardi Vallauri it back into question, thus challenging the presumed specifically linguistic and even grammatical nature of the biological bases of language, in particular those that are seen as located in the brain. Università Roma Tre References Akmajian, Adrian, Richard A. Demers, Ann K. Farmer and Robert M. Harnish (1984). Linguistics: An Introduction to Language and Communication. (1st ed. 1979) Cambridge, MA: Cambridge University Press. Anderson, John (1983). The Architecture of Cognition. Cambridge, MA: Harvard University Press. Arbib, Michael A. and Jane C. Hill (1988). Language acquisition: Schemas replace universal grammar. In Explaining Language Universals, John A. Hawkins (ed.), 56–72. Oxford: Blackwell. Berwick, Robert C. (1985). The acquisition of Syntactic Knowledge. Cambridge, MA: MIT Press. Berwick, Robert C. and Amy S. Weinberg (1984). The Grammatical Basis of Linguistic Performance. Cambridge, MA: MIT Press. Bloomfield, Leonard (1933). Language. New York: Henry Holt. Boncinelli, Edoardo (1999). Il cervello, la mente e l’anima. Milano: Mondadori. Bowerman, Melissa (1988). The ‘no negative evidence’ problem: How do children avoid constructing an overly general grammar? In Explaining Language Universals, John A. Hawkins (ed.), 73–101. Oxford: Blackwell. Braine, Martin D. S. (1971a). The acquisition of language in infant and child. In The Learning of Language, Carroll E. Reed (ed.). New York: Scribners. — (1971b). On two types of models of the internalization of grammars. In The Ontogenesis of Grammar, Dan I. Slobin (ed.), 153–186. New York: Academic Press. — (1992). What sort of innate structure is needed to “bootstrap” into syntax? Cognition 45: 77–100. Braine, Martin D. S. and Patricia J. Brooks (1995). Verb argument structure and the problem of avoiding an overgeneral grammar. In Beyond Names for Things: Young Children’s Acquisition of Verbs, Michael Tomasello and William E. Merriman (eds.), 353–376. Hillsdale, NJ: Erlbaum. Brooks, Patricia and Michael Tomasello (1999). How young children constrain their argument structure constructions. Language 75: 720–738. Brooks, Patricia, Michael Tomasello, Lawrence Lewis and Kelly Dodson (1999). Children’s overgeneralization over fixed transitivity verbs: The entrenchment hypothesis. Child Development 70: 1325–1337. Carruthers, Peter (2002). The cognitive functions of language. Behavioral and Brain Sciences 25 (6): 657–674. Cavalli Sforza, L. Luca, Paolo Menozzi and Alberto Piazza (1994). The History and Geography of Human Genes. Princeton, NJ: Princeton University Press. Cecchetto, Carlo and Luigi Rizzi (2000). A Naturalistic Approach to Language. In Naturalism in the Cognitive Sciences and in the Philosophy of Mind, Sandro Nannini and Hans Joerg Sandkühler (eds.), 117–130. Bern: Peter Lang. Chomsky, Carol S. (1969). The Acquisition of Syntax in Children from 5 to 10. Cambridge, MA: MIT Press. 42. This affects the literature not only in linguistics, but also in other cognitive sciences that “rely on” it: see Pullum and Scholz (2002: 11) The relation between mind and language 385 Chomsky, Noam (1962). Explanatory models in linguistics. In Logic, Methodology and Philosophy of Science, Ernest Nagel, Patrick Suppes and Alfred Tarski (eds.), 528–550. Stanford: Stanford University Press. — (1965). Aspects of the Theory of Syntax. Cambridge, MA: MIT Press. — (1980). Rules and Representations. Oxford: Blackwell. — (1981a). Lectures on Government and Binding. Dordrecht: Foris. — (1981b). Principles and parameters in syntactic theory. In Explanations in Linguistics, Norbert Hornstein and David Lightfoot (eds.), 32–75. London: Longman. — (1982). Some Concepts and Consequences of the Theory of Government and Binding. Cambridge, MA: MIT Press. — (1986). Knowledge of Language: Its Nature, Origin and Use. New York: Praeger. — (1987). Transformational grammar: Past. present and future. In Studies in English Language and Literature, 33–80. Kyoto University. — (1988). Language and Problems of Knowledge. Cambridge, MA: MIT Press. — (1991). Some notes on economy of derivation and representation. In Principles and Parameters in Comparative Grammar, Robert Freidin (ed.), 417–454. Cambridge, MA: MIT Press. Cook, Vivian J. (1991) The poverty-of-the-stimulus argument and multi-competence. Second Language Research 7 (2): 103–117. Cook, Vivian J. and Mark Newson (1996). Chomsky’s Universal Grammar (2nd edition). Oxford: Blackwell. Crain, Stephen (1991). Language acquisition in the absence of experience. Behavioral and Brain Sciences 14: 597–650. Crain, Stephen and Diane Lillo-Martin (1999). An Introduction to Linguistic Theory and Language Acquisition. Malden: Blackwell. Crain, Stephen and Mineharu Nakayama (1987). Structure dependence in children’s language. Language 63: 522–543. Crain, Stephen and Rosalind Thornton (1998). Investigations in Universal Grammar: A Guide to Experiments on the Acquisition of Syntax and Semantics. Cambridge, MA: MIT Press. Dell, François (1981). On the learnability of optional phonological rules. Linguistic Inquiry 12: 31–37. Diessel, Holger and Michael Tomasello (2000). The development of relative constructions in early child speech. Cognitive linguistics 11: 131–152. — (2001). The acquisition of finite complemental clauses in English: A usage based approach to the development of grammatical constructions. Cognitive Linguistics 12: 97–141. Edelman, Gerald M. (1992). Bright Air, Brilliant Fire: On the Matter of Mind. New York: Basic Books. Fodor, Jerry A. (1983). The Modularity of Mind. Cambridge, MA: MIT Press. Fodor, Jerry A., Thomas G. Bever and Merrill F. Garrett (1974). The Psychology of Language: An Introduction to Psycholinguistics and Generative Grammar. New York: McGraw-Hill. Gleitman, Lila R. (1982). Maturational determinants of language growth. Cognition 10: 103–114. Gleitman, Lila R. and Eric Wanner (1982). Language acquisition: The state of the state of the art. In Language Acquisition: The State of the Art, Lila Gleitman and Eric Wanner (eds.), 3–48. Cambridge: Cambridge University Press. Gleitman, Lila R. and Eric Wanner (eds.) (1982). Language Acquisition: The State of the Art. Cambridge: Cambridge University Press. Gold, E. Mark (1967). Language identification in the limit. Information and Control 10: 447–474. Hauser, Marc D., Noam Chomsky and W. Tecumseh Fitch (2002). The faculty of language: What is it, who has it, and how did it evolve? Science 298: 1569–1579. Hawkins, John A. (ed.) (1988). Explaining Language Universals. Oxford: Blackwell. Heukstra, Teun and Jan G. Kooij (1988). The innateness hypothesis. In Explaining Language Universals, John A. Hawkins (ed.), 31–55. Oxford: Blackwell. Hyams, Nina (1986). Language Acquisition and the Theory of Parameters. Dordrecht: Reidel. 386 Edoardo Lombardi Vallauri Jackendoff, Ray (1997). The Architecture of the Language Faculty. Cambridge, MA: MIT Press. Jenkins, Lyle (2000). Biolinguistics: Exploring the Biology of Language. Cambridge: Cambridge University Press. Legate, Julie Anne and Charles D. Yang (2002). Empirical re-assessment of stimulus poverty arguments. The Linguistic Review 19 (1–2): 151–162. Lieberman, Philip (1984). The Biology and Evolution of Language. Cambridge, MA and London: Harvard University Press. — (1991). Uniquely Human: The Evolution of Speech, Thought, and Selfless Behavior. Cambridge, MA: Harvard University Press. Lombardi Vallauri, Edoardo (1999). Spiegare gli universali linguistici. In Logos dell’essere, logos della norma, Luigi Lombardi Vallauri (ed.), 711–747. Bari: Adriatica. Medawar, Peter Brian (1967). The Art of Soluble. London: Methuen. Medawar, Peter Brian and Jean S. Medawar (1983). Aristotle to Zoos: A Philosophical Dictionary of Biology. Cambridge, MA: Harvard University Press. Montalbetti, Mario and Ken Wexler (1985). Binding is Linking. In Proceedings of the Western Conference on Formal Linguistics 4: 225–245. Morgan, James L. (1986). From Simple Input to Complex Grammar. Cambridge, MA: MIT Press. Moro, Andrea (1988). Introduzione. In Linguaggio e problemi della conoscenza, Italian translation of Chomsky (1988), VII–XIII. Bologna: Il Mulino, 1992. — (2002). “Linguistica mendeliana”, ovvero quali domande su genetica e grammatica? Lingue e linguaggio 1: 39–58. Piattelli Palmarini, Massimo (1990). Evoluzione, selezione e cognizione: Dall’apprendimento al posizionamento di interruttori mentali. Sistemi intelligenti II (3): 277–322. Piattelli Palmarini, Massimo (ed.) (1979). Théories du langage, théories de l’apprentissage: Le débat entre Jean Piaget et Noam Chomsky. Paris: Editions du Seuil. Pinker, Stephen (1984). Language learnability and language development. Cambridge, MA: Harvard University Press. — (1994) The Language Instinct: How the Mind Creates Language. New York: Morrow. Pullum, Geoffrey K. and Barbara C. Scholz (2002). Empirical assessment of stimulus poverty arguments. The Linguistic Review 19 (1–2): 9–50. Putnam, Hilary (1967). The ‘innateness hypothesis’ and explanatory models in linguistics. Synthese 17: 12–22. Rizzi, Luigi (1993). Some notes on linguistic theory and language development: The case of root infinitives. Language Acquisition 3: 371–393. — (1994). Early null subjects and root null subjects. In Language Acquisition Studies in Generative Grammar, Teun Hoekstra and Bonnie Schwartz (eds.), 371–393. Amsterdam: Benjamins. — (2000). Comparative Syntax and Language Acquisition. London and New York: Routledge. Sampson, Geoffrey R. (2002). Exploring the richness of the stimulus. The Linguistic Review 19 (1–2): 73–104. Slobin, Dan (1975). On the nature of talk to children. In Foundations of Language Development, Eric H. Lenneberg and Elizabeth Lenneberg (eds.), 283–297. New York, Academic Press. Tomasello, Michael (2000a). Do young children have adult syntactic competence? Cognition 74: 209–253. — (2000b). First steps toward a usage-based theory of language acquisition. Cognitive Linguistics 11 (1/2): 61–82. — (2003). Constructing a Language: A Usage-based Theory of Language Acquisition. Cambridge, MA: Harvard University Press. Wexler, Kenneth and Yu-Chin Chien (1985). The development of lexical anaphors and pronouns. Papers and Reports on Child Language Development 24: 138–149. Wexler, Kenneth and Peter W. Culicover (1980). Formal Principles of Language Acquisition. Cambridge, MA: MIT Press. The relation between mind and language 387 Wexler, Kenneth. and Rita Manzini (1987). Parameters and learnability in binding theory. In Parameter Setting, Thomas Roeper and Edwin Williams (eds.), 41–76. Dordrecht: Reidel.
© Copyright 2026 Paperzz