The evolutionary biologist Eörs Szathmáry observed that linguists “would rather share each other’s toothbrush than each other’s terminology.”1 This is far from an isolated view. Peter Hagoort, an eminent cognitive neuroscientist, voiced similar concerns.
When even at major linguistics conferences the program contains presentation[s] such as “A short tour through the minefield of linguistic terminology,” one should realize that this state of affairs is a serious threat to the influence that linguists exert on the research agenda of cognitive science broadly.2
Tanja Kupisch and Jason Rothman, psycholinguists working primarily on bilingual development, recently noted that “[o]nce offered to the public domain, terminology can have far-reaching and long-lasting effects, even—perhaps especially—when these are unintended by their original promoters.”3
An effort to improve the terminological clarity and coherence of theoretical and experimental linguistics is long overdue. In this respect, linguists might consider following the lead of psychologists in identifying and discussing lists of inaccurate, ambiguous, misused, and polysemous terms.4 The focus throughout should be on key notions of the field. Terms such as feature, parameter, (grammaticality/acceptability) judgment, (language) universal, and Universal Grammar are omnipresent in linguistics.5 These notions are fundamental to the discipline and their misuse has important implications, not only for the coherence of the field, but also for its standing in the broader context of cognitive science. The following terms do not all satisfy the same criteria of inaccuracy, ambiguity, and misuse. The degrees to which they exhibit these characteristics vary, and this is part of the problem. This review will not focus solely on the conceptual clarity of these ten terms, but also on their inconsistent usage.
1. Universal Grammar and Language Universals
In Noam Chomsky’s work, Universal Grammar (UG) is the source of our innate ability to acquire and use a natural language.6 The faculty of language consists of certain principles that are innate, common to our species, and available prior to any systematic exposure to a given language.7 UG is sometimes identified with linguistic universals; but this is a mistake. When Chomsky talks about language universals, he does not refer to properties that are universally attested to in all languages, but to computational properties of the mind that are universal because they arise from a species-universal innate ability.8 A property P in a given language can reflect a universal computational principle even if P is not attested to in another language.9
There is no reason to assume that linguistic universals, understood as properties that are shared across languages, are necessarily derived from UG. Although most languages settle on a consistent word order,10 this preference does not reflect the imperatives of UG. Infants process and reshape input in a way that promotes the regularization of harmonic patterns.11 In fact, as Chomsky has argued, language development is guided by principles of general cognition, which work together with linguistic experience and UG.12 Given that a complete list of all the UG principles has not been compiled, the possibility that these principles are, for the most part, not language-specific, but have cognitive, third factor roots, cannot be ruled out.
2. Parameter
In its early days, UG was thought to consist of principles, which are invariant across languages, and a finite set of open parameters that are fixed by experience. The head-directionality parameter, for example, determines the position of the defining element of a phrase in relation to its complements. Some languages such as English are head-initial; others, such as Japanese, head-final. Setting one parameter enables predictions about the others. The original aim was to capture the panorama of cross-linguistic variation while positing as few parameters as possible.13 How many are a few? Twenty? Eighty? 1,534? 1,465,462? Nobody has an answer to this question, and I suspect nobody ever will. Over the last two decades, the term “parameter” has been applied to every possible point of variation across languages, eventually giving rise to parameters of various sizes and domains of prediction—macroparameters, mesoparameters, and nanoparameters. This terminological fluidity is not innocent. Apart from having stripped the notion of a parameter of its original substance, it endowed the term with various custom-made definitions and objects of inquiry. Anybody can postulate a new nanoparameter, even on the basis of one lexical item. In such cases, the term “parameter” is still used despite it no longer carrying the same crucial implications as in its original context. Microparameters, for instance, make no robust, whole-grammar predictions of setability. Yet we cannot afford to dispose of these predictions. The notion that UG consists of parameters, which provide a guide into the variation space, relied on these predictions. One cannot simply use the term “parameter,” or variants of the term, with an arbitrary meaning.
Terminological fluidity is not without consequences. In the case of the term “parameter,” its misuse over the years raised serious concerns about the biological plausibility of an innate endowment for language that consists of millions of minimal points of variation. “If the number of parameters,” Frederick Newmeyer observed, “needed to handle the different grammars of the world’s languages, dialects, and (possibly) idiolects is in the thousands (or, worse, millions), then ascribing them to an innate UG to my mind loses all semblance of plausibility.”14
3. Feature
If parameters come with various definitions and are often custom made, so are features. “Custom made” refers to proposals of the following form: feature A is proposed for language B on the basis of an item/structure C. In such proposals, the postulated primitive is purely data driven and designed specifically to fit a portion of the data.15 As partial descriptions of linguistic objects,16 features can be anything from roundedness in phonology to the extended projection principle (EPP) in syntax.17 These are both very different. Calling almost anything a feature creates both disciplinary and interdisciplinary concerns.
Where do features come from? For some scholars, the answer is UG itself: “The study of the feature inventory of UG requires a massive database compiled on the basis of detailed studies of particular grammars.”18 Ascribing an inordinate number of primitives to UG is not an innocent move. It should not be a desirable move either, because this would mean that the basic atoms of linguistics are not derived, but postulated as innate. Linguists propose new features at will, but there is no theory of UG that can explain the existence of the requisite number of primitives. How many exactly? Again, there is no answer, because anything can be a partial description of a linguistic object. As Norbert Hornstein has pointed out, the real problem is that we have no hint of a theory of features.
Postulating the existence of a feature A in order to explain a structure B does not mean that we have shed light on our biological ability for language.19 Such inventions come at a high cost.20 Both field-internal terminological coherence and field-external visibility are diminished as a result. Using the same term to talk about very different things inevitably reduces internal coherence and consistency. Meanwhile, a biologist would have a hard time understanding what kind of primitives are classified as features in linguistics. This is problematic under the view that features are part of UG—our biological endowment for language. One could, of course, say that “features are properties of syntactic atoms,” whereby “a feature [plural] for example is used analogously to chemists’ use of H for the real-world thing hydrogen.”21 This analogy is unpersuasive. The definition of a chemical element is specific and unambiguous, made exclusively on the basis of the number of protons within its nucleus. But linguists use the EPP in different ways across research groups.22 Additionally, we know exactly how many chemical elements have been discovered or created: 118. We can and do keep a count. There is a consensus that this is the right number; it is not a matter of one’s definition or viewpoint. The same claim cannot be made for features.
4. The Linguistic Genotype
Linguists often talk about the genetic endowment for language; and this suggests that a part of the human genome is dedicated to language. There is no dispute that our species is biologically language-ready. This does not justify the claim that a portion of our genome is dedicated to language. According to Ana Villar and Antoni Gomila:
While it is obvious that any human mental capability requires a genetic make-up as long as it is a biological phenomenon, it cannot be simply assumed that there is going to be some part of the genome specialized for every mental faculty, or for language in particular. Not only because genes can be pleiotropic (i.e., involved in very different processes), but also because they don’t work in isolation.23
A direct link between genetic makeup and linguistic phenotypic traits is untenable. First, genes do not code for grammatical properties. Second, a set of genes can affect multiple, potentially unrelated, phenotypes so that no direct, one-to-one relation exists between genes and phenotypic traits. Third, genes are not the only players in ontogenetic processes. Epigenetic factors also regulate behavioral functions.24
Chomsky uses the term “genetic endowment.” But he is also one of the few linguists to acknowledge that working out the process of genetically fixing the values of linguistic primitives is a far from trivial task. Spelling out the right version of innateness that underlies our ability to do language is an enormous challenge. Hornstein rightly suggested that linguists interested in meeting it would effectively be biologists, regardless of their success.25
5. Faculty of Language in the Narrow Sense
In 2002, Chomsky, Marc Hauser, and W. Tecumseh Fitch introduced a distinction between the language faculty in a narrow and broad sense (FLN and FLB, respectively).26 FLN was described as the “only uniquely human component” of the language faculty, while FLB included “a sensory-motor system, a conceptual-intentional system, and the computational mechanisms for recursion, providing the capacity to generate an infinite range of expressions from a finite set of elements.”27 Although FLB is clearly defined, and some of its components are unambiguously identified, FLN is harder to pin down.
Chomsky, Hauser, and Fitch defined FLN as a component unique to human language, possibly consisting only of recursion. In 2005, they suggested that
[t]he contents of FLN are to be empirically determined, and could possibly be empty, if empirical findings showed that none of the mechanisms involved are uniquely human or unique to language, and that only the way they are integrated is specific to human language.28
In 2009, Fitch then argued that “FLN was intended to have a considerably narrower scope, perhaps even denoting an empty set, but has been read simply as ‘language’ by some and ‘Universal Grammar’ by others.”29 This last view points to the terminological confusion that arose in the various readings of the term “FLN.” This confusion is unfortunate because the original distinction was “intended as a terminological aid to interdisciplinary discussion and rapprochement.”30
The identification of FLN and UG is wrong. If UG equals FLN, and if FLN is, indeed, an empty set—a possibility that Chomsky has once again left open in his latest book with Robert Berwick31—scholars outside generative linguistics would inevitably question the need to assume a UG-shaped form of innateness. Furthermore, if UG and FLN are indeed the same, why are two terms needed to denote one object?
But, of course, FLN is not the same thing as UG. FLN is unique to humans and unique to language. UG does not have this character.32 UG is relevant to both FLN’s language-specific properties, and FLB’s non-language-specific properties. Clearly, FLN is not tantamount to UG. Any suggestions to the contrary represent a misuse of two important terms in linguistics. It is a mistake that many linguists have made.33
6. Hardwired
Aside from inborn reflexes, remarkably few abilities are hardwired in humans.34
Linguists have talked about hardwired UG,35 hardwired principles and parameters,36 hardwired syntactic categories,37 hardwired Merge,38 hardwired semantic relations,39 and hardwired systems of lexicalizable concepts.40 According to common assumptions, something hardwired is both pre-programmed and behaviorally inflexible. Is the development of our innate ability for language behaviorally inflexible? Do children develop language in a way that is not affected by the environment? Feral children do not receive linguistic input in infancy. What impedes their development of innate syntactic categories? How can we meaningfully talk about a critical period for first language acquisition if a great deal of linguistic machinery is hardwired and will be inexorably manifested?
Few linguistic concepts are hardwired. It is not entirely clear that the brain itself is hardwired. This is evident from the lack of consensus on a number of critical issues, such as the lifetime of neuronal spines.41
7. The Metaphors of Language Development
In Aspects of the Theory of Syntax, Chomsky argued that
[l]inguistic theory is concerned primarily with an ideal speaker-listener, in a completely homogeneous speech-community, who knows its language perfectly and is unaffected by such grammatically irrelevant conditions as memory limitations, distractions, shifts of attention and interest, and errors (random or characteristic) in applying his knowledge of the language in actual performance.42
In his later work, Chomsky introduced the notion of instantaneous acquisition.43 This is a metaphor for an idealized version of acquisition, one that abstracts away from actual developmental stages on the assumption that, barring extreme cases, these stages are uniform and have no impact on the acquired grammar.44 Chomsky took care to highlight that these are idealizations.45
Eventually, Chomsky’s completely homogeneous speech community and ideal speaker became embodied in the idea that, absent a severe pathology, adult performance is essentially homogeneous.46 This is demonstrably false, since cognitive phenotypes are not always binary. Individuals with a pathogenic variant of a gene can be impaired in different ways. This may lead to different cognitive phenotypes, at times not even reaching a cutoff point where the diagnosis of a specific pathology can be made.47 Variation can be found in the absence of any pathology even among speakers of the same language, and even within a speaker past the acquisition period.
Over the last few decades, some linguists converted these idealizations into facts. Consider the claim of linguist Guglielmo Cinque:
The problem of parameter setting is usually posed in the context of the simplifying assumption that acquisition is instantaneous (cf. Chomsky 1975, pp. 119–22). My point is that such an assumption may not be just a convenient (and innocuous) idealization. It may well prove to be the only pertinent way to approach the question. Considerations of the actual stages of acquisition … run the risk of distorting matters.48
Cinque suggests that we can approach the question of language acquisition by focusing on primitives he takes to be innately available, while ignoring both the way acquisition unfolds and the role of the environment. The problem is that the metaphor of instantaneous acquisition assumes that innateness and the environment are fully separable. But innateness does not work alone because genes do not work alone. Development is rarely hardwired,49 and language acquisition cannot be an instantaneous, almost reflex-like realization of innate structures and primitives. The metaphors of language development were once useful. It is through their subsequent use that they became unhelpful. Catchy metaphors, of course, make for popular quotes that conveniently express half-truths. And it is for this reason that they should never be taken too seriously or treated as dogma. According to Norbert Wiener and Arturo Rosenblueth, “the price of metaphor is eternal vigilance.”50
8. Grammaticality Judgment
Despite being frequently used, “grammaticality judgment” is a problematic term. Some linguists have argued that acceptability and grammaticality are distinct;51 others have argued the reverse. Do the terms “acceptability” and “grammaticality” refer to the same thing? If they do, why don’t we just settle on one term? If they do not, do speakers provide different kinds of answers when asked about the grammaticality or the acceptability of a sentence in their native language?
A grammatical sentence conforms to the rules of a given language. An acceptable sentence refers to a speaker’s perception of the stimulus in terms of her own linguistic repertoire.52 There are sentences that are not acceptable or in use—but for reasons that have nothing to do with violating linguistic rules. The sentence, “That that that Bill left Mary amused Sam is interesting is sad,” is a grammatical sentence of English.53 Native speakers would not produce it often and they would certainly not rate it as acceptable as the very similar, “It is sad that it is interesting that it amused Sam that Bill left Mary.”54
If grammaticality also reflects cognitive biases, asking a speaker to provide grammaticality judgments means asking for judgments about the interactions of all linguistic and cognitive factors that determine the limits of grammar. No speaker can provide this. Nor can any linguist. The term “grammaticality judgments” has been correctly labelled a misnomer. Speakers have intuitions only about their perception of linguistic stimuli.55
9. Bilingual Advantage
The effort of mentally juggling two linguistic systems can be linked to better performance in certain cognitive tasks.56 Thus the bilingual advantage. The term itself is neither ambiguous, problematic, nor misused. But its use calls for caution. As various scholars tested an ever-growing number of populations in order to discover the magnitude of the bilingual advantage, it became clear that one can also talk about a bilingual disadvantage.57 A more useful term that emerged in the same literature is “trade-off.”58 In this sense, an advantage in one measure may bring along other effects, hence the term “bilingual advantage” might only provide half of the picture with respect to bilingual effects on cognition. Talking about the overall impact or effect of bilingualism on cognition may come closer to acknowledging the surrounding implications.
10. Optimal Design and Perfect System
Hauser, Chomsky, and Fitch suggest that FLN can be thought of as “a kind of ‘optimal solution’ to the problem of linking the sensory-motor and conceptual-intentional systems,”59 and of satisfying the interface conditions between FLN and FLB. This view is also present in Chomsky’s earlier work under the name “strong minimalist thesis” (SMT).60 In other work, the language system has been described as perfect.61 These notions are elusive. Most scholars understand them as denoting principles of parsimony, economy, and elegance,62 but the way these properties are attributed to language is unclear. Do parsimony, economy, and elegance characterize language, or theories about language?63
Language develops some of its characteristics by adaptation to environmental triggers. There is evidence that over time, speakers develop effort–accuracy trade-offs.64 Languages are also reshaped to become more learnable.65 One might say that the language system is becoming more optimal in response to adaptation pressures. This trend toward efficiency is probably not due to the language system itself. The human mind has been described as making optimal use of its cognitive resources across domains.66 It is unclear whether the optimality of language is due to the recruitment of general principles of rationality and adaptation, or whether language, possibly by being unique at some level, stands out from other cognitive modules.
It is harder to argue that all linguistic theories are optimal, especially in relation to the preceding suggestions that innateness consists of an unknown, ever-growing number of features, parameters, and other primitives.
Taking Stock
The objective of this essay is to attain a higher level of terminological clarity and coherence within the field of linguistics. If successful, this process may help improve its visibility in neighboring fields, such as other parts of psychology, biology, and neuroscience. This visibility is, in fact, extremely important. As scientists of language, linguists should offer robust and cohesive theories and this is particularly true for claims about innateness, which must make sense from a biological point of view. Some months ago, I attended a conference that dealt with certain linguistic primitives. A talk was given by a senior linguist who argued for the removal of a specific primitive from UG. UG was mentioned at the beginning of the talk and never referred to again. Nor was it explained why the feature was allocated to UG in the first place. UG, it should be stressed, is defined as our biological endowment for language, and claims made about UG in linguistics must be plausible in biology. I asked the speaker how the feature in question came to be part of UG. I was told that this had been a common assumption among theoretical linguists since the 1970s. I then asked how theories of UG might support such a claim. I found it hard to see how a biologist would be convinced that the biological endowment for language encodes all of these grammatical features. Defending our theories from interdisciplinary criticisms, I suggested, might be extremely difficult. To my surprise, the speaker responded that he had no interest in interdisciplinary work. These considerations, I realized, simply did not enter the picture. For the speaker, UG is simply a repository of linguistic primitives that can be disconnected from human biology.
This is why terminological clarity matters.