comprehension with regard to language, understanding what others say (or sign or write)
What is the average kindergartener almost as good at as you are? Not much, with one important exception: using language. By 5 years of age, most children have mastered the basic structure of their native language or languages (the possibility of bilingualism is to be assumed whenever we refer to “native language”), whether spoken or manually signed. Although their vocabulary and powers of expression are less sophisticated than yours, their sentences are as grammatically correct as those produced by the average college student. This is a remarkable achievement.
217
production with regard to language, speaking (or writing or signing) to others
Language use requires comprehension, which refers to understanding what others say (or sign or write), and production, which refers to actually speaking (or signing or writing). As we have observed for other areas of development, infants’ and young children’s ability to understand precedes their ability to produce. Children understand words and linguistic structures that other people use months or even years before they include them in their own utterances. This is, of course, not unique to young children; you no doubt understand many words that you never actually use. In our discussion, we will be concerned with the developmental processes involved in both comprehension and production, as well as the relation between them.
generativity refers to the idea that through the use of the finite set of words and morphemes in humans’ vocabulary, we can put together an infinite number of sentences and express an infinite number of ideas
How do languages work? Despite the fact that there are thousands of human languages, they share overarching similarities. All human languages are similarly complex, with different pieces combined at different levels to form a hierarchy: sounds are combined to form words, words are combined to form sentences, and sentences are combined to form stories, conversations, and other kinds of narratives. Children must acquire all of these facets of their native language. The enormous benefit that emerges from this combinatorial process is generativity; by using the finite set of words in our vocabulary, we can generate an infinite number of sentences, expressing an infinite number of ideas.
phonemes the elementary units of meaningful sound used to produce languages
However, the generative power of language carries a cost for young language learners: they must deal with its complexity. To appreciate the challenge presented to children learning their first language, imagine yourself as a stranger in a strange land. Someone walks up to you and says, “Jusczyk daxly blickets Nthlakapmx.” You would have absolutely no idea what this person had just said. Why?
phonological development the acquisition of knowledge about the sound system of a language
First, you would probably have difficulty perceiving some of the phonemes that make up what the speaker is uttering. Phonemes are the units of sound in speech; a change in phoneme changes the meaning of a word. For example, “rake” and “lake” differ by only one phoneme (/r/versus/l/), but the two words have quite different meanings to English speakers. Different languages employ different sets of phonemes; English uses just 45 of the roughly 200 sounds found across the world’s languages. The phonemes that distinguish meaning in any one language overlap with, but also differ from, those in other languages. For example, the sounds/r/and/l/are a single phoneme in Japanese, and do not carry different meanings. Furthermore, combinations of sounds that are common in one language may never occur in others. When you read the stranger’s utterance in the preceding paragraph, you probably had no idea how to pronounce the word Nthlakapmx, because some of its sound combinations do not occur in English (though they do occur in other languages). Thus, the first step in children’s language learning is phonological development: the mastery of the sound system of their language.
morphemes the smallest units of meaning in a language, composed of one or more phonemes
Another reason you would not know what the stranger had said to you, even if you could have perceived the sounds being uttered, is that you would have had no idea what the sounds mean. The smallest units of meaning are called morphemes. Morphemes, alone or in combination, constitute words. The word dog, for example, contains one morpheme. The word dogs contains two morphemes, one designating a familiar furry entity (dog) and the second indicating the plural (-s). Thus, the second component in language acquisition is semantic development, that is, learning the system for expressing meaning in a language, including word learning.
218
semantic development the learning of the system for expressing meaning in a language, including word learning
However, even if you were told the meaning of each individual word the stranger had used, you would still not understand the utterance unless you knew how words are put together in the stranger’s language. To express an idea of any complexity, we combine words into sentences, but only certain combinations are allowed in any given language. Syntax refers to the permissible combinations of words from different categories (nouns, verbs, adjectives, etc.). In English, for example, the order in which words can appear in a sentence is crucial: “Lila ate the lobster” does not mean the same thing as “The lobster ate Lila.” Other languages indicate which noun did the eating and which noun was eaten by adding morphemes like suffixes to the nouns. For example, a Russian noun ending in “a” is likely to refer to the entity doing the eating, while the same noun ending in “u” is likely to refer to the thing that was eaten. The third component in language learning, then, is syntactic development, that is, learning how words and morphemes are combined.
syntax rules in a language that specify how words from different categories (nouns, verbs, adjectives, and so on) can be combined
Finally, a full understanding of the interaction with the stranger would necessitate having some knowledge of the cultural rules and contextual variations for using language. In some societies, for example, it would be quite bizarre to be addressed by a stranger in the first place, whereas in others it would be commonplace. You would also need to know how to go beyond the speaker’s specific words to understand what the speaker was really trying to communicate—to use factors such as the context and the speaker’s emotional tone to read between the lines and to learn how to hold a conversation. Acquiring an understanding of how language is typically used is referred to as pragmatic development.
syntactic development the learning of the syntax of a language
Our example of the bewilderment one experiences when listening to an unfamiliar language is useful for delineating the components of language use. However, when we, as adults, hear someone speaking an unfamiliar language, we already know what a language is. We know that the sounds the person is uttering constitute words, that words are combined to form sentences, that only certain combinations are acceptable, and so on. In other words, in contrast with young language learners, adults have considerable metalinguistic knowledge—that is, knowledge about language and its properties.
pragmatic development the acquisition of knowledge about how language is used
Thus, learning language involves phonological, semantic, syntactic, and pragmatic development, as well as meta-linguistic knowledge. The same factors are involved in learning a sign language, in which the basic linguistic elements are gestures rather than sounds. There are more than 200 languages, including American Sign Language (ASL), that are based on gestures, both manual and facial. They are true languages and are as different from one another as spoken languages are. The course of acquisition of a sign language is remarkably similar to that of a spoken language.
metalinguistic knowledge an understanding of the properties and function of language—that is, an understanding of language as language
What does it take to be able to learn a language in the first place? Full-fledged language is achieved only by humans, so, obviously, one requirement is the human brain. But a single human, isolated from all sources of linguistic experience, could never learn a language; hearing (or seeing) language is a crucial ingredient for successful language development.
219
The key to full-fledged language development lies in the human brain. Language is a species-specific behavior: only humans acquire language in the normal course of development. Furthermore, it is species-universal: language learning is achieved by typically developing infants across the globe.
In contrast, no other animals naturally develop anything approaching the complexity or generativity of human language, even though they can communicate with one another. For example, birds claim territorial rights by singing (Marler, 1970), and vervet monkeys can reveal the presence and identity of predators through specific calls (Seyfarth & Cheney, 1993).
Researchers have had limited success in training nonhuman primates to use complex communicative systems. One early effort was an ambitious project in which a dedicated couple raised a chimpanzee (Vicki) with their own children (Hayes & Hayes, 1951). Although Vicki learned to comprehend some words and phrases, she produced virtually no recognizable words. Subsequent researchers attempted to teach nonhuman primates sign language. Washoe, a chimpanzee, and Koko, a gorilla, became famous for their ability to communicate with their human trainers and caretakers using manual signs (Gardner & Gardner, 1969; Patterson & Linden, 1981). Washoe could label a variety of objects and could make requests (“more fruit,” “please tickle”). The general consensus is that, however impressive Washoe’s and Koko’s “utterances” were, they do not qualify as language, because they contained little evidence of syntactic structure (Terrace et al., 1979; Wallman, 1992).
The most successful sign-learning nonhuman is Kanzi, a great ape of the bonobo species. Kanzi’s sign-learning began when he observed researchers trying to teach his mother to communicate with them by using a lexigram board, a panel composed of a few graphic symbols representing specific objects and actions (“give,” “eat,” “banana,” “hug,” and so forth) (Savage-Rumbaugh et al., 1993). Kanzi’s mother never caught on, but Kanzi did, and over the years his lexigram vocabulary increased from 6 words to more than 350. He is now very adept at using his lexigram board to answer questions, to make requests, and even to offer comments. He often combines signs, but whether they can be considered syntactically structured sentences is not clear.
There are also several well-documented cases of nonprimate animals that have learned to respond to spoken language. Kaminski, Call, and Fischer (2004) found that Rico, a border collie, knew more than 200 words and could learn and remember new words using the same kinds of processes that toddlers use. Alex, an African Grey parrot, learned to produce and understand basic English utterances, although his skills remained at a toddler level (Pepperberg, 1999).
Whatever the ultimate decision regarding the extent to which trained nonhuman animals should be credited with language, several things are clear. Even their most basic linguistic achievements come only after a great deal of concentrated human effort, whereas human children master the rudiments of their language with little explicit teaching. Furthermore, while the most advanced nonhuman communicators do combine symbols, their utterances show limited evidence of syntactic structure, which is a defining feature of language (Tomasello, 1994). In short, only the human brain acquires a communicative system with the complexity, structure, and generativity of language. Correspondingly, we humans are notoriously poor at learning the communicative systems of other species (Harry Potter’s ability to speak Parseltongue with snakes aside). There is an excellent match between the brains of animals of different species and their respective communicative systems.
220
Brain–language relations A vast amount of research has examined the relationship between language and brain function. It is clear that language processing involves a substantial degree of functional localization. At the broadest level, there are hemispheric differences in language functioning that we discussed to some extent in Chapter 3. For the 90% of people who are right-handed, language is primarily represented and controlled by the left hemisphere.
Left-hemisphere specialization appears to emerge very early in life. Studies using neuroimaging techniques have demonstrated that newborns and 3-month-olds show greater activity in the left hemisphere when exposed to normal speech than when exposed to reversed speech or silence (Bortfeld, Fava, & Boas, 2009; Dehaene-Lambertz, Dehaene, & Hertz-Pannier, 2002; Pena et al., 2003). In addition, EEG studies show that infants exhibit greater left-hemisphere activity when listening to speech but greater right-hemisphere activity when listening to nonspeech sounds (Molfese & Betz, 1988). An exception to this pattern of lateralization occurs in the detection of pitch in speech, which in infants, as in adults, tends to involve the right hemisphere (Homae et al., 2006).
Although it is evident that the left hemisphere predominantly processes speech from birth, the reasons for this are not yet known. One possibility is that the left hemisphere is innately predisposed to process language but not other auditory stimuli. Another possibility is that speech is localized to the left hemisphere because of its acoustic properties. In this view, the auditory cortex in the left hemisphere is tuned to detect small differences in timing, whereas the auditory cortex in the right hemisphere is tuned to detect small differences in pitch (e.g., Zatorre et al., 1992; Zatorre & Belin, 2001; Zatorre, Belin, & Penhune, 2002). Because speech turns on small differences in timing (as you will see when we discuss voice onset time), it may be a more natural fit for the left hemisphere.
critical period for language the time during which language develops readily and after which (sometime between age 5 and puberty) language acquisition is much more difficult and ultimately less successful
Critical period for language development If you were to survey your classmates who have studied another language, we predict you would discover that those who learned a foreign language in adolescence found the task to be much more challenging than did those who learned the foreign language in early childhood. A considerable body of evidence suggests that, in fact, the early years constitute a critical period during which language develops readily. After this period (which ends sometime between age 5 and puberty), language acquisition is much more difficult and ultimately less successful.
Relevant to this hypothesis, there are several reports of children who barely developed language at all after being deprived of early linguistic experience. The most famous case in modern times is Genie, who was discovered in appalling conditions in Los Angeles in 1970. From the age of approximately 18 months until she was rescued at age 13, Genie’s parents kept her tied up and locked alone in a room. During her imprisonment, no one spoke to her; when her father brought her food, he growled at her like an animal. At the time of her rescue, Genie’s development was stunted—physically, motorically, and emotionally—and she could barely speak. With intensive training, she made some progress, but her language ability never developed much beyond the level of a toddler’s: “Father take piece wood. Hit. Cry” (Curtiss, 1977, 1989; Rymer, 1993).
221
Does this extraordinary case support the critical-period hypothesis? Possibly, but it is difficult to know for sure. Genie’s failure to develop full, rich, language after her discovery might have resulted as much from the bizarre and inhumane treatment she suffered as from linguistic deprivation.
Other areas of research provide much stronger evidence for the critical-period hypothesis. As noted in Chapter 3, adults, who are well beyond the critical period, are more likely to suffer permanent language impairment from brain damage than are children, presumably because other areas of the young brain (but not the older brain) are able to take over language functions (see M. H. Johnson, 1998). Moreover, adults who learned a second language after puberty use different neural mechanisms to process that language than do adults who learned their second language from infancy (e.g., Kim et al., 1997; Pakulak & Neville, 2011). These results strongly suggest that the neural circuitry supporting language learning operates differently (and better) during the early years.
In an important behavioral study, Johnson and Newport (1989) tested the English proficiency of Chinese and Korean immigrants to the United States who had begun learning English either as children or as adults. The results, shown in Figure 6.1, reveal that knowledge of key aspects of English grammar was related to the age at which these individuals began learning English, but not to the length of their exposure to the language. The most proficient were those who had begun learning English before the age of 7.
A similar pattern has been described for first-language acquisition in the Deaf community: individuals who acquired ASL as a first language when they were children become more proficient signers than do individuals who acquired ASL as a first language as teens or adults (Newport, 1990). Johnson and Newport also observed a great deal of variability among “late learners”—those who were acquiring a second language, or a sign language as their first formal language, at puberty or beyond. As in the findings we predicted for your survey of classmates, some individuals achieved nativelike skills, whereas the language outcomes for others were quite poor. For reasons that are still unknown, some individuals continue to be talented language learners even after puberty, while most do not.
Newport (1990) proposed an intriguing hypothesis to explain these results and, more generally, to explain why children are usually better language learners than adults. According to her “less is more” hypothesis, perceptual and memory limitations cause young children to extract and store smaller chunks of the language than adults do. Because the crucial building blocks of language (the meaning-carrying morphemes) tend to be quite small, young learners’ limited cognitive abilities may actually facilitate the task of analyzing and learning language.
The evidence for a critical period in language acquisition has some very clear practical implications. For one thing, deaf children should be exposed to sign language as early as possible. For another, foreign-language exposure at school, discussed in Box 6.1, should begin in the early grades in order to maximize children’s opportunity to achieve native-level skills in a second language.
Possession of a human brain is not enough for language to develop. Children must also be exposed to other people using language—any language, signed or spoken. Adequate experience hearing others talk is readily available in the environment of almost all children around the world. Much of the speech directed to infants occurs in the context of daily routines—during thousands of mealtimes, diaper changes, baths, and bedtimes, as well as in countless games like peekaboo and nursery rhymes like the “Itsy Bitsy Spider.”
222
bilingualism the ability to use two languages
The topic of bilingualism, the ability to use two languages, has attracted substantial attention in recent years as increasing numbers of children are developing bilingually. Indeed, almost half the children in the world are regularly exposed to more than one language, often at home with parents who speak different languages. Remarkably, despite the fact that bilingual children have twice as much to learn as monolingual children, they show little confusion and language delay. In fact, there is evidence to suggest that being bilingual improves aspects of cognitive functioning in childhood and beyond.
Bilingual learning can begin in the womb. Newborns prenatally exposed to just their native language prefer it over other languages, whereas newborns whose mothers spoke two languages during pregnancy prefer both languages equally (Byers-Heinlein, Burns, & Werker, 2010). Bilingual infants are able to discriminate the speech sounds of their two languages at roughly the same pace that monolingual infants distinguish the sounds of their one language (e.g., Albareda-Castellot, Pons, & Sebastián-Gallés, 2011; Sundara, Polka, & Molnar, 2008). How might this be, given that bilingual infants have twice as much to learn? One possibility is that bilinguals’ attention to speech cues is heightened relative to that of monolinguals. For example, bilingual infants are better than monolingual infants at using purely visual information (a silent talking face) to discriminate between unfamiliar languages (Sebastián-Gallés et al., 2012).
For the most part, children who are acquiring two languages do not seem to confuse them; indeed, they appear to build two separate linguistic systems. When language mixing does occur, it usually reflects a gap of knowledge in one language that the child is trying to fill in with the other, rather than a confusing of the two language systems (e.g., Deuchar & Quay, 1999; Paradis, Nicoladis, & Genesee, 2000).
Children developing bilingually may appear to lag behind slightly in each of their languages because their vocabulary is distributed across two languages (Oller & Pearson, 2002). That is, a bilingual child may know how to express some concepts in one language but not the other. However, both the course and the rate of language development are generally very similar for bilingual and monolingual children (Genesee & Nicoladis, 2006). And as we have noted, there are cognitive benefits to bilingualism: children who are competent in two languages perform better on a variety of measures of executive function and cognitive control than do monolingual children (Bialystok & Craik, 2010; Costa, Hernandez, & Sebastián-Gallés, 2008). Recent results reveal similar effects for bilingual toddlers (Poulin-Dubois et al., 2011). Even bilingual infants appear to show greater cognitive flexibility in learning tasks (Kovács & Mehler, 2009a, b). The link between bilingualism and improved cognitive flexibility likely lies in the fact that bilingual individuals have had to learn to rapidly switch between languages, both in comprehension and production.
More difficult issues arise with respect to second-language learning in the school setting. Some countries with large but distinct language communities, like Canada, have embraced bilingual education. Others, including the United States, have not. The debate over bilingual education in the United States is tied up with a host of political, ethnic, and racial issues. One side of this debate advocates total immersion, in which children are communicated with and taught exclusively in English, the goal being to help them become proficient in English as quickly as possible. The other side recommends an approach that initially provides children with instruction in basic subjects in their native language and gradually increases the amount of instruction provided in English (Castro et al., 2011).
In support of the latter view, there is evidence that (1) children often fail to master basic subject matter when it is taught in a language they do not fully understand; and (2) when both languages are integrated in the classroom, children learn the second language more readily, participate more actively, and are less frustrated and bored (August & Hakuta, 1998; Crawford, 1997; Hakuta, 1999). This approach also helps prevent situations where children might become less proficient in their original language as a result of being taught a second one in school.
223
Infants identify speech as something important very early. When given the choice, newborns prefer listening to speech rather than to artificial sounds (Vouloumanos et al., 2010). Intriguingly, newborns also prefer nonhuman primate (rhesus macaque) vocalizations to nonspeech sounds, and show no preference for speech over macaque vocalizations until 3 months of age (Vouloumanos et al., 2010). These results suggest that infants’ auditory preferences are fine-tuned through experience with human language during their earliest months.
infant-directed speech (IDS) the distinctive mode of speech that adults adopt when talking to babies and very young children
Infant-directed speech Imagine yourself on a bus listening to a stranger who is seated behind you and speaking to someone. Could you guess whether the stranger was addressing an infant or an adult? We have no doubt that you could, even if the stranger was speaking an unfamiliar language. The reason is that in virtually all cultures, adults adopt a distinctive mode of speech when talking to babies and very young children. This special way of speaking was originally dubbed “motherese” (Newport, Gleitman, & Gleitman, 1977). The current term, infant-directed speech (IDS), recognizes the fact that this style of speech is used by both males and females, including parents and nonparents alike. Indeed, even young children adopt it when talking to babies (Shatz & Gelman, 1973).
CHARACTERISTICS OF INFANT-DIRECTED SPEECH The most obvious quality of IDS is its emotional tone. It is speech suffused with affection—“the sweet music of the species,” as Darwin (1877) put it. Another obvious characteristic of IDS is exaggeration. When people speak to babies, their speech is slower, and their voice is often higher pitched, than when they speak to adults, and they swoop abruptly from high pitches to low pitches and back again. Even their vowels are clearer (Kuhl et al., 1997). All this exaggerated speech is accompanied by exaggerated facial expressions. Many of these characteristics have been noted in adults speaking such languages as Arabic, French, Italian, Japanese, Mandarin Chinese, and Spanish (see de Boysson-Bardies, 1996/1999), as well as in deaf mothers signing to their infants (Masataka, 1992).
Beyond expressing emotional tone, caregivers can use various pitch patterns of IDS to communicate important information to infants even when infants don’t know the meaning of the words uttered. For example, a word uttered with sharply falling intonation tells the baby that their caregiver disapproves of something, whereas a cooed warm sound indicates approval. These pitch patterns serve the same function in language communities ranging from English and Italian to Japanese (Fernald et al., 1989). Interestingly, infants exhibit appropriate facial emotion when listening to these pitch patterns, even when the language is unfamiliar (Fernald, 1993).
IDS also seems to aid infants’ language development. To begin with, it draws infants’ attention to speech itself. Indeed, infants prefer IDS to adult-directed speech (Cooper & Aslin, 1994; Pegg, Werker, & McLeod, 1992), even when it is in a language other than their own. For example, both Chinese and American infants listened longer to a recording of a Cantonese-speaking woman talking to a baby in IDS than to the same woman speaking normally to an adult friend (Werker, Pegg, & McLeod, 1994). Some studies suggest that infants’ preference for IDS may emerge because it is “happy speech”; when speakers’ affect is held constant, the preference disappears (Singh, Morgan, & Best, 2002). Perhaps because they pay greater attention to IDS, infants learn and recognize words better when the words are presented in IDS than when they are presented in adult-directed speech (Ma et al., 2011; Singh et al., 2009; Thiessen, Hill, & Saffran, 2005).
224
Although IDS is very common throughout the world, it is not universal. In some cultures, such as the Kwara’ae of the Solomon Islands, the Ifaluk of Micronesia, and the Kaluli of Papua New Guinea, it is believed that because infants cannot understand what is said to them, there is no reason for caregivers to speak to them (Le, 2000; Schieffelin & Ochs, 1987; Watson-Gegeo, & Gegeo, 1986). For example, young Kaluli infants are carried facing outward so that they can engage with other members of the group (but not with their caregiver), and if they are spoken to by older siblings, the mother will speak for them (Schieffelin & Ochs, 1987). Thus, even if they are not addressed directly by their caregivers, these infants are still immersed in language.
That infants begin life equipped with the two basic necessities for acquiring language—a human brain and a human environment—is, of course, only the beginning of the story. Of all the things we learn as humans, languages are arguably the most complex; so complex, in fact, that scientists have yet to be able to program computer systems to acquire a human language. The overwhelming complexity of language is further reflected in the difficulty most people have in learning a new language after puberty. How, then, do infants and young children manage to acquire their native language with such astounding success? We turn now to the many steps through which that remarkable accomplishment proceeds.
Acquiring a language involves listening and speaking (or watching and signing) and requires both comprehending what other people communicate and producing intelligible speech (or signs). Infants start out paying attention to what people say or sign, and they know a great deal about language long before their first linguistic productions.
prosody the characteristic rhythm, tempo, cadence, melody, intonational patterns, and so forth with which a language is spoken
The first step in language learning is figuring out the sounds of one’s native language. As you saw in Chapter 2, the task usually begins in the womb, as fetuses develop a preference for their mother’s voice and the language they hear her speak. The basis for this very early learning is prosody, the characteristic rhythmic and intonation patterns with which a language is spoken. Differences in prosody are in large part responsible for why languages—from Japanese to French to Swahili—sound so different from one another.
Speech perception also involves distinguishing among the speech sounds that make a difference in a given language. To learn English, for example, one must distinguish between bat and pat, dill and kill, Ben and bed. As you will see next, young infants do not have to learn to hear these differences: they perceive many speech sounds in very much the same way that adults do.
categorical perception the perception of speech sounds as belonging to discrete categories
Categorical perception of speech sounds Both adults and infants perceive speech sounds as belonging to discrete categories. This phenomenon, referred to as categorical perception, has been established by studying people’s response to speech sounds. In this research, a speech synthesizer is used to gradually and continuously change one speech sound, such as/b/, into a related one, such as/p/. These two phonemes are on an acoustic continuum; they are produced in exactly the same way, except for one crucial difference—the length of time between when air passes through the lips and when the vocal cords start vibrating. This lag, referred to as voice onset time (VOT), is shorter for/b/(less than 25 milliseconds [ms]) than for/p/(greater than 25 ms). (Try saying “ba” and “pa” alternately several times, with your hand on your throat, and you will likely experience this difference in VOT.)
voice onset time (VOT) the length of time between when air passes through the lips and when the vocal cords start vibrating
225
To study the perception of VOT, researchers create recordings of speech sounds that vary along this VOT continuum, so that each successive sound is slightly different from the one before, with/b/gradually changing into/p/. What is surprising is that adult listeners do not perceive this continuously changing series of sounds (Figure 6.2). Instead, they hear/b/repeated several times and then hear an abrupt switch to/p/. All the sounds in this continuum that have a VOT of less than 25 ms are perceived as/b/, and all those that have a VOT greater than 25 ms are perceived as/p/. Thus, adults automatically divide the continuous signal into two discontinuous categories—/b/and/p/. This perception of a continuum as two categories is a very useful perceptual ability because it allows one to pay attention to sound differences that are meaningful in one’s native language, such as, in English, the difference between/b/and/p/, while allowing meaningless differences, such as the difference between a/b/with a 10 ms VOT versus a/b/with a 20 ms VOT, to be ignored.
Young infants draw the same sharp distinctions between speech sounds. This remarkable fact was established using the habituation technique familiar to you from previous chapters. In the original, classic study (one of the 100 most frequently cited studies in psychology), 1- and 4-month-olds sucked on a pacifier hooked up to a computer (Eimas et al., 1971). The harder they sucked, the more often they’d hear repetitions of a single speech sound. After hearing the same sound repeatedly, the babies gradually sucked less enthusiastically (habituation). Then a new sound was played. If the infants’ sucking rate increased in response to the new sound, the researchers inferred that the infants discriminated the new sound from the old one (dishabituation).
The crucial factor in this study was the relation between the new and old sounds—specifically, whether they were from the same or different phonemic categories. For one group of infants, the new sound was from a different category; thus, after habituation to a series of sounds that adults perceive as/b/, sucking now produced a sound that adults identify as/p/. For the second group, the new sound was within the same category as the old one (i.e., adults perceive them both as/b/). A critical feature of the study is that for both groups, the new and old sounds differed equally in terms of VOT.
226
As Figure 6.3 shows, after habituating to/b/, the infants increased their rate of sucking when the new sound came from a different phonemic category (/p/instead of/b/). Habituation continued, however, when the new sound was within the same category as the original one. Since this classic study, researchers have established that infants show categorical perception of numerous speech sounds (Aslin, Jusczyk, & Pisoni, 1998).
A fascinating outcome of this research is the discovery that young infants actually make more distinctions than adults do. This rather surprising phenomenon occurs because any given language uses only a subset of the large variety of phonemic categories that exist. As noted earlier, the sounds/r/and/l/make a difference in English, but not in Japanese. Similarly, speakers of Arabic, but not of English, perceive a difference between the/k/sounds in “keep” and “cool.” Adults simply do not perceive differences in speech sounds that are not important in their native language, which partly accounts for why it is so difficult for adults to become fluent in a second language.
In contrast, infants can distinguish between phonemic contrasts made in all the languages of the world—about 600 consonants and 200 vowels. For example, Kikuyu infants in Africa are just as good as American babies at discriminating English contrasts not found in Kikuyu (Streeter, 1976). Studies done with infants from English-speaking homes have shown that they can discriminate non-English distinctions made in languages ranging from German and Spanish to Thai, Hindi, and Zulu (Jusczyk, 1997).
This research reveals an ability that is innate, in the sense that it is present at birth, and experience-independent, because infants can discriminate between speech sounds they have never heard before. Being born able to distinguish speech sounds of any language is enormously helpful to infants, priming them to start learning whichever of the world’s languages they hear around them. Indeed, the crucial role of early speech perception is reflected in a positive correlation between infants’ speech-perception skills and their later language skills. Babies who were better at detecting differences between speech sounds at 6 months scored higher on measures of vocabulary and grammar at 13 to 24 months of age (Tsao, Liu, & Kuhl, 2004).
Developmental changes in speech perception During the last months of their first year, infants increasingly home in on the speech sounds of their native language, and by 12 months of age, they have “lost” the ability to perceive the speech sounds that are not part of it. In other words, their speech perception has become adultlike. This shift was first demonstrated by Werker and her colleagues (Werker, 1989; Werker & Lalonde, 1988; Werker & Tees, 1984), who studied infants ranging in age from 6 to 12 months. The infants, all from English-speaking homes, were tested on their ability to discriminate speech contrasts that are not used in English but that are important in two other languages—Hindi and Nthlakapmx (a language spoken by indigenous First Nations people in the Canadian Pacific Northwest). The researchers used a simple conditioning procedure, shown in Figure 6.4. The infants learned that if they turned their head toward the sound source when they heard a change in the sounds they were listening to, they would be rewarded by an interesting visual display. If the infants turned their heads in the correct direction immediately following a sound change, the researchers inferred that they had detected the change.
227
Figure 6.5 shows that at 6 to 8 months of age, English-learning infants readily discriminated between the sounds they heard; they could tell one Hindi syllable from another, and they could also distinguish between two sounds in Nthlakapmx. At 10 to 12 months of age, however, the infants no longer perceived the differences they had detected a few months before. Two Hindi syllables that had previously sounded different to them now sounded the same. Other research indicates that a similar change occurs slightly earlier for the discrimination of vowels (Kuhl et al., 1992; Polka & Werker, 1994). Interestingly, this perceptual narrowing does not appear to be an entirely passive process. Kuhl, Tsao, and Liu (2003) found that infants learned more about the phonetic structure of Mandarin from a live interaction with a Mandarin speaker than from watching a videotape of one.
Is this process of perceptual narrowing limited to speech? To answer this question, a recent study asked whether this narrowing process also occurs in ASL (Palmer et al., 2012). The researchers began by determining whether infants who had never been exposed to ASL were able to discriminate between highly similar ASL signs that are differentiated by the shape of the hand. They found that 4-month-olds could, in fact, discriminate between the signs. However, by 14 months of age, only infants who were learning ASL were able to detect the difference between the hand shapes; those who were not learning ASL had lost their ability to make this perceptual discrimination. Perceptual narrowing is thus not limited to speech. Indeed, this narrowing process may be quite broad; recall the discussion of perceptual narrowing in the domains of face perception and musical rhythm discussed in Chapter 5.
Thus, after the age of 8 months or so, infants begin to specialize in their discrimination of speech sounds, retaining their sensitivity to sounds in the native language they hear every day, while becoming increasingly less sensitive to nonnative speech sounds. Indeed, becoming a native listener is one of the greatest accomplishments of the infant’s first year of postnatal life.
228
word segmentation the process of discovering where words begin and end in fluent speech
As infants begin to tune into the speech sounds of their language, they also begin to discover another crucial feature of the speech surrounding them: words. This is no easy feat. Unlike the words typed on this page, there are no spaces between words in speech, even in IDS. What this means is that most utterances infants hear are strings of words without pauses between them, like “Lookattheprettybaby! Haveyou-everseensuchaprettybabybefore?” They then have to figure out where the words start and end. Remarkably, they begin the process of word segmentation during the second half of the first year.
In the first demonstration of infant word segmentation, Jusczyk & Aslin (1995) used a head-turn procedure designed to assess infants’ auditory preferences. In this study, 7-month-olds first listened to passages of speech in which a particular word was repeated from sentence to sentence—for example, “The cup was bright and shiny. A clown drank from the red cup. His cup was filled with milk.” After listening to these sentences several times, infants were tested using the head-turn preference procedure to see whether they recognized the words repeated in the sentences. In this method, flashing lights mounted near two loudspeakers located on either side of an infant are used to draw the infant’s attention to one side or the other. As soon as the infant turns to look at the light, an auditory stimulus is played through the speaker, and it continues as long as the infant is looking in that direction. The length of time the infant spends looking at the light—and hence listening to the sound—provides a measure of the degree to which the infant is attracted to that sound.
Infants in this study were tested on repetitions of words that had been presented in the sentences (such as cup) or words that had not (such as bike). The researchers found that infants listened longer to words that they had heard in the passages of fluent speech, as compared with words that never occurred in the passages. This result indicates that the infants were able to pull the words out of the stream of speech—a task so difficult that even sophisticated speech-recognition software often fails at it.
How do infants find words in pause-free speech? They appear to be remarkably good at picking up regularities in their native language that help them to find word boundaries. One example is stress patterning, an element of prosody. In English, the first syllable in two-syllable words is much more likely to be stressed than the second syllable (as in “English,” “often,” and “second”). By 8 months of age, English-learning infants expect stressed syllables to begin words and can use this information to pull words out of fluent speech (Curtin, Mintz, & Christiansen, 2005; Johnson & Jusczyk, 2001; Jusczyk, Houston, & Newsome, 1999; Thiessen & Saffran, 2003).
distributional properties the phenomenon that in any language, certain sounds are more likely to appear together than are others
Another regularity to which infants are surprisingly sensitive concerns the distributional properties of the speech they hear. In every language, certain sounds are more likely to appear together than are others. Sensitivity to such regularities in the speech stream was demonstrated in a series of statistical-learning experiments in which babies learned new words based purely on regularities in how often a given sound followed another (Aslin, Saffran, & Newport, 1998; Saffran, Aslin, & Newport, 1996). The infants listened to a 2-minute recording of four different three-syllable “words” (e.g., tupiro, golabu, bidaku, padoti) repeated in random order with no pauses between the “words.” Then, on a series of test trials, the babies were presented with the “words” they had heard (e.g., bidaku, padoti) and with sequences that were not words (such as syllable sequences that spanned a word boundary—for example, kupado, made up from the end of bidaku and the beginning of padoti).
229
Using the same kind of preferential listening test described for the Juscyzk & Aslin (1995) study on the previous page, the researchers found that infants discriminated between the words and the sequences that were not words. To do so, the babies must have registered that certain syllables often occurred together in the sample of speech they heard. For example, “bi” was always followed by “da” and “da” was always followed by “ku,” whereas “ku” could be followed by “tu,” “go,” or “pa.” Thus, the infants used recurrent sound patterns to fish words out of the passing stream of speech. This ability to learn from distributional properties extends to real languages as well; English-learning infants, for example, can track similar statistical patterns when listening to Italian IDS (Pelucchi, Hay, & Saffran, 2009).
Identifying these regularities in speech sounds supports the learning of words. After repeatedly hearing novel “words” such as timay and dobu embedded in a long stream of speech sounds, 17-month-olds readily learned those sounds as labels for objects (Graf Estes et al., 2007). Similarly, after hearing Italian words like mela and bici embedded in fluent Italian speech, 17-month-olds who had no prior exposure to Italian readily mapped those labels to objects (Hay et al., 2011). Having already learned the sound sequences that made up the words apparently made it easier for the infants to associate the words with their referents.
Probably the most salient regularity for infants is their own name. Infants as young as 4½ months will listen longer to repetitions of their own name than to repetitions of a different but similar name (Mandel, Jusczyk, & Pisoni, 1995). Just a few weeks later, they can pick their own name out of background conversations (Newman, 2005). This ability helps them to find new words in the speech stream. After hearing “It’s Jerry’s cup!” a number of times, 6-month-old Jerry is more likely to learn the word cup than if he had not heard it right after his name (Bortfeld et al., 2005). Over time, infants recognize more and more familiar words, making it easier to pluck new ones out of the speech that they hear.
Infants are exceptional in their ability to identify patterns in the speech surrounding them. They start out with the ability to make crucial distinctions among speech sounds but then narrow their focus to the sounds and sound patterns that make a difference in their native language. This process lays the groundwork for their becoming not just native listeners but also native speakers.
In their first months, babies are getting ready to talk. The repertoire of sounds they can produce is initially extremely limited. They cry, sneeze, sigh, burp, and smack their lips, but their vocal tract is not sufficiently developed to allow them to produce anything like real speech sounds. Then, at around 6 to 8 weeks of age, infants begin to coo—producing long, drawn-out vowel sounds, such as “ooohh” or “aaahh.” Young infants entertain themselves with vocal gymnastics, switching from low grunts to high-pitched cries, from soft murmurs to loud shouts. They click, smack, blow raspberries, squeal, all with apparent fascination and delight. Through this practice, infants gain motor control over their vocalizations.
While their sound repertoire is expanding, infants become increasingly aware that their vocalizations elicit responses from others, and they begin to engage in dialogues of reciprocal ooohing and aaahing, cooing and gooing with their parents. With improvement in their motor control of vocalization, they imitate the sounds of their “conversational” partners, even producing higher-pitched sounds when interacting with their mothers and lower-pitched sounds when interacting with their fathers (de Boysson-Bardies, 1996/1999).
230
babbling repetitive consonant–vowel sequences (“bababa…”) or hand movements (for learners of signed languages) produced during the early phases of language development
Babbling Sometime between 6 and 10 months of age, but on average at around 7 months, a major milestone occurs: babies begin to babble. Standard babbling involves producing syllables made up of a consonant followed by a vowel (“pa,” “ba,” “ma”) that are repeated in strings (“papapa”). Contrary to the long-held belief that infants babble a wide range of sounds from their own and other languages (Jakobson, 1941/1968), research has revealed that babies actually babble a fairly limited set of sounds, some of which are not part of their native language (de Boysson-Bardies, 1996/1999).
Native language exposure is a key component in the development of babbling. Although congenitally deaf infants produce vocalizations similar to those of hearing babies until around 5 or 6 months of age, their vocal babbling occurs very late and is quite limited (Oller & Eilers, 1988). However, some congenitally deaf babies do “babble” right on schedule—those who are regularly exposed to sign language. Infants exposed to ASL babble manually. They produce repetitive hand movements that are components of full ASL signs, just as vocally babbled sounds are repeated components of spoken words (Petitto & Marentette, 1991). Thus, like infants learning a spoken language, infants learning signed languages seem to experiment with the elements that are combined to make meaningful words in their native language (Figure 6.6).
As their babbling becomes more varied, it gradually takes on the sounds, rhythm, and intonational patterns of the language infants hear daily. In a simple but clever experiment, French adults listened to the babbling of a French 8-month-old and an 8-month-old from either an Arabic- or Cantonese-speaking family. When asked to identify which baby was the French one in each pair, the adults chose correctly 70% of the time (de Boysson-Bardies, Sagart, & Durand, 1984). Thus, before infants utter their first meaningful words, they are, in a sense, native speakers of a language.
Early interactions Before we turn to the next big step in language production—uttering recognizable words—it is important to consider the social context that promotes language development in most societies. Even before infants start speaking, they display the beginnings of communicative competence: the ability to communicate intentionally with another person.
The first indication of communicative competence is turn-taking. In a conversation, mature participants alternate between speaking and listening. Jerome Bruner and his colleagues (Bruner, 1977; Ratner & Bruner, 1978) have proposed that learning to take turns in social interactions is facilitated by parent–infant games, such as peekaboo and “give and take,” in which caregiver and baby take turns giving and receiving objects. In these “dialogues,” the infant has the opportunity to alternate between an active and a passive role, as in a conversation in which one alternates between speaking and listening. These early interactions give infants practice in bidirectional communication, providing infants with a scaffold to learn how to use language to converse with others. Indeed, recent research suggests that caregivers’ responses to infant babbling may serve a similar function. When an adult labels an object for an infant just after the infant babbles, the infant’s learning of the label is more greatly enhanced than when the labeling occurs in the absence of babbling (Goldstein et al., 2010). The results of this study suggest that babbling may serve as a signal to the caregiver that the infant is attentive and ready to learn. This early back-and-forth may also provide infants with practice in conversational turn-taking.
231
As discussed in Chapter 4, successful communication also requires intersubjectivity, in which two interacting partners share a mutual understanding. The foundation of intersubjectivity is joint attention, which, early on, is established by the parent’s following the baby’s lead, looking at and commenting on whatever the infant is looking at. By 12 months of age, infants have begun to understand the communicative nature of pointing, with many also being capable of meaningful pointing themselves (Behne et al., 2012).
We have thus seen that infants spend a good deal of time getting ready to talk. Through babbling, they gain some initial level of control over the production of sounds that are necessary to produce recognizable words. As they do so, they already begin to sound like their parents. Through early interactions with their parents, they develop interactive routines similar to those required in the use of language for communication. We will now turn our attention to the processes that lead to infants’ first real linguistic productions: words.
When babies first begin to segment words from fluent speech, they are simply recognizing familiar patterns of sounds without attaching any meaning to them. But then, in a major revolution, they begin to recognize that words have meaning.
reference in language and speech, the associating of words and meaning
The problem of reference The first step for infants in acquiring the meanings of words is to address the problem of reference, that is, to start associating words and meaning. Figuring out which of the multitude of possible referents is the right one for a particular word is, as the philosopher Willard Quine (1960) pointed out, a very complex problem. If a child hears someone say “bunny” in the presence of a rabbit, how does the child know whether this new word refers to the rabbit itself, to its fuzzy tail, to the whiskers on the right side of its nose, or to the twitching of its nose? That the problem of reference is a real problem is illustrated by the case of a toddler who thought “Phew!” was a greeting, because it was the first thing her mother said on entering the child’s room every morning (Ferrier, 1978).
Infants begin associating highly familiar words with their highly familiar referents surprisingly early on. When 6-month-olds hear either “Mommy” or “Daddy,” they look toward the appropriate person (Tincoff & Jusczyk, 1999). Infants gradually come to understand the meaning of less frequently heard words, with the pace of their vocabulary-building varying greatly from one child to another. Remarkably, parents are often unaware of just how many words their infants recognize. Using a computer monitor, Bergelson and Swingley (2012) showed infants pairs of pictures of common foods and body parts and tracked the infants’ eye gaze when one of the pictures was named. They found that even 6-month-olds looked to the correct picture significantly more often than would be expected by chance, demonstrating that they recognized the names of these items. Strikingly, most of their parents reported that the infants did not know the meanings of these words. So not only do infants understand far more words than they can produce; they also understand far more words than even their caregivers realize.
232
One of the remarkable features of infants’ early word recognition is how rapidly they understand what they are hearing. To illuminate the age-related dynamics of this understanding, Fernald and her colleagues presented infants with images depicting pairs of familiar objects, such as a dog and a baby, and observed how quickly the infants moved their eyes to the correct object after hearing its label used (e.g., “Where’s the baby?”). The researchers found that whereas 15-month-olds waited until they had heard the whole word to look at the target object, 24-month-olds looked at the correct object after hearing only the first part of its label, just as adults do (Fernald et al., 1998; Fernald, Perfors, & Marchman, 2006; Fernald, Swingley, & Pinto, 2001). Older infants can also use context to help them recognize words. For example, those who are learning a language that has a grammatical gender system (like Spanish or French) can use the gender of the article preceding the noun (la versus el in Spanish; la versus le in French) to speed their recognition of the noun itself (Lew-Williams & Fernald, 2007; Van Heugten & Shi, 2009). Other visual-fixation research has shown that older infants can even recognize familiar words when they are mispronounced (e.g., “vaby” for “baby,” “gall” for “ball,” “tog” for “dog,” etc.), though their recognition is slower than when they hear the words pronounced correctly (Swingley & Aslin, 2000).
early word production Gradually, infants begin to say some of the words they understand, with most producing their first words between 10 and 15 months of age. The words a child is able to say are referred to as the child’s productive vocabulary.
What counts as an infant’s “first word”? It can be any specific utterance consistently used to refer to something or to express something. Even with this loose criterion, identification of an infant’s earliest few words can be problematic. For one thing, doting parents often misconstrue their child’s babbling as words. For another, early words may differ from their corresponding adult forms. For example, Woof was one of the first words spoken by the boy whose linguistic progress was illustrated at the beginning of this chapter. It was used to refer to the dog next door—both to excitedly name the animal when it appeared in the neighbors’ yard and to wistfully request the dog’s presence when it was absent.
Infants’ early word productions are limited by their ability to pronounce words clearly enough that an adult can recognize them. To make life easier for themselves, infants adopt a variety of simplification strategies (Gerken, 1994). For example, they leave out the difficult bits of words, turning banana into “nana,” or they substitute easier sounds for hard-to-say ones—“bubba” for brother, “wabbit” for rabbit. Sometimes they reorder parts of words to put an easier sound at the beginning of the word, as in the common “pasketti” (for spaghetti) or the more idiosyncratic “Cagoshin” (the way the child quoted at the beginning of the chapter continued for several years to say Chicago).
233
Once children start talking, what do they talk about? The early productive vocabularies of children in the United States include names for people, objects, and events from the child’s everyday life (Clark, 1979; K. Nelson, 1973). Children name their parents, siblings, pets, and themselves, as well as other personally important objects such as cookies, juice, and balls. Frequent events and routines are also labeled—“up,” “bye-bye,” “night-night.” Important modifiers are also used—“mine,” “hot,” “all gone.” Table 6.1 reveals substantial cross-linguistic similarities in the content of the first 10 words of children in the United States, Hong Kong, and Beijing. As the table shows, many of infants’ first words in the three societies referred to specific people or were sound effects (Tardif et al., 2008).
In the early productive vocabularies of children learning English, nouns predominate. One reason may be that because nouns label entities—whereas verbs represent relations among entities—the meanings of nouns are easier to pick up from observation than are the meanings of verbs (Gentner, 1982). Similarly, words that are easier to picture—that are more imageable—are easier for infants and toddlers to learn (McDonough et al., 2011). Another reason is that middle-class American mothers (the group most frequently studied) engage in frequent bouts of object-labeling for their infants—“Look, there’s a turtle! Do you see the turtle?” (Fernald & Morikawa, 1993)—and the proportion of nouns in very young children’s vocabularies is related to the proportion of nouns in their mother’s speech to them (Pine, 1994). Significantly, the pattern of object-labeling by mothers differs across cultures and contexts. For example, Japanese mothers label objects far less often than do American mothers (Fernald & Morikawa, 1993). In the context of toy play, Korean mothers use more verbs than nouns, a pattern very different from that observed in English-speaking mothers (Choi, 2000). And indeed, infants in Korea learn nouns and verbs at the same rate, unlike English-learning infants (Choi & Gopnik, 1995).
234
holophrastic period the period when children begin using the words in their small productive vocabulary one word at a time
Initially, infants say the words in their small productive vocabulary only one word at a time. This phase is referred to as the holophrastic period, because the child typically expresses a “whole phrase”—a whole idea—with a single word. For example, a child might say “Drink!” to express the desire for a glass of juice. Children who produce only one-word utterances are not limited to single ideas; they manage to express themselves by stringing together successive one-word utterances. An example is a little girl with an eye infection who pointed to her eye, saying “Ow,” and then after a pause, “Eye” (Hoff, 2001).
overextension the use of a given word in a broader context than is appropriate
What young children want to talk about quickly outstrips the number of words in their limited vocabularies, so they make the words they do know perform double duty. One way they do this is through overextension—using a word in a broader context than is appropriate, as when children use dog for any four-legged animal, daddy for any man, moon for a dishwasher dial, or hot for any reflective metal (Table 6.2). Most overextensions represent an effort to communicate rather than a lack of knowledge, as demonstrated by research in which children who overextended some words were given comprehension tests (Naigles & Gelman, 1995). In one study, children were shown pairs of pictures of entities for which they generally used the same label—for instance, a dog and a sheep, both of which they normally referred to as “dog.” However, when asked to point to the sheep, they chose the correct animal. Thus, these children understood the meaning of the word sheep, but because it was not in their productive vocabulary, they used a related word that they knew how to say in order to talk about the animal.
235
Within a family, parents often notice significant linguistic differences between their children. Within a given community, such differences can be greatly magnified. In a single kindergarten classroom, for example, there may be a tenfold difference in the number of words used by different children. What accounts for these differences?
The number of words children know is intimately related to the number of words that they hear, which, in turn, is linked to their caregivers’ vocabularies. One of the key determinants of the language children hear is the socioeconomic status of their parents. In a seminal study, Hart and Risley (1994) recorded the speech that 42 parents used with their children over the course of 2½ years, from before the infants were talking until they were 3 years of age. Some of the parents were upper-middle class, others were working class, and others were on welfare. The results were astonishing: the received linguistic experience of the average child whose parents were on welfare (616 words per hour) was half that of the average working-class parents’ child (1251 words per hour) and less than one-third that of the average child in a professional family (2153 words per hour). The researchers did the math and suggested that after 4 years, an average child with upper-middle-class parents would have accumulated experience with almost 45 million words, compared with 26 million for an average child in a working-class family and 13 million for an average child with parents on welfare.
How do these differences affect children’s language development? Not surprisingly, children from higher-SES groups have larger vocabularies than those of children from lower-SES groups (Fenson et al., 1994; Huttenlocher et al., 1991). Indeed, a study of high-SES and mid-SES mothers and their 2-year-old children found that SES differences in maternal speech (e.g., number and length of utterances, richness of vocabulary, and sentence complexity) predicted some of the differences in children’s spoken vocabularies (Hoff, 2003). For instance, higher-SES mothers tended to use longer utterances with their children than did mid-SES parents, giving their children access not only to more words but also to more complex grammatical structures (Hoff, 2003). Such differences in maternal speech even affect how quickly toddlers recognize familiar words: children whose mothers provided more maternal speech at 18 months were faster at recognizing words at 24 months than were children whose mothers provided less input (Hurtado, Marchman, & Fernald, 2008).
Similar findings emerge in the school context. For example, when preschool children with low language skills are placed in classrooms with peers who also have low language skills, they show less language growth than do their counterparts who are placed with classmates who have high language skills (Justice et al., 2011). These peer effects have important implications for programs like Head Start (discussed fully in Chapter 8), which are designed to enhance language development and early literacy for children living in poverty. Unfortunately, congregating children from lower-SES families together in the same preschool classrooms may limit their ability to “catch up.” However, there is the possibility that negative peer effects may be offset by positive teacher effects. For example, one study found that children whose preschool teachers used a rich vocabulary showed better reading comprehension in 4th grade than did children whose preschool teachers used a more limited vocabulary (Dickinson & Porche, 2011).
These results suggest that for a variety of reasons, parents’ SES affects the way they talk to their children; in turn, those individual differences have a substantial influence on the way their children talk. These differences can be intensified by the linguistic abilities of children’s peers and teachers. For children in low-SES environments, the potential negative effects of these influences may be offset by interventions ranging from increased access to children’s books (which provide enriched linguistic environments) to enhancing teacher training in low-income preschool settings (Dickinson, 2011). Regardless of the source, what goes in is what comes out: we can only learn words and grammatical structures that we hear (or see or read) in the language surrounding us.
236
Word learning After the appearance of their first words, children typically plod ahead slowly, reaching a productive vocabulary of 50 or so words by around 18 months of age. At this point, the rate of learning appears to accelerate, leading to what appears to be a “vocabulary spurt” (e.g., L. Bloom, 1973; McMurray, 2007). Although scholars disagree about whether learning actually speeds up for all or even most children (Bloom, 2000), it is clear that children’s communicative abilities are growing rapidly (Figure 6.7).
What accounts for the skill with which young children learn words? When we look closely, we see that there are multiple sources of support for learning new words, some coming from the people around them, and some generated by the children themselves.
ADULT INFLUENCES ON WORD LEARNING In addition to using IDS, which makes word learning easier for infants, adults facilitate word learning by highlighting new words. For example, they stress new words, or place them at the end of a sentence. They also tend to label objects that are already the focus of the child’s attention, thereby reducing uncertainty about the referent (Masur, 1982; Tomasello, 1987; Tomasello & Farrar, 1986). Their repetition of words also helps; young children are more likely to acquire words their parents use frequently (Huttenlocher et al., 1991). Another stimulus to word learning that adults provide involves naming games, in which they ask the child to point to a series of named items—“Where’s your nose?” “Where’s your ear?” “Where’s your tummy?” There is also some evidence that parents may facilitate their children’s word learning by maintaining spatial consistency with the objects they are labeling. For instance, in a study in which parents labeled novel objects for their infants, the infants learned the names of the objects more readily when the objects were in the same location each time they were labeled (Samuelson et al., 2011). Presumably, consistency in the visual environment helps children map words onto objects and events in that environment. In Box 6.3, we discuss recent research focused on a currently popular means by which some parents try to “outsource” word learning: technology.
fast mapping the process of rapidly learning a new word simply from hearing the contrastive use of a familiar and the unfamiliar word
CHILDREN’S CONTRIBUTIONS TO WORD LEARNING When confronted with words they haven’t heard before, children actively exploit the context in which the new word was used in order to infer its meaning. A classic study by Carey and Bartlett (1978) demonstrated fast mapping—the process of rapidly learning a new word simply from hearing the contrastive use of a familiar word and the unfamiliar word. In the course of everyday activities in a preschool classroom, an experimenter drew a child’s attention to two trays—one red, the other an uncommon color the child would not know by name—and asked the child to get “the chromium tray, not the red one.” The child was thus provided with a contrast between a familiar term (red) and an unfamiliar one (chromium). From this simple contrast, the participants inferred which tray they were supposed to get and that the name of the color of that tray was “chromium.” After this single exposure to a novel word, about half the children showed some knowledge of it 1 week later by correctly picking out “chromium” from an array of paint chips.
Some theorists have proposed that the many inferences children make in the process of learning words are guided by a number of assumptions (sometimes referred to as principles, constraints, or biases) that limit the possible meanings children entertain for a new word. For example, children expect that a given entity will have only one name, an expectancy referred to as the mutual exclusivity assumption by Woodward and Markman (1998). Early evidence for this assumption came from a study in which 3-year-olds saw pairs of objects—a familiar object for which the children had a name and an unfamiliar one for which they had no name. When the experimenter said, “Show me the blicket,” the children mapped the novel label to the novel object, the one for which they had no name (Markman & Wachtel, 1988). Even 16-month-old infants do the same (Graham, Poulin-Dubois, & Baker, 1998). (See Figure 6.8a.) Interestingly, bilingual and trilingual infants, who are accustomed to hearing more than one name for a given object, are less likely to follow the mutual exclusivity principle (Byers-Heinlein & Werker, 2009).
237
Markman and Woodward (Markman, 1989; Woodward & Markman, 1998) also proposed the whole-object assumption, according to which children expect a novel word to refer to a whole object rather than to a part, property, action, or other aspect of the object. Thus, in the case of Quine’s rabbit problem (see page 231), the whole-object assumption leads children to map the label “bunny” to the whole animal, not just to its tail or the twitching of its nose.
pragmatic cues aspects of the social context used for word learning
When confronted with novel words, children also exploit a variety of pragmatic cues to their meaning by paying attention to the social contexts in which the words are used. For example, children use an adult’s focus of attention as a cue to word meaning. In a study by Baldwin (1993), an experimenter showed 18-month-olds two novel objects and then concealed them in separate containers. Next, the experimenter peeked into one of the containers and commented, “There’s a modi in here.” The adult then removed and gave both objects to the child. When asked for the “modi,” the children picked the object that the experimenter had been looking at when saying the label. Thus, the infants used the relation between eye gaze and labeling to learn a novel name for an object before they had ever seen it (see Figure 6.8b).
Another pragmatic cue that children use to draw inferences about a word’s meaning is intentionality (Tomasello, 2007). For instance, in one study, 2-year-olds heard an experimenter announce, “Let’s dax Mickey Mouse.” The experimenter then performed two actions on a Mickey Mouse doll, one carried out in a coordinated and apparently intentional way, followed by a pleased comment (“There!”), and the other carried out in a clumsy and apparently accidental way, followed by an exclamation of surprise (“Oops!”). The children interpreted the novel verb dax as referring to the action the adult apparently intended to perform (Tomasello & Barton, 1994). Infants can even use an adult’s emotional response to infer the name of a novel object that they cannot see (Tomasello, Strosberg, & Akhtar, 1996). In a study establishing this fact, an adult announced her intention to “find the gazzer.” She then picked up one of two objects and showed obvious disappointment with it. When she gleefully seized the second object, the infants inferred that it was a “gazzer.” (Figure 6.8c depicts another instance in which a child infers the name of an unseen object from an adult’s emotional expression.)
The degree to which preschool children take a speaker’s intention into account is shown by the fact that if an adult’s labeling of an object conflicts with their knowledge of that object, they will nevertheless accept the label if the adult clearly used it intentionally (Jaswal, 2004). When an experimenter simply used the label “dog” in referring to a picture of a catlike animal, preschool children were reluctant to extend the label to other catlike stimuli. However, they were much more willing to do so when the experimenter made it clear that he really intended his use of the unexpected label by saying, “You’re not going to believe this, but this is actually a dog.” Similarly, if a child has heard an adult refer to a cat as “dog,” the child will be reluctant to subsequently learn a new word used by that “untrustworthy” adult (e.g., Koenig & Harris, 2005; Koenig & Woodward, 2010; Sabbagh & Shafman, 2009).
238
Another way young children can infer the meaning of novel words is by taking cues from the linguistic context in which the words are used. In one of the first experiments on language acquisition, Brown (1957) established that the grammatical form of a novel word influences children’s interpretation of it. He showed preschool children a picture of a pair of hands kneading a mass of material in a container (Figure 6.9). The picture was described to one group of children as “sibbing,” to another as “a sib,” and to a third as “some sib.” The children subsequently interpreted sib as referring to the action, the container, or the material, depending on which grammatical form (verb, count noun, or mass noun) of the word they had heard.
239
Two- and 3-year-old children also use the grammatical category of novel words to help interpret their meaning (e.g., Hall, 1994; Hall, Waxman, & Hurwitz, 1993; Markman & Hutchinson, 1984; Waxman, 1990). Hearing “This is a dax,” they assume that dax refers to an object, as well as to other objects from the same category. In contrast, “This is a dax one” suggests that dax refers to a property of the object (e.g., its color or texture), while “This is dax” suggests that dax is a proper noun (a name). Even infants and toddlers can draw on some of these links to interpret the meaning of novel words (e.g., Booth & Waxman, 2009; Waxman & Hall, 1993; Waxman & Markow, 1995, 1998).
Children’s interpretation of novel words applied to objects is particularly guided by the objects’ shape, possibly because shape is a good cue to category membership. Children readily extend a novel noun to novel objects of the same shape, even when those objects differ dramatically in size, color, and texture (Graham & Poulin-Dubois, 1999; Landau, Smith, & Jones, 1988; L. B. Smith, Jones, & Landau, 1992). Thus, a child who hears a U-shaped wooden block called “a dax” will assume that dax also refers to a U-shaped object covered in blue fur or to a U‑shaped piece of red wire—but not to a wooden block of a different shape (Figure 6.10). A shape bias is also evident in young children’s spontaneous extension of familiar words to novel objects that are vaguely similar to familiar entities (e.g., a cone might be referred to as a “mountain”) (Samuelson & Smith, 2005).
Another potentially useful cue to word meaning is the repeated correspondence between words the child hears and objects the child observes in the world. Any single scene is ambiguous. For example, if the child sees four different novel objects together and hears the label “dax,” the child has no way of knowing which object is the dax (this ambiguity is similar to that in Quine’s “bunny” example described earlier). But across experiences, the child might observe that whenever “dax” is said, one of those four objects is always present, and thus that object is probably the dax. Through this process of cross-situational word learning, even infants can narrow down the possible meanings of new words (e.g., L. Smith & Yu, 2008; Vouloumanos & Werker, 2009).
syntactic bootstrapping the strategy of using the grammatical structure of whole sentences to figure out meaning
Children also use the grammatical structure of whole sentences to figure out meaning—a strategy referred to as syntactic bootstrapping (Fisher, 1999; Fisher, Gleitman, & Gleitman, 1991; Gertner, Fisher, & Eisengart, 2006; Yuan & Fisher, 2009). An early demonstration of this phenomenon involved showing 2-year-olds a videotape of a duck using its left hand to push a rabbit down into a squatting position while both animals waved their right arms in circles (Figure 6.11) (Naigles, 1990). As they watched, some children were told “The duck is kradding the rabbit”; others were told “The rabbit and the duck are kradding.” All the children then saw two videos side by side, one showing the duck pushing on the rabbit and the other showing both animals waving their arms in the air. Instructed to “Find kradding,” the two groups looked at the event that matched the syntax they had heard while watching the initial video. Those who had heard the first sentence took “kradding” to mean what the duck had been doing to the rabbit, whereas those who had heard the second sentence thought it meant what both animals had been doing. Thus, the children had arrived at different interpretations for a novel verb based on the structure of the sentence in which it was embedded.
240
When adults enter a new language culture, they often have access to many technological aids—from pocket dictionaries to smartphone apps for translations—to help them get around and ask for what they need. They can also harness technology to learn a new language, from collegiate digital language labs to commercial training programs like Rosetta Stone.
What about infants? They, too, are often immersed in technology. Indeed, over the past two decades, there was a “brainy-baby” craze in which businesses earned hundreds of millions of dollars marketing all manner of electronic games, toys, and DVDs that claimed to enhance babies’ intellectual growth. Some of these claims were laughable. For example, one of your authors purchased a teething ring for her baby and was astonished to read on the packaging that the ring supposedly improves an infant’s language development by enhancing early oral-motor skills. Other claims seemed more plausible but were subsequently called into serious question by developmental research—so much so that the companies producing these products (e.g., “Baby Einstein”) were forced to stop promoting their “educational” value.
However, there is still concern over technology marketed for children younger than 2 years because it reduces the time infants and toddlers spend actively engaged with caregivers and objects, their best source for learning. In a large-scale study that included more than 1000 children younger than 2 years, researchers examined the association between TV/DVD viewing and vocabulary development (Zimmerman, Cristakis, & Meltzoff, 2007). Crucially, the authors controlled for family demographics (including parents’ education, family income, race/ethnicity, and so forth) that may influence media use. For 8- to 16-month-olds, there was a negative correlation between the amount of time infants watched DVDs and parental reports of the infants’ vocabularies: the more the babies watched, the lower their vocabulary scores. Note that this negative relationship was observed only for DVDs marketed as “educational” for babies; other types of programming (including both educational and noneducational children’s television programs) showed no association with vocabulary development. For the older infants in the study (between 17 and 24 months of age), the negative relationship disappeared. Contrary to many parents’ beliefs, the amount of time parents spent watching TV with their children did not affect the children’s vocabulary scores. These results are consistent with the American Academy of Pediatrics’ recommendation to minimize screen time for children younger than 2 years.
241
In the most rigorous study to date, DeLoache and colleagues (2010) used random assignment and an objective test of vocabulary to determine whether a best-selling “educational” DVD had an impact on language development. The DVD in question was marketed to infants at least 12 months old. The researchers randomly assigned 12- to 18-month-olds into four groups. Infants in the video-with-interaction group watched the video with a parent 5 times a week over 4 weeks; the parent was asked to interact naturally with the infant while watching. Infants in the video-without-interaction group received the same amount of exposure, but without a parent watching along with them. This mimics a common situation at home, where parents might be in the same room but are engaged in another activity. Infants in the parent-teaching group did not watch the video at all. Instead, their parents were given a list of 25 words featured on the video and asked to teach the infants those words in whatever way felt most natural to them. Finally, infants in the control group received no intervention, serving as a baseline for typical vocabulary development.
At the beginning and end of the study, the infants were tested on a subset of the words featured on the DVD. Infants in the DVD-viewing conditions showed no more advancement in vocabulary than the infants in the control group did. The infants who showed the greatest vocabulary development were the infants in the parent-teaching group. Interestingly, the performance of infants in the DVD-viewing conditions was unrelated to how much parents thought their infant had learned from the DVD. However, there was a correlation between how much the parents liked the DVD themselves and how much they thought their infant had learned: the more parents liked the DVD, the more likely they were to overestimate its positive effects.
Although this area of research is still relatively new, the results thus far suggest that outsourcing infants’ vocabulary development to their passive observation of educational media does not match up with how infants actually learn. Even the best technology cannot substitute for a caregiver who is interacting with the child. The importance of such interaction was highlighted by a study in which toddlers learned new words from an adult in one of three different conditions: (1) in-person interaction with the adult; (2) interaction with the adult via video link; and (3) observation of a video of the adult interacting with another adult. The word-learning outcomes for the first two conditions were the same—and superior to those in which toddlers were passive observers (O’Doherty et al., 2011). Questions about electronic media for infants are only going to increase, given the delight with which young children, and even infants, have embraced the apps on their parents’ smartphones and iPads. As with any type of activity, a little here and there probably doesn’t hurt. But it is important to approach any claims of “educational value” with a great deal of skepticism.
As you can see, infants and young children have a remarkable ability to learn new words as object names. Under some circumstances, they are also able to learn nonlinguistic “labels” for objects. Infants between 13 and 18 months of age map an experimenter’s gestures or nonverbal sounds (e.g., squeaks and whistles) onto novel objects just as readily as they map words (Namy, 2001; Namy & Waxman, 1998; Woodward & Hoyne, 1999). By 20 to 26 months of age, however, they accept only words as names. And when novel labels are presented via computer rather than through interaction with an adult, even 12-month-olds accept only words as labels, not other nonverbal sounds (MacKenzie, Graham, & Curtin, 2011). Thus, infants learn quite early on that strings of phonemes are more likely to carry meaning than other types of sounds.
242
A major landmark in early language development is achieved when children start combining words into sentences, an advance that enables them to express increasingly complex ideas. The degree to which children develop syntax, and the speed with which they do it, is what most distinguishes their language abilities from those of nonhuman primates.
First sentences Most children begin to combine words into simple sentences by the end of their second year. However, in another example of comprehension preceding production, young children know something about word combinations well before they produce any. In one of the first demonstrations of infants’ sensitivity to word order, Hirsh-Pasek and Golinkoff (1991) showed infants two videotaped scenes—one of a woman kissing some keys while holding up a ball and the other of the woman holding up the keys while kissing the ball. Thus, the same three elements—kissing, keys, and a ball—were present in both scenes. Yet when the infants heard the sentence “She’s kissing the keys” or “She’s kissing the ball,” they looked preferentially at the appropriate scene.
telegraphic speech the term describing children’s first sentences that are generally two-word utterances
Children’s first sentences are two-word combinations; their separate utterances of “More,” “Juice,” and “Drink” become “More juice” and “Drink juice.” These two-word utterances have been described as telegraphic speech because, just as in telegrams, nonessential elements are missing (Brown & Fraser, 1963). Consider the following examples of standard two-word utterances: “Read me,” “Mommy tea,” “Ride Daddy,” “Hurt knee,” “All wet,” “More boy,” “Key door,” “Andrew sleep” (Braine, 1976). These primitive sentences lack a number of elements that would appear in adult utterances, including function words (such as a, the, in), auxiliary verbs (is, was, will, be), and word endings (indicating plurals, possessives, or verb tenses). Children’s early sentences possess this telegraphic quality in languages as diverse as English, Finnish, Dholuo (Kenya), and Kaluli (Papua New Guinea) (de Boysson-Bardies, 1996/1999). For young children learning languages like English, in which word order is crucial for meaning, their early sentences follow a consistent word order: a child might say “Eat cookie” but would be unlikely ever to say “Cookie eat.”
Many children continue to produce one- and two-word utterances for some time, whereas others quickly move on to sentences consisting of three or more words. Figure 6.12 shows the rapid increase in the mean length of utterances of three children in Roger Brown’s (1973) classic study of language development. As you can see from the figure, Eve started her explosive increase in sentence length much earlier than the other two children did. The length of children’s utterances increases in part because they begin to systematically incorporate some of the elements that were missing from their telegraphic speech.
Once children are capable of producing four-word sentences, typically at around 2½ years of age, they begin to produce complex sentences containing more than one clause (Bowerman, 1979): “Can I do it when we get home?” “I want this doll because she’s big” (Limber, 1973).
243
Grammar: a tool for building new words and sentences As noted at the beginning of this chapter, human languages are generative: through the use of the finite set of words and morphemes in their native vocabulary, individuals can create an infinite number of sentences and express an infinite number of ideas. Each language has a particular set of rules (and exceptions) that governs how linguistic elements can be combined. The power of language derives from the mastery of these regularities, which allows individuals to produce and understand language beyond the specific words and sentences to which they have been exposed. How does this mastery come about, especially in the early years of life?
Much of the research on this topic has focused on morphemes that are added to nouns and verbs. In English, nouns are made plural by adding–s, and verbs are put into the past tense by adding–ed, with some notable exceptions (e.g., men, went). Young children recognize these formations and are able to generalize them to novel words. In a classic experiment by Berko (1958), for example, preschoolers were shown a picture of a nonsense animal, which the experimenter referred to as “a wug.” Then a picture of two of the creatures was produced, and the experimenter said, “Here are two of them; what are they?” Children as young as 4 readily answered correctly: “Wugs.” Since the children had never heard the word wug before, their ability to produce the correct plural form for this totally novel word demonstrated that they could generalize beyond the other plurals they had previously heard. The results of this study are taken as evidence that the participants understood how English pluralization works.
overregularization speech errors in which children treat irregular forms of words as if they were regular
Evidence for generalization also comes from the way children treat irregular cases. Consider the plural of man and the past tense of go. Children initially use the correct irregular forms of these words, saying “men” for the plural of man and “went” for the past tense of go. However, they then start making occasional overregularization errors, in which they treat irregular forms as if they were regular. For instance, a child who previously said “men” and “went” may begin producing novel forms such as “mans” and “goed,” as well as “foots,” “feets,” “breaked,” “broked,” and even “branged” and “walkeded” (Berko, 1958; Kuczaj, 1977; Xu & Pinker, 1995). Before eventually mastering irregular forms, children sometimes alternate between these overregularization errors and correct irregular forms (Marcus, 1996; 2004). The following dialogue between a 2½-year-old and his father illustrates this kind of error, as well as the difficulty of correcting it:
Child: I used to wear diapers. When I growed up (pause)
Father: When you grew up?
Child: When I grewed up, I wore underpants.
(Clark, 1993)
Parents play a role in their children’s grammatical development, although a more limited one than you might expect. Clearly, they provide a model of grammatically correct speech. In addition, they frequently fill in missing parts of their children’s incomplete utterances, as when a parent responds to “No bed” by saying, “You really don’t want to go to bed right now, do you?”
One might think that parents also contribute to their children’s language development by correcting their grammatical errors. In fact, parents generally ignore even wildly ungrammatical mistakes, accepting sentences such as “I magicked it,” “Me no want go,” or “I want dessert in front of dinner” (Bryant & Polkosky, 2001; Brown & Hanlon, 1970). It would be hard to do otherwise, since so much of children’s speech is like this. And, as the parent who tried to correct his son’s use of “growed” discovered, such efforts are largely futile anyway. In general, parents are more likely to correct factual errors than grammatical errors.
244
Given this lack of correction, how do infants figure out the ways in which the syntax of their native language works? One approach to answering this question in the lab involves creating miniature languages—known as artificial grammars—and determining which types of linguistic patterns infants are able to learn. After just brief exposure, infants as young as 8 months can learn fairly complex patterns, generalizing beyond the specific items they have heard (e.g., Gerken, Wilson, & Lewis, 2005; Gómez, 2002; Lany & Saffran, 2010; Marcus et al., 1999; Saffran et al., 2008). For example, infants who have heard a list of three-word sequences in which the second word is repeated, such as “le di di, wi je je, de li li…” recognize the pattern when it is presented in new syllables, such as “ko ga ga” (Marcus et al., 1999). Ongoing research focuses on the degree to which these laboratory studies are tapping into the same processes that infants use to acquire native language syntax.
Learning how to combine words to create interpretable sentences is the crowning achievement in language acquisition. Possibly no linguistic development is more stunning than the progress children make in a few years from simple two-word utterances to grammatically correct complex sentences. Even their errors reveal an increasingly sophisticated understanding of the structure of their native language. This accomplishment is made all the more impressive given the relative lack of parental feedback they receive. The manner in which this learning unfolds is the focus of current research.
Although young children are eager to participate in conversations with others, their conversational skills lag well behind their burgeoning language skills. For one thing, much of very young children’s speech is directed to themselves, rather than to another person. And this is true not just in solitary play: as much as half of young children’s speech in the company of other children or adults is addressed to themselves (Schoeber-Peterson & Johnson, 1991). Vygotsky ((1934/1962) believed that this private speech of young children serves an important regulatory function: children talk to themselves as a strategy to organize their actions (Behrend, Rosengren, & Perlmutter, 1992). Gradually, private speech is internalized as thought, and children become capable of mentally organizing their behavior, so they no longer need to talk out loud to themselves.
collective monologue conversation between children that involves a series of non sequiturs, the content of each child’s turn having little or nothing to do with what the other child has just said
As noted in Chapter 4, when young children converse with other children, their conversations tend to be egocentric. Piaget (1923/1926) labeled young children’s talk with their peers as collective monologues. Even when they take turns speaking, their conversations tend to be a series of non sequiturs, with the content of each child’s turn having little or nothing to do with what the other child has just said. The following conversation between two preschoolers is a good example of what Piaget observed:
Jenny: My bunny slippers…are brown and red and sort of yellow and white. And they have eyes and ears and these noses that wiggle sideways when they kiss.
Chris: I have a piece of sugar in a red piece of paper. I’m gonna eat it but maybe it’s for a horse.
Jenny: We bought them. My mommy did. We couldn’t find the old ones. These are like the old ones. They were not in the trunk.
Chris: Can’t eat the piece of sugar, not unless you take the paper off.
(Stone & Church, 1957, pp. 146–147)
Gradually, children’s capacity for sustained conversation increases. In a longitudinal study of parent–child conversations of four children from the age of 21 months to 36 months, Bloom, Rocissano, and Hood (1976) found that the proportion of children’s utterances that were on the same topic and added new information to what an adult had just said more than doubled (from around 20% to more than 40%). In contrast, the proportion of utterances on unrelated topics fell dramatically (from around 20% to almost 0).
245
narratives descriptions of past events that have the basic structure of a story
A particular aspect of young children’s conversations that changes dramatically in the preschool period is the extent to which they talk about the past. At most, 3-year-olds’ conversations include occasional brief references to past events. In contrast, 5-year-olds produce narratives—descriptions of past events that have the form of a story (Miller & Sperry, 1988; K. Nelson, 1993). One thing that makes longer, more coherent narratives possible is better understanding of the basic structure of stories (Peterson & McCabe, 1988; Shapiro & Hudson, 1991; Stein, 1988).
Parents actively assist their children to develop the ability to produce coherent accounts of past events by providing what has been referred to as scaffolding (discussed in Chapter 4) for their children’s narratives (Bruner, 1975). An effective way to structure children’s conversations about the past is to ask them elaborative questions, that is, questions that enable them to say something—anything—that advances the story:
Mother: And what else happened at the celebrations?
Child: I don’t know.
Mother: We did something special with all the other children.
Child: What was it?
Mother: There were a whole lot of people over at the beach, and everyone was doing something in the sand.
Child: What was it?
Mother: Can’t you remember what we did in the sand? We were looking for something.
Child: Umm, I don’t know.
Mother: We went digging in the sand.
Child: Umm, and that was when um the yellow spade broke.
Mother: Good girl, I’d forgotten that. Yes, the yellow spade broke, and what happened?
Child: Um, we had to um dig with the other end of the yellow bit one.
Mother: That’s right. We used the broken bit, didn’t we?
Child: Yeah.
(Farrant & Reese, 2002)
The child in this conversation does not actually say much, but the parent’s questions help the child think about the event, and the parent also provides a conversational model. Those toddlers whose parents scaffold their early conversations by asking useful, elaborative questions produce better narratives on their own a few years later (Fivush, 1991; McCabe & Peterson, 1991; Reese & Fivush, 1993).
A crucial aspect of becoming a good conversational partner is the pragmatic development that allows children to understand how language is used to communicate. Such understanding is essential with utterances that require listeners to go beyond the words they are hearing to grasp their actual meaning—as in instances of rhetorical questioning, sarcasm, irony, and the use of hyperbole or understatement to make a point. Children’s pragmatic abilities develop over the course of the preschool years, facilitating communication with adults and peers. In particular, they learn to take the perspective of their conversational partner, something that is clearly lacking in the example of the “conversation” between preschoolers Jenny and Chris, quoted earlier. Kindergarten-aged listeners are able to make use of a conversational partner’s perspective (e.g., by considering what information relevant to the conversation the partner does or doesn’t have) to figure out what the partner means, and to provide a pertinent response (Nadig & Sedivy, 2002; Nilsen & Graham, 2009). The development of this ability is related to children’s level of executive function; as children become more able to control their tendency to assume their own perspective, it becomes easier for them to take the perspective of a conversational partner.
246
Children also learn to use information other than words to interpret meaning. For example, older preschoolers can exploit the vocal affect of an ambiguous statement to figure out a speaker’s intention (Berman, Chambers, & Graham, 2010). When presented with two dolls—one intact, the other broken—and directed to “Look at the doll,” 4-year-olds (but not 3-year-olds) looked at the intact doll when the instruction was given with positive affect, and at the broken doll when the instruction was given with negative affect.
We thus see that young children put their burgeoning linguistic skills to good use, becoming more effective communicative partners. Initially, they need substantial support from a more competent partner, but their conversational skills increase quite regularly. Children’s growing understanding of narrative structure and their emerging ability to take other people’s perspectives are crucial components in the development of their conversational skills.
From 5 or 6 years of age on, children continue to develop language skills, although with less dramatic accomplishments. For example, the ability to sustain a conversation, which grew so dramatically in the preschool years, continues to improve into adulthood. School-aged children become increasingly capable of reflecting upon and analyzing language, and they master more complex grammatical structures, such as the use of passive constructions.
One consequence of schoolchildren’s more reflective language skills is their increasing appreciation of the multiple meanings of words, which is responsible for the emergence of the endless series of puns, riddles, and jokes with which they delight themselves and torture their parents (Ely & McCabe, 1994). They also are able to learn the meaning of new words simply from hearing them defined (Pressley, Levin, & McDaniel, 1987), a factor that helps their comprehension vocabulary expand—from the 10,000 words that the average 6-year-old knows to the 40,000 words estimated for 5th graders (Anglin, 1993) to the average college-student vocabulary that has been estimated to be as high as 150,000 words (Miller & Gildea, 1987).
As you have seen throughout this chapter, there is ample evidence for both nature and nurture in the process of language development. The two key prerequisites for language acquisition are (1) a human brain and (2) experience with a human language. The former clearly falls on the side of nature, and the latter, on the side of nurture. Despite the obvious interaction between the two, the nature–nurture debate continues to rage fiercely in the area of language development. Why?
247
The study of language development emerged from a theoretical debate about the processes through which language is acquired. In the 1950s, B. F. Skinner (1957) wrote a book entitled Verbal Behavior, in which he presented a behaviorist theory of language development. As you saw in Chapter 1, behaviorists believed that development is a function of learning through reinforcement and punishment of overt behavior. Skinner argued that parents teach children to speak by means of the same kinds of reinforcement techniques that are used to train animals to perform novel behaviors.
In what was probably the most influential book review ever published, Noam Chomsky (1959) countered Skinner by pointing out some of the reasons why language cannot be learned through the processes of reinforcement and punishment. One key reason was noted earlier in this chapter: we can understand and produce sentences that we have never heard before (generativity). If language-learning proceeds by means of reinforcement and punishment, how could we know that a sentence like “Colorless green ideas sleep furiously” is a grammatical English sentence, whereas “Green sleep colorless furiously ideas” is not (Chomsky, 1957)? Similarly, how could children produce words they have never heard before, like wented or mouses? The explanation of such instances must be that we know details about the structure of our native language that we have not been taught—facts that are unobservable and thus impossible to reinforce—contrary to Skinner’s proposal.
248
Universal Grammar a proposed set of highly abstract, unconscious rules that are common to all languages
In his own explanation of language development, Chomsky proposed that humans are born with a Universal Grammar, a hard-wired set of principles and rules that govern grammar in all languages. Chomsky’s account, which has been central to the development of the modern discipline of linguistics, is consistent with the fact that, despite many surface differences, the underlying structures of the world’s languages are fundamentally similar. His strongly nativist account also provides an explanation for why most children learn language with exceptional rapidity, while nonhumans (who presumably lack a Universal Grammar) do not. (The Universal Grammar hypothesis has been highly relevant to investigations of emerging languages like the Nicaraguan Sign Language discussed in Box 6.4, a language in which children are creating new grammatical structures.)
People around the world spontaneously accompany their speech with gestures. The naturalness of gesturing is revealed by the fact that blind people gesture as they speak just as much as sighted individuals do, even when they know their listener is also blind (Iverson & Goldin-Meadow, 1998).
Gesturing starts early: infants often produce recognizable, meaningful gestures before they speak recognizable words. According to Acredolo and Goodwyn (1990), many “baby signs” are invented by children themselves. One child in their research signed “alligator” by putting her hands together and opening and closing them to imitate snapping jaws; another indicated “dog” by sticking her tongue out as if panting; another signaled “flower” by sniffing. Infants gain earlier motor control of their hands than of their vocal apparatus, facilitating the use of signs during the first year.
Interestingly, there is a relationship between early gesturing and later vocabulary (M. L. Rowe, Ozcaliskan, & Goldin-Meadow, 2008). The more children gestured at 14 months, the larger their spoken vocabulary was at 42 months. Moreover, differences in the amount of gesturing by high- and low-SES families are one factor that influences the SES effects discussed in Box 6.2 (M. L. Rowe & Goldin-Meadow, 2009).
Especially dramatic evidence of intimate connections between gesture and language comes from remarkable research on children who have created their own gesture-based languages. Goldin-Meadow and colleagues studied congenitally deaf American and Chinese children whose hearing parents had little or no proficiency in any formal sign language (Feldman, Goldin-Meadow, & Gleitman, 1978; Goldin-Meadow, 2003; Goldin-Meadow & Mylander, 1998). These children and their parents created “home signs” in order to communicate with one another, and the children’s gesture vocabulary quickly outstripped that of their parents.
More important, the children (but not the parents) spontaneously imposed a structure—a rudimentary grammar—on their gestures. Both groups of children used a grammatical structure that occurs in some languages but not those of their parents. As a result, the sign systems of the children were more similar to one another than to those of their own parents. The children’s signs were also more complex than those of their parents. A similar phenomenon occurs when deaf children are learning to sign from parents who themselves learned a conventional sign language, like ASL, but whose signing is ungrammatical (usually because they learned to sign later in life). In such cases, deaf children have been reported to spontaneously impose structure that is more consistent than the signs that their parents produce (Singleton & Newport, 2004).
The most extensive and extraordinary example of language creation by children comes from the invention of Nicaraguan Sign Language (NSL), a completely new language that has been evolving over the past 30 years. In 1979, a large-scale education program for the deaf began in Nicaragua (Senghas & Coppola, 2001). The program brought hundreds of deaf children together in two schools in the city of Managua. For most of the children, it was their first exposure to other deaf people.
249
The teachers in the schools knew no formal sign language, nor did the children, who had only the simple home signs they had used to communicate with their families. The children quickly began to build on one another’s existing informal signs, constructing a “pidgin” sign language—a relatively crude, limited communication system. The language was used by the students, both in and outside school, and was learned by each new group of children who entered the community.
What happened next was astonishing. As younger students entered the schools, they rapidly mastered the rudimentary system used by the older students and then gradually transformed it into a complex, fully consistent language (NSL) with its own grammar. The most fluent signers in the NSL community are currently the youngest children, both because NSL has evolved into a real language and because they acquired it at an earlier age.
Recently, another emerging signed language was discovered in the Negev desert of Israel (Sandler et al., 2005). The Al-Sayyid Bedouin Sign Language (ABSL) is in its third generation and is about 75 years old. Unlike NSL, ABSL is acquired from birth, because deaf children in this community typically have at least one deaf adult in their extended family. The grammatical structure of ABSL does not resemble those of the local spoken languages (Arabic and Hebrew).
These reports of language invention by deaf children are not just fascinating stories: they provide evidence for the child’s contribution to language learning. Over several generations, children have been taking improvised signing that is simple and inconsistent and transforming it into structures much closer to those observed in established languages. Whether this process reflects the operation of a Chomsky-style Universal Grammar or the operation of more general learning mechanisms remains unknown. Regardless, the discovery that children go beyond the linguistic input they receive, spontaneously refining and systematizing these emerging languages, is one of the most fascinating findings in the field of language development.
250
Current theories all acknowledge some of Chomsky’s crucial observations. For example, any account of language development must be able to explain why all human languages share so many characteristics. Theories must also explain how it is that language users, from infants to adults, are able to generalize beyond the specific words and sentences they have been exposed to. But the ways in which various accounts handle these facts differ along two key dimensions. The first dimension is the degree to which these explanations lie within the child (nature) versus within the environment (nurture). The second dimension pertains to the child’s contributions: Did the cognitive and neural mechanisms underlying language learning evolve solely to support language learning (domain-specific), or are they used for learning many different kinds of things (domain-general)?
With respect to the first dimension, theorists have countered Chomsky’s argument about the universality of language structure by pointing out that there are also universals in children’s environments. Parents all over the world need to communicate about certain things with their children, and these things are likely to be reflected in the language that children learn. For example, recall Table 6.1, which shows the remarkable overlap in the earliest words acquired across three diverse cultures (Tardif et al., 2008). These similarities reflect what parents want to talk about with their infants, and what infants want to communicate about.
Indeed, accounts focused on social interaction maintain that virtually everything about language development is influenced by its communicative function. Children are motivated to interact with others, to communicate their own thoughts and feelings, and to understand what other people are trying to communicate to them (Bloom, 1991; Bloom & Tinker, 2001; Snow, 1999). According to this position, children gradually discover the underlying regularities in language and its use by paying close attention to the multitude of clues available in the language they hear, the social context in which language is used, and the intentions of the speaker (e.g., Tomasello, 2008). Some of these conventions may be learned by the same kinds of reinforcement methods originally proposed by Skinner. For example, Goldstein and colleagues have found that both the sounds infants make when babbling and the rate at which they produce them can be influenced by parental reinforcement, such as smiling and touching in response to the babbling (Goldstein, King, & West, 2003; Goldstein & Schwade, 2008). Whether these types of social behaviors can influence less overt aspects of language development, such as the acquisition of syntax, remains unclear.
modularity hypothesis the idea that the human brain contains an innate, self-contained language module that is separate from other aspects of cognitive functioning
What of the second dimension—the domain-specificity of the processes underlying language acquisition? According to the strongly nativist view espoused by Chomsky, the cognitive abilities that support language development are highly specific to language. As Steven Pinker (1994) describes it, language is “a distinct piece of the biological makeup of our brains…distinct from more general abilities to process information or behave intelligently” (p. 18). This claim is taken one step further by the modularity hypothesis, which proposes that the human brain contains an innate, self-contained language module that is separate from other aspects of cognitive functioning (Fodor, 1983). The idea of specialized mental modules is not limited to language. As you will see in Chapter 7, innate, special-purpose modules have been proposed to underlie a variety of functions, including perception, spatial skills, and social understanding.
An alternative view suggests that the learning mechanisms underlying language development are actually quite general. Although these learning abilities might be innate, their evolutionary development was not restricted to language learning. For example, researchers have demonstrated that the distributional learning mechanisms discussed earlier in this chapter also help infants track sequences of musical notes, visual shapes, and human actions (e.g., Fiser & Aslin, 2001; Kirkham, Slemmer, & Johnson, 2002; Roseberry et al., 2011; Saffran et al., 1999). Similarly, the fast-mapping mechanisms that support rapid word learning are also used by toddlers to learn facts about objects (Markson & Bloom, 1997). Also relevant is the fact that the “less is more” hypothesis for the critical period for language development, discussed earlier in this chapter, is not tied specifically to language (Newport, 1990). The ability to extract small chunks of information is likely useful in other domains as well, such as music, which also consists of small pieces (notes, chords) organized into higher-level structures (melody, harmony). Finally, recent theories concerning developmental language disorders (discussed in Box 6.5) invoke aspects of general cognitive function, not just language.
251
Throughout this chapter, we have emphasized both what is similar and what diverges across children and across cultures over the course of language development. The most significant individual differences fall under the category of developmental language disorders. These range from delays that often disappear by school age to life-long challenges.
Estimates of the number of preschoolers with language disorders range from 2% to 19% (H. D. Nelson et al., 2006). The size of this range reflects the fact that language disorders are often not diagnosed until a child enters school. Of this group, many of the children are considered late talkers. This label is applied to toddlers who are developing typically in other domains but whose vocabulary development is lagging at or below the 10th percentile. Some of these children are so-called “late bloomers” who will go on to have normal or near-normal language skills. Recent research shows that late-talking toddlers with better word recognition skills are the most likely to catch up (Fernald & Marchman, 2012). Children who fail to catch up—roughly 7% of school-aged children in the United States—are diagnosed with specific language impairment (SLI). These children exhibit challenges in many language-related tasks, including speech perception, word segmentation, and grammatical comprehension (e.g., Evans, Saffran, & Robe-Torres, 2009; Fonteneau & van der Lely, 2008; Rice, 2004; Ziegler et al., 2005). They may also exhibit more general challenges in working memory, sequence learning, and processing speed (e.g., Leonard et al., 2007; Tomblin, Mainela-Arnold, & Zhang, 2007).
Children diagnosed with Down syndrome, fragile-X syndrome, or autism spectrum disorders (ASD) tend to be significantly delayed across all aspects of language development, including both language production and comprehension. Indeed, challenges in communication are one of the diagnostic criteria for ASD. For children with ASD, early language abilities are predictive of later outcomes, including response to treatment (e.g., Stone & Yoder, 2001; Szatmari et al., 2003). Interestingly, younger siblings of children with ASD, who are themselves at greater risk for ASD, have a higher rate of language delay than do their peers (Gamliel et al., 2009).
Another group of children who may develop language disorders are deaf children. As we noted earlier, if these children have early exposure to a natural sign language like ASL, they will follow a typical language development trajectory. However, 90% of deaf children are born to hearing parents, and many of these children do not have access to sign language. In the absence of hearing, it is very difficult to learn a spoken language. An increasingly popular intervention for profoundly deaf infants, children, and adults is the use of cochlear implants (CIs), surgically implanted devices that translate auditory input into electrical stimulation of the auditory nerve. The signal that CIs provide is quite degraded relative to typical acoustic hearing. Nevertheless, many deaf infants and children are able to learn spoken language with the help of CIs, though the level of success varies widely across individuals. Consistent with other critical-period findings, implantation at a young age (e.g., prior to age 3) is better than at later ages (e.g., Houston & Miyamoto, 2010). Even at the outset of learning, though, deaf infants and toddlers who perceive speech through CIs are less accurate and slower to recognize words than are their hearing peers (Grieco-Calub, Litovsky, & Saffran, 2009). Providing these learners with bilingual input (natural sign language and spoken language through CIs) may offer the most successful route to language acquisition.
252
connectionism a type of information-processing approach that emphasizes the simultaneous activity of numerous interconnected processing units
As in other areas of child development, computational modeling has played an important role in the development of modern theoretical perspectives. By using computational models, researchers can specify both the innate structure and the environmental input to a computerized learner and attempt to determine what is crucial when simulating children’s language acquisition. One influential perspective oriented around computational modeling is connectionism, a type of information-processing theory that emphasizes the simultaneous activity of numerous interconnected processing units. Connectionist researchers have developed computer simulations of various aspects of cognitive development, including language acquisition (e.g., Elman et al., 1996). The software learns from experience, gradually strengthening certain connections among units in ways that mimic children’s developmental progress. Connectionist accounts have achieved impressive success with respect to modeling specific aspects of language development, including children’s acquisition of the past tense in English and the development of the shape bias for word learning (e.g., Rumelhart & McClelland, 1986; Samuelson, 2002). However, connectionist models are always open to criticism regarding the features that were built into the models in the first place (e.g., do they have the same “innate” constraints as infants?), and how well the input provided to them matches the input received by real children.
The process of comprehending and producing language, whether spoken or signed, involves the development of many different kinds of knowledge and skills. In the space of a few years, children take giant steps in mastering the phonology, semantics, syntax, and pragmatics of their native language. This remarkable achievement is made possible by the joint prerequisites of a human brain and exposure to human communication.
The current theoretical accounts of language development differ with respect to how much emphasis they put on nature and nurture. Nativists like Chomsky emphasize innate linguistic knowledge and language-specific learning mechanisms, whereas other theorists argue that language learning can emerge from general-purpose learning mechanisms. Children’s motivation to understand and interact with other people is also central to many theories. The vast literature on language development provides some support for all of these views, but none of them provide the full story of children’s acquisition of this vastly complex and arguably unique of all human abilities.