Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2016 Dec 1.
Published in final edited form as: Trends Cogn Sci. 2015 Oct 9;19(12):713–716. doi: 10.1016/j.tics.2015.09.006

Apples and Oranges: Developmental Discontinuities in Spoken-Language Processing?

Sarah C Creel 1,*, Carolyn Quam 2
PMCID: PMC4673018  NIHMSID: NIHMS728519  PMID: 26456261

Abstract

Much research focuses on speech processing in infancy, sometimes generating the impression that speech-sound categories do not develop further. Yet other studies suggest substantial plasticity throughout mid-childhood. Differences between infant versus child and adult experimental methods currently obscure how language processing changes across childhood, calling for approaches that span development.

The Missing Link

While we know a great deal about the beginnings of spoken-language comprehension, a missing puzzle piece is the link from early perceptual sensitivities and word learning, to later perceptual memory, associative learning, and word-meaning mapping. Not uncommonly, language-development conferences and journals contain claims that young infants are speech ‘experts’. This emphasis on early abilities minimizes evidence of substantial plasticity in speech perception through primary school [1,2]. In the language-development literature, the fact that infants only discriminate sounds or learn words under a very narrow set of circumstances is often regarded as a mere methodological concern – yet it may reflect something about the learning infants still need to do to reach maturity.

Here we discuss a major barrier to seeing the continuous developmental picture of speech perception and word-meaning mapping: most experimental paradigms used with infants versus older learners differ substantially. Infant-appropriate tasks are exquisitely sensitive to subtle or incomplete knowledge – by necessity – while tasks used with older children and adults often require explicit indications of more complex knowledge or abilities. As a result, comparing learning or competence in infancy versus later childhood sometimes seems like comparing apples and oranges. This makes it difficult to characterize what changes across development. In an informal database search of speech-perception/word-recognition from birth to 5 years, only 15 of 512 total papers (3%) used the same method with both infants under 12 months old and preschoolers or older learners. In the following, we discuss two major causes of paradigm-based discrepancies across development, highlight some consequences of these discrepancies, and provide concrete suggestions for bridging the gap.

Causes of Paradigm-Based Discrepancies across Development

Different Tasks are Used to Test Purportedly the Same Abilities

Investigations of ostensibly the same ability in infants versus older learners (e.g., sound discrimination or word learning) often differ in both how sounds are presented to listeners, and what types of responses are prompted. Because young infants cannot indicate knowledge explicitly, infant tasks rely on implicit responses, such as looking times. Infants’ sound discrimination is often tested by presenting one sound category repeatedly (such as ‘ba ba ba…’, or a repeating set of voices [3]) until signs of boredom appear (often, ceasing to look at a picture on a computer screen). Then, the sound is switched (‘da’) to see whether infants ‘perk up’ and begin looking again, indicating discrimination (of ‘ba’ versus ‘da’). Repeated exposure to the initial sound ‘ba’ dramatically increases the memory strength for that sound. This ‘pumped-up’ memory may be crucial to detecting when the sound changes. Supporting this are findings from a neural method called mismatch negativity (MMN). MMN is an electrical brain response to a mismatching element (B) in a sequence such as A, A, A, … B. The MMN response is thought to reflect sensory or perceptual memory. Critically, the more A's that precede the oddball B element, the stronger the neural response [4].

While discrimination tasks are sometimes used with older learners, and are also argued to assess short-term sound discrimination, they are typically very different from discrimination tests with infants. They often present just two sounds (e.g., ‘ba ba’ or ‘ba da’), and prompt a ‘Same’ or ‘Different’ judgment. Infant paradigms are likely more sensitive, both because they do not require an explicit response – reducing task demands – and because they build up a stronger representation of the first sound as it is repeated throughout the initial exposure.

Older Learners are Tested on More Complex Abilities

Another way in which tasks for older learners are more demanding is that older learners are often required not just to differentiate sounds, but to learn an association between sounds and objects – for example learning names for objects. Whereas infants are most often tested on sound discrimination, older learners are often asked to learn words or otherwise associate sounds to visual categories, which entails learning and differentiating several sound-meaning mappings. For example, the ability to tell apart different voices is typically assessed via simple discrimination in infancy [3], but older learners might be asked to associate voices to different characters [5]. There are some infant tasks that require associative learning [6]. However, infants typically complete just a few brief test items [6], while word-learning tasks (or other associative-learning tasks) with older learners typically require participants to maintain recently learned associations for dozens of test items. This is a significant difference considering that newly-formed word-meaning mappings in children may be relatively fragile in memory, decaying in as little as 5 min [7].

Consequences of Paradigm-Based Discrepancies: Paradoxical Developmental Patterns

The above differences in infant and child tasks and their memory requirements can create a puzzling picture of how perceptual abilities change between infancy and adulthood – in some cases leading to the appearance of greater sophistication in infancy than in later childhood. For example, infant learners discriminate a new voice among other voices [3], yet preschoolers have difficulty mapping two different voices to two different characters [5].

In another well-known case, by 12 months, infants have tuned their discrimination of similar-sounding consonants such as ‘b’ versus ‘d’ to their native language, only discriminating native contrasts [8]. However, children cannot reliably distinguish newly-learned word labels differing only in these same sounds until 17–20 months [6]. Researchers debate whether this is due to infants’ immature phonological knowledge [9], or because task difficulty masks good phonological knowledge [10]. In a third case, infants respond with different facial expressions to positive versus negative adult speech [11]. Yet well into preschool, children struggle to associate emotional speech patterns with the concepts ‘happy’ versus ‘sad’ [12].

Superficially, these three cases suggest that, paradoxically, infants are more mature listeners than older children, who recover abilities by adulthood – a U-shaped developmental pattern (Box 1). However, apparent U-shaped patterns may well be attributable to the increased complexity or memory demands of tasks used with older learners. The true developmental trajectory may be a slow, steady increase in perceptual knowledge. Unfortunately, because relatively little work bridges age and task divides, it is often difficult to discern what changes between infancy and later childhood.

Box 1. Is There a Developmental Discontinuity in Using Pitch Cues to Vocal Emotion?

Seminal work [11] has demonstrated that infants respond to the emotional content and pragmatic functions (e.g., approval versus disapproval) of adult speech. By as early as 5 months, infants smile more to parents’ expressions of approval, which tend to be produced higher in the vocal range and with more sweeping contours (Figure IA, left) and are more likely to show negative affect in response to prohibitions (Figure IA, right) [11]. While infants are sensitive to emotions and pragmatic functions in speech, older children tested in more complex tasks show protracted development of interpretation of vocal emotion. For example, several studies have shown that preschool children weight the words that are spoken more heavily than the vocal emotion when these cues conflict. Even when words are intentionally neutral (Figure IB), children cannot map vocal expressions to the emotions ‘happy’ or ‘sad’ until about age 4 years (Figure IC) [12].

Proposed Solutions

Task differences outlined above currently obscure the developmental picture of perceptual and memory abilities that support language. The following measures would help to better characterize the full course of perceptual development.

First, more studies should focus on a longer span of development. One approach is to use different tasks with the same age group to obtain task equivalencies, facilitating cross-age comparisons across studies. Another approach is to relate each study's findings to existing literature on earlier and later ages, carefully considering differences in tasks, memory demands, and materials.

Second, matched methodologies across wide age ranges are possible with certain paradigms. While adults are unlikely to demonstrate sound discrimination by sucking on a non-nutritive nipple (a common infant paradigm), nor are infants likely to spend an hour in the laboratory learning 40 new words, other paradigms can better connect disparate ages. These include neural methods, such as MMN, to assess perceptual sensitivity and eye tracking, which can be used from infancy through adulthood to test learned associations between auditory words and their pictured images. When age differences emerge in matched methodologies, we can have more confidence that they reflect real developmental changes rather than methodological confounds. Then, we can begin trying to explain their causes and consequences.

Figure I. Apparent Developmental Discontinuity in Sensitivity to Vocal Emotion.

Figure I

(A) Infants’ facial expressions can demonstrate sensitivity to pragmatic functions of speech such as approval (left) or disapproval (right). (B) Happy (top) versus sad (bottom) vocal expressions. (C) Gradual improvement in preschoolers’ ability to identify the vocal emotions in (B). Diagonal line denotes linear fit; thick horizontal line denotes chance responding (0.50). (B, C) Adapted from [12], with permission from John Wiley & Sons.

Acknowledgments

C.Q. was supported by National Institutes of Health K99 DC013795. S.C.C. was supported by National Science Foundation (NSF) BCS-1057080 and NSF BCS-1230003. Thanks to Jeff Elman for comments on an earlier draft of this paper.

References

  • 1.Hazan V, Barrett S. The development of phonemic categorization in children aged 6–12. J. Phon. 2000;28:377–396. [Google Scholar]
  • 2.Oh GE, et al. A one-year longitudinal study of English and Japanese vowel production by Japanese adults and children in an English-speaking setting. J. Phon. 2011;39:156–157. doi: 10.1016/j.wocn.2011.01.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Johnson EK, et al. Infant ability to tell voices apart rests on language experience. Dev. Sci. 2011;14:1002–1011. doi: 10.1111/j.1467-7687.2011.01052.x. [DOI] [PubMed] [Google Scholar]
  • 4.Cowan N, et al. Memory prerequisites of mismatch negativity in the auditory event-related potential (ERP) J. Exp. Psychol. Learn. Mem. Cogn. 1993;19:909–921. doi: 10.1037//0278-7393.19.4.909. [DOI] [PubMed] [Google Scholar]
  • 5.Creel SC, Jimenez SR. Differences in talker recognition by preschoolers and adults. J. Exp. Child Psychol. 2012;113:487–509. doi: 10.1016/j.jecp.2012.07.007. [DOI] [PubMed] [Google Scholar]
  • 6.Stager CL, Werker JF. Infants listen for more phonetic detail in speech perception than in word-learning tasks. Nature. 1997;388:381–382. doi: 10.1038/41102. [DOI] [PubMed] [Google Scholar]
  • 7.Horst JS, Samuelson LK. Fast mapping but poor retention by 24-month-old infants. Infancy. 2008;13:128–157. doi: 10.1080/15250000701795598. [DOI] [PubMed] [Google Scholar]
  • 8.Werker JF, Tees R. Cross-language speech perception: evidence for perceptual reorganization during the first year of life. Infant Behav. Dev. 1984;7:49–63. [Google Scholar]
  • 9.Apfelbaum KS, McMurray B. Using variability to guide dimensional weighting: associative mechanisms in early word learning. Cogn. Sci. 2011;35:1105–1138. doi: 10.1111/j.1551-6709.2011.01181.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Werker JF, Fennell CT. Infant speech perception and later language acquisition: methodological underpinnings. In: Colombo J, et al., editors. Infant Pathways to Language: Methods, Models, and Research Directions. Lawrence Erlbaum Associates; 2008. pp. 85–98. [Google Scholar]
  • 11.Fernald A. Approval and disapproval: infant responsiveness to vocal affect in familiar and unfamiliar languages. Child Dev. 1993;64:657–674. [PubMed] [Google Scholar]
  • 12.Quam C, Swingley D. Development in children's interpretation of pitch cues to emotions. Child Dev. 2012;83:236–250. doi: 10.1111/j.1467-8624.2011.01700.x. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES