Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
letter
. 2010 Apr 28;107(18):E73. doi: 10.1073/pnas.1003007107

Mapping of heard speech into articulation information and speech acquisition

John R Skoyles 1,1
PMCID: PMC2889576  PMID: 20427741

Yuen et al. (1) find that hearing an “incongruent distractor” syllable shifts the tongue's place of articulation when pronouncing a second printed syllable. More specifically, hearing a /t/-initial phone in the distractor causes increased alveolar region tongue-palate contact when a subject seeks to pronounce a printed syllable with an initial /k/ or /s/. (The distractor/printed syllables were nonword stimuli such as /kib/ and /seeb/.) Yuen et al. (1) seek to interpret this effect as a speech phenomenon only in terms of potential speech perception processes. This interpretation overlooks the well-attested existence of a capability within the faculty of speech for overheard speech to be mapped directly and automatically into articulations independently of perception/comprehension.

Brain lesion research shows that such direct articulation mapping occurs in a separate path from speech perception (2). Research with speech shadowing shows that this mapping can be as fast or faster than that when people seek to produce as rapidly as possible the same speech sound (3). The echolalia found in some mentally retarded individuals shows the independence of this mapping from speech comprehension because such speech repetition occurs in the absence of other speech abilities (4).

Although not so obviously a part of speech as its production and perception, the capacity for a speaker to directly map input into output is critical for its acquisition and the intergenerational transmission of spoken vocabulary. This is because it enables a speaker after hearing the pronunciation of an unfamiliar word to incorporate that pronunciation directly into their conversation and spoken vocabulary (5). Consistent with this, competence in nonword repetition is the best predictor (apart from age) of the rate with which children acquire new spoken vocabulary (5). The mapping of novel heard pronunciations strategically needs to be processed automatically upon all words. This is because speakers cannot know in advance that a word is unfamiliar. As a result they can know a word is in need of input into motor output mapping only after the opportunity to do this has gone. Thus, speakers if they are to incorporate unfamiliar words into their spoken vocabulary must by default map all spoken input to ensure the needing mapping has been done.

Yuen et al. (1) argue that no strategic need exists for subjects in their experiment to recruit the motor system into their task in regard to the distractor words. However, the existence of speech mapping as a separate process from speech perception/production provides a parsimonious—although overlooked—explanation. No reason therefore exists for Yuen et al. (1) to interpret their findings, “in the context of a broader emerging framework, whereby the perception of action entails activation of the motor system.” Such general and non-speech-specific processes might exist but speech repetition processes linked to the need to map unfamiliar speech input into speech output for the propagation and learning of spoken vocabulary are already sufficient to account for their data.

Footnotes

The author declares no conflict of interest.

References

  • 1.Yuen I, Davis MH, Brysbaert M, Rastle K. Activation of articulatory information in speech perception. Proc Natl Acad Sci USA. 2010;107:592–597. doi: 10.1073/pnas.0904774107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.McCarthy R, Warrington EK. A two-route model of speech production, evidence from aphasia. Brain. 1984;107:463–485. doi: 10.1093/brain/107.2.463. [DOI] [PubMed] [Google Scholar]
  • 3.Porter RJ, Lubker JF. Rapid reproduction of vowel-vowel sequences: Evidence for a fast and direct acoustic-motoric linkage in speech. J Speech Hear Res. 1980;23:593–602. [PubMed] [Google Scholar]
  • 4.Fay WH, Coleman RO. A human sound transducer/ reproducer: Temporal capabilities of a profoundly echolatic child. Brain Lang. 1977;4:396–402. doi: 10.1016/0093-934x(77)90034-7. [DOI] [PubMed] [Google Scholar]
  • 5.Masur EF. Infants’ early verbal imitation and their later lexical development. Merrill-Palmer Q. 1995;41:286–306. [Google Scholar]

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES