Abstract
Understanding speech in a background of competing noise is challenging, especially for individuals with hearing loss or deficits in auditory processing ability. The ability to hear in background noise cannot be predicted from the audiogram, an assessment of peripheral hearing ability; therefore, it is important to consider the impact of central and cognitive factors on speech-in-noise perception. Auditory processing in complex environments is reflected in neural encoding of pitch, timing, and timbre, the crucial elements of speech and music. Musical expertise in processing pitch, timing, and timbre may transfer to enhancements in speech-in-noise perception due to shared neural pathways for speech and music. Through cognitive-sensory interactions, musicians develop skills enabling them to selectively listen to relevant signals embedded in a network of melodies and harmonies, and this experience leads in turn to enhanced ability to focus on one voice in a background of other voices. Here we review recent work examining the biological mechanisms of speech and music perception and the potential for musical experience to ameliorate speech-in-noise listening difficulties.
Keywords: Brain stem, music, speech in noise, timing, pitch
From busy classrooms to crowded restaurants, our world is a noisy place. To participate fully in today’s society, we must selectively attend and listen to one voice among competing voices and other noises. Listening in noise is particularly challenging for clinical populations, such as children with learning impairments (i.e., dyslexia, autism spectrum disorder, attention deficit disorder, auditory processing disorder, specific language impairment)1–8 and older adults with and without hearing loss.9–13 The problem with hearing in noise is often not an issue of audibility—the individual can hear but cannot understand what is being said. Other factors including central auditory processing and cognitive abilities interact with peripheral hearing status, determining the degree of successful communication in noise. To understand the message, the individual must process the three basic components of sound: pitch, timing, and timbre. Processing these elements is important for understanding both speech and music, and the musician’s ability to tune into music (along with enhanced working memory and attention) appears to transfer to the ability to tune into speech in background noise. Music is said to be biologically powerful,14,15 with effects of musical training on neural functions lasting for a lifetime. The auditory system is malleable and adaptive, making constant changes based on immediate sensory and behavioral contexts and past experiences, including musical training.16 In this review, we summarize recent work examining the neural mechanisms of speech-in-noise (SIN) perception, including encoding of stimulus regularities, timing, and pitch, and the effects of musical experience on the neural encoding of speech, particularly in noise.
SUBCORTICAL APPROACH TO EXAMINING NEURAL SOUND ENCODING
The use of auditory brain stem responses to study the neural encoding of speech and music has been fruitful for several reasons. Unlike the cortical response, the brain stem response to complex stimuli (cABR) mimics the stimulus remarkably well, both visually and auditorially.17,18 The periodicity of the stimulus peaks can be seen in both the response waveform as well as the fast Fourier transform of the response (Fig. 1). In addition, when the brain stem response waveform is converted to a sound file, it actually sounds like the stimulus.18 The cABR is particularly suited for testing individuals as the response is highly reliable,19 and timing differences of fractions of milliseconds can be clinically significant.20–24 The brain stem faithfully represents the three main features of speech and music: timing (onsets/offsets and envelope of the response), pitch (encoding of the fundamental frequency), and timbre (harmonics).17,25 Moreover, the response originates primarily in the rostral brain stem, and the spectral and timing features such as those that contribute to the auditory perception of emotion are represented with remarkable fidelity.26 This transparency in the response then allows for the examination of musician differences in the subcortical encoding of these features.
The brain stem is traditionally referred to as the “old brain” and is considered to be evolutionarily ancient, part of the reptilian brain. Do not be misled; the brain stem is “smart,” part of a reciprocally interactive sensory and cognitive network, engaging in multiple aspects of the nervous system. Current views of the brain stem are changing, and there is evidence that the brain stem can reveal much to us about the processing of auditory signals. The brain stem’s multifaceted processing is evident given the anatomic complexity of the multiple nuclei and crossed and uncrossed pathways within the brain stem. In addition to the vast system of afferent fibers carrying sensory information to the midbrain (inferior colliculus) and the auditory cortex, there is an extensive system of descending efferent fibers that synapse all along the auditory pathway extending even to the outer hair cells of the basilar membrane.27 In fact, the efferent fiber count may actually exceed the number of afferent fibers, which is definitely counterintuitive if the principal purpose of the auditory system is to passively convey acoustic information from the receptor to the auditory cortex for final and more complex processing. Indeed, top-down tuning of sensory function modulates subcortical responses by sharpening tuning, augmenting stimulus features, increasing the signal-to-noise ratio, excluding irrelevant information, increasing response efficiency, controlling contextual influences, modulating plasticity, and promoting learning.26,28–31 For example, top-down modulation of the pathway from the cortex to the inferior colliculus has been effectively demonstrated in ferrets, in which pharmacological blockage impeded auditory learning.30
The effects of musical training on auditory brain stem processing are another demonstration of top-down modulation, and these effects are not limited to musical sound processing but generalize to speech encoding and other nonmusical neural functions. The facets of learning involved in musical training provide an explanation for these effects. Musical training requires active listening and engagement with musical sounds, including interpretation or mapping sounds to meaning, as reviewed in Kraus and Chandrasekaran.16 For example, the dynamics of music (tempo, volume, etc.) are used to signal mood and can differentiate otherwise identical note patterns. The ability to attach meaning to dynamics and other characteristics of music requires attention to the fine-grained properties of music (pitch, timing, and timbre) as well as cognitive skills relating to working memory and attention. Many aspects of musical training focus attention on these specific aspects of sound: the violinist must listen to pitch cues to tune his or her instrument, the instrumentalist in an orchestra monitors the timing cues of other instruments, and the conductor relies on timbre cues to distinguish between instruments.
TRANSFER EFFECTS OF MUSICAL EXPERIENCE
Musical experience generalizes to enhanced function in other areas of the auditory system, with previous studies showing transfer effects to cortical, cerebellar, and other areas.16 For example, musicians demonstrate enhanced performance on auditory-based cognitive tasks, such as auditory memory or attention.32,33 Interestingly, this advantage is not seen in the visual domain, suggesting that the musicians have selective improvements in the auditory domain rather than improved overall neural function. Length of musical training also relates to measures of auditory attention, indicating that this improved ability is likely not innate but rather develops with experience (Fig. 2).32 The effects of musical experience extend all the way to the cochlea,34–37 with stronger top-down activity demonstrated by greater olivocochlear efferent suppression of biomechanical activity in musicians compared with nonmusicians.
At the subcortical level, musicians have higher brain stem amplitudes in response to both music and speech than nonmusicians.38 In fact, musical training generalizes to linguistic pitch encoding, with musicians showing enhancements in subcortical pitch processing to the most complex vowel contours present in Mandarin Chinese, but not to simple contours,39 demonstrating that the enhancement is selective rather than an overall gain effect. Another demonstration of selective enhancement is seen in subcortical responses to a musical chord, in which musicians have larger responses to the harmonics of the upper tone in a musical chord, which are important for melody perception, but do not have larger responses to the harmonics of the lower tone or the fundamental frequency (F0) of either upper or lower tones.31 Similar results are found in response to emotional nuances in sound.26 Compared with nonmusicians, musicians have greater representation of the F0 in the response corresponding to the complex portion of a baby’s cry, but actually have smaller F0 representation for the periodic portion (easier to lock onto), suggesting that this training also results in greater processing efficiency (Fig. 3).
HEARING SPEECH IN NOISE
Understanding speech in a background of competing talkers is a highly complex task, in many ways similar to the task of the musician who must segregate and selectively attend to competing melodies from many different instruments in an orchestra or ensemble. As sound travels from the ear to the mid brain to the primary auditory cortex, the signal is processed and shaped by a myriad of interacting factors. We have already noted that musicians have enhanced auditory memory and attention, cognitive factors that also contribute to enhanced SIN perception. As stated previously, music and speech both consist of the basic elements of pitch, timing, and timbre. Here we provide examples of how the subcortical representation of these elements relates to SIN perception.
Detection of Stimulus Regularities
The brain is constantly bombarded by stimuli, which must be processed and interpreted. To provide some perspective, consider the amount of data that enters the brain in just 1 second. It is roughly estimated that in the visual system, one million retinal ganglion cells fire at a rate of 1 to 10 spikes per second, which at one bit per spike adds up to 10 million bits of data per second. In the auditory system, 30,000 auditory nerve fibers fire at a rate of 0 to 300 spikes per second, or 9 million bits of data per second at 1 bit per spike. It is impossible to process this much information in real time; therefore, our neural systems must have a mechanism for managing and interpreting these data. One of the ways this is accomplished is through detection of patterns or stimulus regularities in an ongoing stream of data, which has been demonstrated in animal40,41 and human studies.42–46
The brain stem’s response to stimulus regularities was evaluated in a group of normal-learning children ages 8 to 12.43 Fast Fourier transforms were calculated for responses recorded to a speech syllable [da] presented in a repetitive, unvarying condition and to the same syllable presented in a varying condition in which the [da] syllable was one of eight different consonant-vowel syllables. The representation of the lower harmonics (H2 and H4) was enhanced in the repetitive compared with the variable condition. Moreover, the extent of enhancement for the repetitive condition was related to a measure of SIN perception—scores on the Hearing in Noise Test (HINT; Biologic Systems Corp., Mundelein, IL; Fig. 4). Online adaptation and enhancement were also found for music (to a repeating melody) with the greatest amplitude increase occurring for the second repeated note within the melody.47 This ability to detect regularities is also related to a test of musical aptitude—the Rhythm Aptitude subtest of Gordon’s Intermediate Measures of Audiation test (GIA Publications, Chicago, IL)—indicating that active engagement with music may improve the ability to detect and encode sound patterns.48 The brain stem’s online detection of patterns, resulting in enhancement of predictable features of the auditory stimulus, is important for providing a stable representation of the stimulus for subsequent processing by the cortex, and a more stable representation in the cortex is likely more resistant to degradation by interfering noise.
Neural Timing
Timing is a crucial element for distinguishing speech sounds; for example, consonants such as [t] and [d] are differentiated solely by voice-onset time. The brain stem’s representation of timing differences is exquisitely precise,49 and it has been suggested that neural timing differences to speech stimuli can be used to assess children with reading disabilities or impairments in language or SIN perception.21–23,50,51
The brain stem nuclei are tonotopic52; therefore, brain stem responses to consonants differing in their formant trajectories during the transition from the consonant to the vowel (e.g., in the syllables [ba], [da], and [ga]) reflect timing differences between these consonants, with [ga] having the shortest peak latency, followed by [da] and then [ga]. These timing differences persist for the duration of the formant transition but disappear in the steady state, the unvarying region of the response corresponding to the vowel. The extent of these timing differences is related to SIN perception in children. Brain stem responses were recorded to the syllables [ba], [ga], and [da] in children ages 8 to 12, and a differentiation score was calculated based on how closely the responses followed the expected timing patterns. Higher differentiation scores were related to better SIN perception (based on HINT; Fig. 5).
Stop consonants are particularly vulnerable to the effects of noise, which causes delays in neural timing as well as decreases in response amplitudes.53–55 These delays are particularly evident in children with poor SIN perception. Brain stem responses to the syllable [da] were presented in quiet and in six-talker babble conditions to children who were divided into top and bottom SIN perception groups based on the HINT. The children in the bottom group had greater delays in response timing compared with those in the top group. For both groups, the noise effect was greatest shortly after the stimulus onset, delaying the peak latencies by a full millisecond. In the top SIN group, the latency shift then reduced and leveled off by 40 milliseconds into the response, but in the bottom SIN group the latency shift did not level off until 60 milliseconds. The first 60 milliseconds of the response correspond to the formant transition of the syllable, the most perceptually vulnerable region of the response,21,56–58 in which the formant frequencies are dynamic. By 60 milliseconds, the formant frequencies are static, making this region of the syllable easier to process (Fig. 6). Therefore, it is not surprising that the greatest group differences were found in the transition region of the response.
Pitch Representation
Pitch is another important element for SIN perception, allowing the listener to identify the speaker and to “tag” the speaker’s voice, thus facilitating the ability to follow one speaker’s voice from competing voices in a noisy environment such as a crowded restaurant or cocktail party. The F0 and lower harmonics, along with other acoustic ingredients, contribute to the perception of pitch.59,60 Psychophysical experiments have demonstrated that F0 differences facilitate the ability to distinguish between competing vowels or sentences.61–66
The perceptual benefit provided by F0 separation is apparent in neural processing.67 Both children and young adults who have greater F0 magnitudes in their cABRs also have better SIN perception.68,69 Song et al found that in young adults, the top SIN group (based on the Speech-in-Noise test [Quick-SIN]; Etymotic Research, Elmhurst, IL) had greater F0 magnitudes than the bottom SIN group in response to the speech syllable [da] presented in background babble.19 This difference in F0 encoding was seen in the transition but not the in steady-state region of the response. The formant transition carries phonetic information70 and its perception can easily be disrupted by noise.56,58 Similarly, in children, the top SIN group (based on HINT scores) had higher F0 magnitudes than the bottom group in the neural response corresponding to the formant transition.69
SIN Model
Structural equation modeling was used to quantify the relationships among factors contributing to SIN perception in children.71 The brain stem construct variable consisted of three factors: degree of neural timing shift, enhancement of stimulus regularities, and magnitude of pitch cues (e.g., F0 and lower harmonics), and the SIN construct was based on HINT scores. This model predicted 56% of the variance in SIN perception, with a p value of 0.033 (Fig. 7). Ongoing work using structural equation modeling to evaluate these relationships in older adults will include cognitive factors and peripheral hearing status.
EFFECTS OF MUSICAL TRAINING
We know that musicians have advantages for encoding both music and speech, but do these advantages extend to SIN perception? As mentioned previously, musicians are particularly adept at extracting relevant signals from a soundscape; therefore, does this skill transfer to the ability to selectively attend to one voice from among competing voices? This musician advantage has been demonstrated for SIN perception as well as auditory working memory,72 a skill that is highly related to SIN perception.73,74 In a group of young adults, musicians had higher scores on SIN assessments (HINT and Quick-SIN) as well as working memory (cluster score from the Woodcock-Johnson Test III Test of Cognitive Abilities75) compared with nonmusicians, providing further evidence that musical training transfers to nonmusical domains. Importantly, years of musical practice correlated with SIN perception, supporting the idea that the extent of training contributes to the advantage for hearing in noise.72
A neural basis for this musician benefit has been demonstrated in the subcortical encoding of sounds. Background (babble) noise had a less degradative effect on brain stem responses in musicians than in nonmusicians.76 The brain stem responses for musicians and nonmusicians were similar in quiet, but noise had a greater (disruptive) effect on the morphology, size, timing, and frequency content of nonmusician’s responses than of musicians. Parbery-Clark et al quantified this effect using stimulus-to-response correlations, which compared the timing and morphology of the response waveform to the stimulus waveform.72 The stimulus-to-response correlations were essentially equivalent in quiet, but in noise the musician’s r value changed only slightly, and the non-musician’s r value dropped significantly. This musician difference also was seen for neural timing: onset and transition peaks were more delayed in noise for nonmusicians than for musicians, and these latency delays were related to SIN perception (Fig. 8). Overall, these results demonstrate that musicians have perceptual and neural advantages for processing SIN.
CONCLUSION
Subcortical responses to complex stimuli can be used to assess auditory processing ability in individuals suspected of deficits in SIN perception. Factors such as the neural enhancement stimulus regularities, strength of neural timing (differentiation of stop consonants and decreased noise-induced delays), and representation of the F0 are important neural signatures associated with better ability to hear in noise. Musicians have enhanced SIN abilities, both perceptually and biologically, indicating that musical training may ameliorate the deleterious effects of background noise on speech understanding. Years of musical practice are related to SIN perception, suggesting that music may be an effective vehicle for developing the auditory skills necessary for hearing well in background noise. Although the examples provided in this review represented lifelong experience, there is also evidence of subcortical malleability following short-term training.77–83 Future work is needed to determine whether musically based auditory training can be used as a management strategy for improving SIN perception in children and older adults who experience particular challenges in background noise.
Acknowledgments
This work was funded by the National Institutes of Health (RO1 DC01510) and the National Science Foundation SGER (0842376). We would especially like to thank the members of the Auditory Neuroscience Laboratory and our participants for their contributions to this work.
Footnotes
Proceedings of the Widex Pediatric Audiology Congress; Guest Editor, André M. Marcoux, Ph.D.
References
- 1.Bradlow AR, Kraus N, Hayes E. Speaking clearly for children with learning disabilities: sentence perception in noise. J Speech Lang Hear Res. 2003;46(1):80–97. doi: 10.1044/1092-4388(2003/007). [DOI] [PubMed] [Google Scholar]
- 2.Geffner D, Lucker JR, Koch W. Evaluation of auditory discrimination in children with ADD and without ADD. Child Psychiatry Hum Dev. 1996;26(3):169–179. doi: 10.1007/BF02353358. [DOI] [PubMed] [Google Scholar]
- 3.Lagacé J, Jutras B, Gagné JP. Auditory processing disorder and speech perception problems in noise: finding the underlying origin. Am J Audiol. 2010;19(1):17–25. doi: 10.1044/1059-0889(2010/09-0022). [DOI] [PubMed] [Google Scholar]
- 4.Moore DR, Ferguson MA, Edmondson-Jones AM, Ratib S, Riley A. Nature of auditory processing disorder in children. Pediatrics. 2010;126(2):e382–e390. doi: 10.1542/peds.2009-2826. [DOI] [PubMed] [Google Scholar]
- 5.Alcántara JI, Weisblatt EJ, Moore BC, Bolton PF. Speech-in-noise perception in high-functioning individuals with autism or Asperger’s syndrome. J Child Psychol Psychiatry. 2004;45(6):1107–1114. doi: 10.1111/j.1469-7610.2004.t01-1-00303.x. [DOI] [PubMed] [Google Scholar]
- 6.Ziegler JC, Pech-Georgel C, George F, Lorenzi C. Speech-perception-in-noise deficits in dyslexia. Dev Sci. 2009;12(5):732–745. doi: 10.1111/j.1467-7687.2009.00817.x. [DOI] [PubMed] [Google Scholar]
- 7.Ziegler JC, Pech-Georgel C, George F, Alario F-X, Lorenzi C. Deficits in speech perception predict language learning impairment. Proc Natl Acad Sci U S A. 2005;102(39):14110–14115. doi: 10.1073/pnas.0504446102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Lewis D, Hoover B, Choi S, Stelmachowicz P. Relationship between speech perception in noise and phonological awareness skills for children with normal hearing. Ear Hear. 2010;31(6):761–768. doi: 10.1097/AUD.0b013e3181e5d188. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Cruickshanks KJ, Wiley TL, Tweed TS, et al. The Epidemiology of Hearing Loss Study. Prevalence of hearing loss in older adults in Beaver Dam, Wisconsin. Am J Epidemiol. 1998;148(9):879–886. doi: 10.1093/oxfordjournals.aje.a009713. [DOI] [PubMed] [Google Scholar]
- 10.Humes LE. Speech understanding in the elderly. J Am Acad Audiol. 1996;7(3):161–167. [PubMed] [Google Scholar]
- 11.Souza PE, Boike KT, Witherell K, Tremblay K. Prediction of speech recognition from audibility in older listeners with hearing loss: effects of age, amplification, and background noise. J Am Acad Audiol. 2007;18(1):54–65. doi: 10.3766/jaaa.18.1.5. [DOI] [PubMed] [Google Scholar]
- 12.Gordon-Salant S. Hearing loss and aging: new research findings and clinical implications. J Rehabil Res Dev. 2005;42(4 Suppl 2):9–24. doi: 10.1682/jrrd.2005.01.0006. [DOI] [PubMed] [Google Scholar]
- 13.Dubno JR, Dirks DD, Morgan DE. Effects of age and mild hearing loss on speech recognition in noise. J Acoust Soc Am. 1984;76(1):87–96. doi: 10.1121/1.391011. [DOI] [PubMed] [Google Scholar]
- 14.Patel A. Music, Language, and the Brain. New York, NY: Oxford University Press, Inc; 2010. [Google Scholar]
- 15.Peretz I. The nature of music from a biological perspective. Cognition. 2006;100(1):1–32. doi: 10.1016/j.cognition.2005.11.004. [DOI] [PubMed] [Google Scholar]
- 16.Kraus N, Chandrasekaran B. Music training for the development of auditory skills. Nat Rev Neurosci. 2010;11(8):599–605. doi: 10.1038/nrn2882. [DOI] [PubMed] [Google Scholar]
- 17.Skoe E, Kraus N. Auditory brain stem response to complex sounds: a tutorial. Ear Hear. 2010;31(3):302–324. doi: 10.1097/AUD.0b013e3181cdb272. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Galbraith GC, Arbagey PW, Branski R, Comerci N, Rector PM. Intelligible speech encoded in the human brain stem frequency-following response. Neuroreport. 1995;6(17):2363–2367. doi: 10.1097/00001756-199511270-00021. [DOI] [PubMed] [Google Scholar]
- 19.Song JH, Nicol T, Kraus N. Test-retest reliability of the speech-evoked auditory brainstem response. Clin Neurophysiol. 2010 doi: 10.1016/j.clinph.2010.07.009. In press Corrected Proof. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Hall J. New Handbook of Auditory Evoked Responses. Boston, MA: Allyn & Bacon; 2007. [Google Scholar]
- 21.Basu M, Krishnan A, Weber-Fox C. Brainstem correlates of temporal auditory processing in children with specific language impairment. Dev Sci. 2010;13(1):77–91. doi: 10.1111/j.1467-7687.2009.00849.x. [DOI] [PubMed] [Google Scholar]
- 22.Banai K, Hornickel J, Skoe E, Nicol T, Zecker S, Kraus N. Reading and subcortical auditory function. Cereb Cortex. 2009;19(11):2699–2707. doi: 10.1093/cercor/bhp024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Billiet CR, Bellis TJ. The relationship between brainstem temporal processing and performance on tests of central auditory function in children with reading disorders. J Speech Lang Hear Res. 2011;54(1):228–242. doi: 10.1044/1092-4388(2010/09-0239). [DOI] [PubMed] [Google Scholar]
- 24.Wible B, Nicol T, Kraus N. Atypical brainstem representation of onset and formant structure of speech sounds in children with language-based learning problems. Biol Psychol. 2004;67(3):299–317. doi: 10.1016/j.biopsycho.2004.02.002. [DOI] [PubMed] [Google Scholar]
- 25.Dhar S, Abel R, Hornickel J, et al. Exploring the relationship between physiological measures of cochlear and brainstem function. Clin Neurophysiol. 2009;120(5):959–966. doi: 10.1016/j.clinph.2009.02.172. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Strait DL, Kraus N, Skoe E, Ashley R. Musical experience and neural efficiency: effects of training on subcortical processing of vocal expressions of emotion. Eur J Neurosci. 2009;29(3):661–668. doi: 10.1111/j.1460-9568.2009.06617.x. [DOI] [PubMed] [Google Scholar]
- 27.Gao E, Suga N. Experience-dependent plasticity in the auditory cortex and the inferior colliculus of bats: role of the corticofugal system. Proc Natl Acad Sci U S A. 2000;97(14):8081–8086. doi: 10.1073/pnas.97.14.8081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Suga N, Xiao Z, Ma X, Ji W. Plasticity and corticofugal modulation for hearing in adult animals. Neuron. 2002;36(1):9–18. doi: 10.1016/s0896-6273(02)00933-9. [DOI] [PubMed] [Google Scholar]
- 29.Luo F, Wang Q, Kashani A, Yan J. Corticofugal modulation of initial sound processing in the brain. J Neurosci. 2008;28(45):11615–11621. doi: 10.1523/JNEUROSCI.3972-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Bajo VM, Nodal FR, Moore DR, King AJ. The descending corticocollicular pathway mediates learning-induced auditory plasticity. Nat Neurosci. 2010;13(2):253–260. doi: 10.1038/nn.2466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Lee KM, Skoe E, Kraus N, Ashley R. Selective subcortical enhancement of musical intervals in musicians. J Neurosci. 2009;29(18):5832–5840. doi: 10.1523/JNEUROSCI.6133-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Strait DL, Kraus N, Parbery-Clark A, Ashley R. Musical experience shapes top-down auditory mechanisms: evidence from masking and auditory attention performance. Hear Res. 2010;261(1–2):22–29. doi: 10.1016/j.heares.2009.12.021. [DOI] [PubMed] [Google Scholar]
- 33.Chan AS, Ho Y-C, Cheung M-C. Music training improves verbal memory. Nature. 1998;396(6707):128–128. doi: 10.1038/24075. [DOI] [PubMed] [Google Scholar]
- 34.Micheyl C, Carbonnel O, Collet L. Medial olivocochlear system and loudness adaptation: differences between musicians and non-musicians. Brain Cogn. 1995;29(2):127–136. doi: 10.1006/brcg.1995.1272. [DOI] [PubMed] [Google Scholar]
- 35.Micheyl C, Khalfa S, Perrot X, Collet L. Difference in cochlear efferent activity between musicians and non-musicians. Neuroreport. 1997;8(4):1047–1050. doi: 10.1097/00001756-199703030-00046. [DOI] [PubMed] [Google Scholar]
- 36.Perrot X, Micheyl C, Khalfa S, Collet L. Stronger bilateral efferent influences on cochlear biomechanical activity in musicians than in non-musicians. Neurosci Lett. 1999;262(3):167–170. doi: 10.1016/s0304-3940(99)00044-0. [DOI] [PubMed] [Google Scholar]
- 37.Brashears SM, Morlet TG, Berlin CI, Hood LJ. Olivocochlear efferent suppression in classical musicians. J Am Acad Audiol. 2003;14(6):314–324. [PubMed] [Google Scholar]
- 38.Musacchia G, Sams M, Skoe E, Kraus N. Musicians have enhanced subcortical auditory and audiovisual processing of speech and music. Proc Natl Acad Sci U S A. 2007;104(40):15894–15898. doi: 10.1073/pnas.0701498104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Wong PCM, Skoe E, Russo NM, Dees T, Kraus N. Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nat Neurosci. 2007;10(4):420–422. doi: 10.1038/nn1872. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Toro JM, Trobalón JB. Statistical computations over a speech stream in a rodent. Percept Psychophys. 2005;67(5):867–875. doi: 10.3758/bf03193539. [DOI] [PubMed] [Google Scholar]
- 41.Hauser MD, Newport EL, Aslin RN. Segmentation of the speech stream in a non-human primate: statistical learning in cotton-top tamarins. Cognition. 2001;78(3):B53–B64. doi: 10.1016/s0010-0277(00)00132-3. [DOI] [PubMed] [Google Scholar]
- 42.Saffran JR, Aslin RN, Newport EL. Statistical learning by 8-month-old infants. Science. 1996;274(5294):1926–1928. doi: 10.1126/science.274.5294.1926. [DOI] [PubMed] [Google Scholar]
- 43.Chandrasekaran B, Hornickel J, Skoe E, Nicol TG, Kraus N. Context-dependent encoding in the human auditory brainstem relates to hearing speech in noise: implications for developmental dyslexia. Neuron. 2009;64(3):311–319. doi: 10.1016/j.neuron.2009.10.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Sussman ES, Bregman AS, Wang WJ, Khan FJ. Attentional modulation of electrophysiological activity in auditory cortex for unattended sounds within multistream auditory environments. Cogn Affect Behav Neurosci. 2005;5(1):93–110. doi: 10.3758/cabn.5.1.93. [DOI] [PubMed] [Google Scholar]
- 45.Sussman E, Steinschneider M. Neurophysiological evidence for context-dependent encoding of sensory input in human auditory cortex. Brain Res. 2006;1075(1):165–174. doi: 10.1016/j.brainres.2005.12.074. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Griffiths TD, Uppenkamp S, Johnsrude I, Josephs O, Patterson RD. Encoding of the temporal regularity of sound in the human brainstem. Nat Neurosci. 2001;4(6):633–637. doi: 10.1038/88459. [DOI] [PubMed] [Google Scholar]
- 47.Skoe E, Kraus N. Hearing it again and again: online subcortical plasticity in humans. PLoS ONE. 2010;5(10):e13645. doi: 10.1371/journal.pone.0013645. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Strait DL, Ashley R, Hornickel J, Kraus N. Context-dependent encoding of speech in the human auditory brainstem as a marker of musical aptitude. Presented at: ARO 33rd MidWinter Meeting; 2010; Anaheim, CA. [Google Scholar]
- 49.Tzounopoulos T, Kraus N. Learning to encode timing: mechanisms of plasticity in the auditory brainstem. Neuron. 2009;62(4):463–469. doi: 10.1016/j.neuron.2009.05.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Hornickel J, Skoe E, Nicol T, Zecker S, Kraus N. Subcortical differentiation of stop consonants relates to reading and speech-in-noise perception. Proc Natl Acad Sci U S A. 2009;106(31):13022–13027. doi: 10.1073/pnas.0901123106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Anderson S, Skoe E, Chandrasekaran B, Kraus N. Neural timing is linked to speech perception in noise. J Neurosci. 2010;30(14):4922–4926. doi: 10.1523/JNEUROSCI.0107-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Kandler K, Clause A, Noh J. Tonotopic reorganization of developing auditory brainstem circuits. Nat Neurosci. 2009;12(6):711–717. doi: 10.1038/nn.2332. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Burkard RF, Sims D. A comparison of the effects of broadband masking noise on the auditory brainstem response in young and older adults. Am J Audiol. 2002;11(1):13–22. doi: 10.1044/1059-0889(2002/004). [DOI] [PubMed] [Google Scholar]
- 54.Russo N, Nicol T, Musacchia G, Kraus N. Brainstem responses to speech syllables. Clin Neurophysiol. 2004;115(9):2021–2030. doi: 10.1016/j.clinph.2004.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Cunningham J, Nicol TG, Zecker SG, Bradlow AR, Kraus N. Neurobiologic responses to speech in noise in children with learning problems: deficits and strategies for improvement. Clin Neurophysiol. 2001;112(5):758–767. doi: 10.1016/s1388-2457(01)00465-5. [DOI] [PubMed] [Google Scholar]
- 56.Tallal P, Stark RE. Speech acoustic-cue discrimination abilities of normally developing and language-impaired children. J Acoust Soc Am. 1981;69(2):568–574. doi: 10.1121/1.385431. [DOI] [PubMed] [Google Scholar]
- 57.Tallal P, Piercy M. Developmental aphasia: rate of auditory processing and selective impairment of consonant perception. Neuropsychologia. 1974;12(1):83–93. doi: 10.1016/0028-3932(74)90030-x. [DOI] [PubMed] [Google Scholar]
- 58.Hedrick MS, Younger MS. Perceptual weighting of stop consonant cues by normal and impaired listeners in reverberation versus noise. J Speech Lang Hear Res. 2007;50(2):254–269. doi: 10.1044/1092-4388(2007/019). [DOI] [PubMed] [Google Scholar]
- 59.Fellowes JM, Remez RE, Rubin PE. Perceiving the sex and identity of a talker without natural vocal timbre. Percept Psychophys. 1997;59(6):839–849. doi: 10.3758/bf03205502. [DOI] [PubMed] [Google Scholar]
- 60.Meddis R, O’Mard L. A unitary model of pitch perception. J Acoust Soc Am. 1997;102(3):1811–1820. doi: 10.1121/1.420088. [DOI] [PubMed] [Google Scholar]
- 61.Bird J, Darwin CJ. Effects of a difference in fundamental frequency in separating two sentences. In: Palmer AR, et al., editors. Psychophysical and Physiological Advances in Hearing. London, United Kingdom: Whurr; 1998. pp. 263–269. [Google Scholar]
- 62.Assmann PF, Summerfield Q. Perceptual segregation of concurrent vowels. J Acoust Soc Am. 1987;82:S120. [Google Scholar]
- 63.Culling JF, Darwin CJ. Perceptual separation of simultaneous vowels: within and across-formant grouping by F0. J Acoust Soc Am. 1993;93(6):3454–3467. doi: 10.1121/1.405675. [DOI] [PubMed] [Google Scholar]
- 64.Oxenham AJ. Pitch perception and auditory stream segregation: implications for hearing loss and cochlear implants. Trends Amplif. 2008;12(4):316–331. doi: 10.1177/1084713808325881. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Brokx JP, Nooteboom SG. Intonation and the perceptual separation of simultaneous voices. J Phonetics. 1982;10:23–26. [Google Scholar]
- 66.de Cheveigne A. Concurrent vowel identification III: a neural model of harmonic interference cancellation. J Acoust Soc Am. 1997;101:2857–2865. [Google Scholar]
- 67.Alain C, Reinke K, He Y, Wang C, Lobaugh N. Hearing two things at once: neurophysiological indices of speech segregation and identification. J Cogn Neurosci. 2005;17(5):811–818. doi: 10.1162/0898929053747621. [DOI] [PubMed] [Google Scholar]
- 68.Song J, Skoe E, Banai K, Kraus N. Perception of speech in noise: neural correlates. J Cogn Neurosci. 2010;0:1–12. doi: 10.1162/jocn.2010.21556. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Anderson S, Skoe E, Chandrasekaran B, Zecker S, Kraus N. Brainstem correlates of speech-in-noise perception in children. Hear Res. 2010;270(1–2):151–157. doi: 10.1016/j.heares.2010.08.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Redford MA, Diehl RL. The relative perceptual distinctiveness of initial and final consonants in CVC syllables. J Acoust Soc Am. 1999;106(3 Pt 1):1555–1565. doi: 10.1121/1.427152. [DOI] [PubMed] [Google Scholar]
- 71.Hornickel J, Chandrasekaran B, Zecker S, Kraus N. Auditory brainstem measures predict reading and speech-in-noise perception in school-aged children. Behav Brain Res. 2011;216(2):597–605. doi: 10.1016/j.bbr.2010.08.051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Parbery-Clark A, Skoe E, Lam C, Kraus N. Musician enhancement for speech-in-noise. Ear Hear. 2009;30(6):653–661. doi: 10.1097/AUD.0b013e3181b412e9. [DOI] [PubMed] [Google Scholar]
- 73.Pichora-Fuller MK, Souza PE. Effects of aging on auditory processing of speech. Int J Audiol. 2003;42(Suppl 2):S11–S16. [PubMed] [Google Scholar]
- 74.Heinrich A, Schneider BA, Craik FI. Investigating the influence of continuous babble on auditory short-term memory performance. Q J Exp Psychol (Colchester) 2008;61(5):735–751. doi: 10.1080/17470210701402372. [DOI] [PubMed] [Google Scholar]
- 75.Woodcock RW, McGrew KS, Mather N. Woodcock-Johnson III Tests of Cognitive Abilities. Itasca, IL: Riverside Publishing; 2001. [Google Scholar]
- 76.Parbery-Clark A, Skoe E, Kraus N. Musical experience limits the degradative effects of background noise on the neural processing of sound. J Neurosci. 2009;29(45):14100–14107. doi: 10.1523/JNEUROSCI.3256-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Song JH, Skoe E, Wong PCM, Kraus N. Plasticity in the adult human auditory brainstem following short-term linguistic training. J Cogn Neurosci. 2008;20(10):1892–1902. doi: 10.1162/jocn.2008.20131. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.de Boer J, Thornton ARD. Neural correlates of perceptual learning in the auditory brainstem: efferent activity predicts and reflects improvement at a speech-in-noise discrimination task. J Neurosci. 2008;28(19):4929–4937. doi: 10.1523/JNEUROSCI.0902-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Russo NM, Nicol TG, Zecker SG, Hayes EA, Kraus N. Auditory training improves neural timing in the human brainstem. Behav Brain Res. 2005;156(1):95–103. doi: 10.1016/j.bbr.2004.05.012. [DOI] [PubMed] [Google Scholar]
- 80.Suga N, Ma X. Multiparametric corticofugal modulation and plasticity in the auditory system. Nat Rev Neurosci. 2003;4(10):783–794. doi: 10.1038/nrn1222. [DOI] [PubMed] [Google Scholar]
- 81.Kumar AU, Hegde M, Mayaleela Perceptual learning of non-native speech contrast and functioning of the olivocochlear bundle. Int J Audiol. 2010;49(7):488–496. doi: 10.3109/14992021003645894. [DOI] [PubMed] [Google Scholar]
- 82.Carcagno S, Plack CJ. Subcortical plasticity following perceptual learning in a pitch discrimination task. J Assoc Res Otolaryngol. 2011;12(1):89–100. doi: 10.1007/s10162-010-0236-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Russo NM, Hornickel J, Nicol T, Zecker S, Kraus N. Biological changes in auditory function following training in children with autism spectrum disorders. Behav Brain Funct. 2010;6:60. doi: 10.1186/1744-9081-6-60. [DOI] [PMC free article] [PubMed] [Google Scholar]