Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2011 Sep 16;108(39):16139–16140. doi: 10.1073/pnas.1112841108

Hearing impairments hidden in normal listeners

Shihab A Shamma 1,1
PMCID: PMC3182733  PMID: 21926326

Early and accurate diagnosis of sensory deficits is a public health concern with major scientific and financial implications. Several fundamental requirements conspire to make this goal a difficult one to achieve; among them is the necessity for well-designed quick tests that can pinpoint the nature of the deficits. Another is the need for a deep understanding of the functional organization of the sensory system and the cues mediating perception.

To appreciate how these requirements shape our current diagnostic paradigms, consider the striking difference between the procedures routinely used to screen children and young adults for visual and auditory impairments. In vision, even the simplest eye examinations test the ability to identify the shapes of letters or the direction of arrows, and hence the acuity of the visual system in encoding the meaningful structure of the visual scene. By contrast, hearing tests are often limited to measuring audiometric thresholds that reveal the just-audible tone intensity as a function of frequency. The relevance of these hearing tests is analogous in vision to asking subjects to indicate the presence or absence of a spot of light falling on different retinal locations. It is therefore understandable why the utility of basic auditory screening procedures is limited to indications of “hard of hearing,” failing in most cases to discover a heterogeneity of hearing disabilities in a population of listeners or to provide an indication of hearing impairments that can benefit from early intervention.

Auditory scientists are of course well aware of these limitations and have devised a variety of alternative tests to probe deeper into auditory function. These efforts, however, are hampered by the lack of an agreed-upon set of basic auditory physiological and perceptual cues whose measurements reliably reflect hearing ability and fidelity. Thus, despite enormous progress, we remain largely uncertain about the neurobiological mechanisms underlying perception of basic attributes of sound, such as pitch (1, 2), localization (3, 4), and timbre (5, 6). One important source of this difficulty stems from a fundamental duality of spectral and temporal cues in early auditory processing that has bedeviled auditory research for more than a century (7). The essence of this controversy is illustrated in Fig. 1. It stems from the duality of cochlear representations (Fig. 1A) and the way these cues become intermingled in the midbrain as they contribute jointly in mediating various auditory perceptual attributes (Fig. 1B).

Fig. 1.

Fig. 1.

Auditory functions and cues. (A) Spectral and temporal cues in cochlear processing. Cochlear analysis of three harmonic tones (100, 200, and 300 Hz) results in spatially well-separated clusters of responses along the tonotopic axis of the auditory nerve (spectral analysis). Each tone also evokes a periodic response pattern time-locked to the frequency of the tone and referred to as the temporal fine structure (TFS). The perceived spectral pattern can be estimated from either the spectral cues (by simple averaging of each channel) and/or from the temporal cues by analysis of the response periodicities. (B) Auditory functions and diagnostics. A conceptual sequence of auditory transformations and the common procedures (in blue) to probe the cues at each stage. Audibility threshold measurements often reflect general cochlear function. More specific insights into its spectral analysis are gained from “bandwidth” estimates (16). Fidelity of the TFS is assessed by detecting thresholds of frequency modulation (FM) (10). Spectral and temporal cues become intimately intermingled as they jointly contribute to generate the perceptual attributes (timbre, pitch, and location) in the midbrain and toward the cortex. Pitch percepts are often evaluated using “melodic judgments and resolution estimates (5). Timbre is a more ambiguous percept that is broadly reflected by measurements of speech intelligibility and musical instrument identification (9). Localization of azimuthal sources near the midline is thought to reflect well the fidelity of the interaural time delays (ITDs) derived from the TFS (17). Finally, auditory cortical function engages numerous cognitive abilities, such as attention and memory, as well as bottom-up spectral and temporal cues through the perceptual attributes. “Speech segregation tests” can be designed to diagnose the contribution of each specific subset of these cues to perception, as demonstrated by Ruggles et al. (8).

The study by Ruggles et al. in PNAS (8) elegantly demonstrates how much can be gained from a careful consideration of auditory principles in diagnosing hearing impairments, both at a scientific level and as a public health goal. Having selected a group of “normal” subjects on the basis of widely used audiometric threshold criteria, the authors design a series of tests with suprathreshold sounds targeted to detect evidence of “temporal dysfunction,” specifically the inability to encode properly the temporal cues thought to be critical for performing well on these tests.

The first test is a speech segregation task, typical of scenarios commonly referred to as the “cocktail party problem” (9), in which the subject has to attend to one of several simultaneous speakers. To do so successfully, the listener has to accurately perceive the location of a target speaker and comprehend her speech while resisting interference from other nearby speakers. This complex process requires not only the encoding of precise temporal information to localize the desired source but also a myriad of other spectral and temporal cues to decode the speech. The authors astutely observe that badly performing subjects comprehended well the received speech, though not that of the intended target speaker. One sensible interpretation of this result is that subjects could not localize well, an auditory function that is strongly mediated by interaural time delay cues that are neurally encoded by the fine-time-structure responses on the auditory nerve [the TFS (temporal fine structure) cues in Fig. 1A].

The second test explores the integrity of these temporal cues more directly by measuring detection thresholds of frequency modulations on a single tone (Fig. 1B), a task that is widely assumed to require precise temporal cues but is entirely unrelated to binaural hearing (10). Again, badly performing subjects on the first test displayed a similar difficulty in this task, suggesting that their deficits had more to do specifically with temporal encoding mechanisms and less to do with the binaural processing of these cues.

Both of the above tests are psychoacoustic, requiring careful controls, extensive measurements, and willing subjects who follow instructions. By contrast, the third test the authors administer is a noninvasive physiological measurement of the auditory brainstem response, the summed neural activity from the cochlear

Much can be gained from a careful consideration of auditory principles in diagnosing hearing impairments.

nucleus and nearby midbrain structures midway up the auditory pathway. Voiced speech and other stimuli containing periodic components evoke a frequency-following response (FFR) that reflects stimulus-locked neural responses, which may be interpreted as an approximate, and admittedly ill-defined, global correlate of the accuracy of temporal encoding (11). Be that as it may, the FFR of the subjects that performed well on the previous two psychoacoustic tests displayed strong responses to the fundamental periodicities in the speech signal. Interestingly, this finding is consistent with many previous reports that have associated enhanced FFR with learned auditory skills, such as in musicians and speakers of tonal languages (12), and weaker brainstem responses (ABR) after exposure to loud noise (13). ABR measurements do not involve a psychoacoustic task requiring comprehension of instructions or speech samples. As such, the test decouples higher cognitive abilities and functions (e.g., deficits in attention or linguistic abilities) from the expression of the basic temporal cues at the earlier auditory stages, thus pointing to the possible origin of the observed impairments. Finally, an attractive aspect of the ABR test is its suitability for infants and other noncooperative subjects, because it requires no feedback or even conscious listening (14).

The implications of the Ruggles et al. study (8) are numerous and important. On the scientific level, this study confirms the critical role of the TFS in sound perception and analysis of complex auditory scenes (15). On the clinical and therapeutic levels, it demonstrates the feasibility and benefits of more sophisticated auditory screening tests. It also points to the urgent need for a larger battery of tests to probe a wider set of cues at different levels of the auditory system (10). Such information will pave the way for the design of more effective hearing aids that are tailored to ameliorate the exact diagnosed impairments. For instance, one may incorporate finer spectral analysis to counter abnormally broad cochlear filters or a crisper representation of onsets and offsets to enhance the temporal modulations. Finally, as the authors point out, the societal implications of better auditory screening are substantial because it makes apparent the dangers of pervasive habits such as listening to blaring music and living with loud environmental sounds, all the while hiding behind the false security of normal audiometric thresholds!

Footnotes

The author declares no conflict of interest.

See companion article on page 15516 of issue 37 in volume 108.

References

  • 1.Oxenham AJ, Bernstein JG, Penagos H. Correct tonotopic representation is necessary for complex pitch perception. Proc Natl Acad Sci USA. 2004;101:1421–1425. doi: 10.1073/pnas.0306958101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Shamma S, Klein D. The case of the missing pitch templates: How harmonic templates emerge in the early auditory system. J Acoust Soc Am. 2000;107:2631–2644. doi: 10.1121/1.428649. [DOI] [PubMed] [Google Scholar]
  • 3.McAlpine D, Jiang D, Palmer A. A neural code for low-frequency sound localization in mammals. Nat Neurosci. 2001;4:396–401. doi: 10.1038/86049. [DOI] [PubMed] [Google Scholar]
  • 4.Popper AN, Fay RR, editors. Sound Source Localization. New York: Springer; 2005. [Google Scholar]
  • 5.Moore BCJ. An Introduction to the Psychology of Hearing. London: Elsevier Academic Press; 2004. [Google Scholar]
  • 6.Shamma S. Physiological bases of timbre perception. In: Michael Gazzaniga., editor. The New Cognitive Neurosciences. Cambridge, MA: MIT Press; 1999. [Google Scholar]
  • 7.Pickles JO. An Introduction to the Physiology of Hearing. Bingley, UK: Emerald Group; 2003. [Google Scholar]
  • 8.Ruggles D, Bharadwaj H, Shinn-Cunningham BG. Normal hearing is not enough to guarantee robust encoding ofsuprathreshold features important in everyday communication. Proc Natl Acad Sci USA. 2011;108:15516–15521. doi: 10.1073/pnas.1108912108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bronkhorst AW. The cocktail party phenomenon: A review on speech intelligibility in multiple-talker conditions. Acust Act Acust. 2000;86:117–128. [Google Scholar]
  • 10.Strelcyk O, Dau T. Relations between frequency selectivity, temporal fine-structure processing, and speech reception in impaired hearing. J Acoust Soc Am. 2009;125:3328–3345. doi: 10.1121/1.3097469. [DOI] [PubMed] [Google Scholar]
  • 11.Smith JC, Marsh JT, Brown WS. Far-field recorded frequency-following responses: Evidence for the locus of brainstem sources. Electroencephalogr Clin Neurophysiol. 1975;39:465–472. doi: 10.1016/0013-4694(75)90047-4. [DOI] [PubMed] [Google Scholar]
  • 12.Wong PC, Skoe E, Russo NM, Dees T, Kraus N. Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nat Neurosci. 2007;10:420–422. doi: 10.1038/nn1872. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Kujawa SG, Liberman MC. Adding insult to injury: Cochlear nerve degeneration after “temporary” noise-induced hearing loss. J Neurosci. 2009;29:14077–14085. doi: 10.1523/JNEUROSCI.2845-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Russo NM, et al. Deficient brainstem encoding of pitch in children with Autism Spectrum Disorders. Clin Neurophysiol. 2008;119:1720–1731. doi: 10.1016/j.clinph.2008.01.108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Lorenzi C, Moore BCJ. Role of temporal envelope and fine structure cues in speech perception: A review. In: Dau T, Buchholz JM, Harte JM, Christiansen TU, editors. Auditory Signal Processing in Hearing-Impaired Listeners, First International Symposium on Auditory and Audiological Research (ISAAR 2007) Holbaek, Denmark: Center-Tryk; 2008. [Google Scholar]
  • 16.Glasberg BR, Moore BCJ. Derivation of auditory filter shapes from notched-noise data. Hear Res. 1990;47:103–138. doi: 10.1016/0378-5955(90)90170-t. [DOI] [PubMed] [Google Scholar]
  • 17.Joris PX, Smith PH, Yin TC. Coincidence detection in the auditory system: 50 years after Jeffress. Neuron. 1998;21:1235–1238. doi: 10.1016/s0896-6273(00)80643-1. [DOI] [PubMed] [Google Scholar]

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES