Abstract
Microelectrode recordings from the human auditory cortex suggest that the tuning of individual neurons can account for sound frequency discrimination thresholds and that this tuning varies in a context-dependent fashion with the type of sound used to measure it.
Our ears can easily distinguish thousands of objects and events: human voices, musical instruments, approaching vehicles, shattering glass, and so on. What all auditory objects have in common is that they contain vibrating parts — for example, strings, membranes, vocal folds, engine valves, buzzing insect wings — which differ greatly in their mechanical properties, so that each emits its own ‘fingerprint’ of sound frequencies. The capacity of the auditory system to resolve the constituent frequencies of each ‘fingerprint’ is therefore a critical step in the identification of auditory objects. A recent paper by Bitterman et al. [1] provides new insights into how this may be done, by illustrating the remarkable selectivity and adaptability of the neural filters in the human central auditory system.
Non-invasive techniques for measuring neural activity — or, in the case of functional imaging, the associated hemodynamic changes — have revolutionized our ability to study the human brain in action. But these are relatively blunt tools that sample the activity of large assemblies of neurons within whole brain regions, rather than at the level of single neurons. Most of our knowledge about information processing by individual neurons has come from electrophysiological recording experiments carried out in animals. Consequently, we are still a long way from identifying the neurophysiological underpinnings of human cognition.
A rare opportunity to carry out comparable electrophysiological studies in humans is, however, provided by epileptic patients in whom intracranial microwire electrodes are implanted to determine where their seizures start. Recordings made from such patients have provided valuable insights into the functional organization of the human auditory cortex [1–4], which is located within the temporal lobe, as well as the relationship between the spiking activity of cortical neurons and functional magnetic resonance imaging measurements made at the same time [5]. In the most recent of these studies, Bitterman et al. [1] used this approach to explore the frequency tuning of single neurons in the human auditory cortex.
Frequency Selectivity in the Auditory System
The most characteristic and well documented feature of auditory neurons in the brain is their frequency selectivity. At every level of the auditory system, from the cochlea of the inner ear to the primary auditory cortex, neurons typically respond to tones over a restricted range of frequencies at the lowest sound intensities to which they are responsive, but to a progressively greater range of frequencies as the sound intensity increases. This frequency tuning is derived from a systematic variation in the biomechanical properties of the basilar membrane, which subdivides the cochlea longitudinally. Consequently, different tone frequencies excite different points along the length of the cochlea, setting up the ‘tonotopic’ maps that are preserved at subsequent levels of the auditory pathway as a result of the spatial order in the neural connections that exist between them.
Frequency selectivity can also be estimated behaviorally by measuring the threshold for detecting a tone of a particular frequency in the presence of masking noise containing energy over different frequency ranges. This shows that the auditory system contains a series of ‘bandpass filters’ or ‘critical bands’, within which different sounds are not easily distinguished and interact in ways that influence how they are perceived [6]. These psychophysical filters have their origin in the frequency analysis performed by the cochlea, but also seem to be influenced by processing that takes place within the brain [7,8].
But the auditory filters appear to be too coarse to account for our perceptual ability to detect a change in the frequency of a sound. Trained human listeners can discriminate tones that differ in frequency by less than 1%, whereas the bandwidth of the filters has been estimated to be much broader at about one fifth of an octave, closer to 15% of the center frequency. Consequently, it is thought that our ability to resolve different tones might be based — at least for the low frequencies where discrimination thresholds are best — more on the neural coding of temporal cues to sound frequency that are present in the stimulus waveform, than a change in the excitation pattern within the cochlea [6,9].
But the frequency selectivity of neurons higher up in the auditory pathway is not just a reflection of the mechanical tuning of the cochlear filters. Microelectrode recordings from neurons in the auditory cortex of different species have revealed a range of tuning curve shapes and bandwidths. In some neurons, frequency selectivity can be preserved over the full range of sound intensities to which the neuron is responsive, most likely as a result of inhibitory processing at higher levels of the auditory pathway [10]. Such sharp frequency tuning contributes, for example, to the ability of certain echolocating bats to judge relative target velocity through a comparison of the constant-frequency components of the animals’ ultrasonic pulses and their Doppler-shifted echoes [11]. The functional significance of the diverse frequency selectivity exhibited by cortical neurons in other, less specialized species is, however, unclear.
Recording from the Human Auditory Cortex
The patients studied by Bitterman et al. [1] were asked to listen to two different classes of acoustic stimuli, either an artificial stimulus known as a random chord sequence or to a movie sound track consisting of spoken dialogue and background music. The electrical activity of individual neurons in the auditory cortex was recorded through the microelectrodes, and a method known as ‘reverse correlation’ was then used to assess the frequency selectivity of the neurons. Reverse correlation assumes, plausibly, that the discharges of the recorded sensory neurons are likely to be triggered by the stimulus episodes that immediately preceded them. By observing a large number of discharges, it is then possible to infer the optimal stimulus for the neuron by asking, using ‘spike triggered averaging’, what all the stimulus episodes that precede these responses have in common [12].
When using this technique, care must be exercised to take the ‘autocorrelation structure’ of the stimuli into account. Imagine that a neuron is sensitive only to frequency A, but if, in the stimulus set used to characterize the neuron, frequency A regularly occurs together with frequency B, then frequency B may feature prominently in the spike-triggered average, even though the neuron is not directly sensitive to that frequency. Bitterman et al. [1] were therefore careful to correct their measurements for stimulus auto-correlation, using methods pioneered by Frederic Theunissen [13].
Representative examples of the sort of neural frequency responses — commonly referred to as ‘spectro-temporal receptive fields’ or STRFs — which Bitterman et al. [1] obtained are shown in Figure 1, next to spectrograms of the stimuli used to measure them. Several aspects of the STRFs shown in Figure 1 are worth noting. For instance, the STRF obtained with the random chord stimulus reveals surprisingly sharp frequency tuning (Figure 1B). The excitatory frequency regions, shown in red, are no more than a tenth of an octave wide, considerably sharper than the frequency tuning afforded by the inner ear.
Figure 1. Probing the frequency selectivity of neurons in the auditory cortex.
(A) Example of a ‘random chord’ sound stimulus. Random chords consist of numerous brief tone pips presented at random time intervals. The sound spectrogram is shown beneath the stimulus waveform. ‘Warm’ (yellow and red) colors in the spectrogram indicate high sound intensities at the times and frequencies shown on the x- and y-axes, respectively. (B) The spec-tro-temporal receptive field (STRF) of a single neuron in the human auditory cortex, estimated by reverse correlation with the random chord stimuli. In the STRF, warm colors show which frequencies excite the neuron and over what time course. (C) Example of a ‘sound track’ stimulus, consisting of a human speech dialogue against faint background music. (D) STRF of the same neuron shown in B, but this time recorded with the sound-track stimulus. Note that the neuron’s frequency tuning adapts to the stimulus and becomes broader (based on data in Bitterman et al. [1]).
The majority of the neurons recorded in this study exhibited similarly sharp frequency tuning. Bitterman et al. [1] used these data to estimate neuronal discrimination thresholds and found that these were comparable to those measured psychophysically in naïve listeners. Their findings therefore suggest that auditory cortical neurons in humans are consistently more selective for sound frequency than in most other mammalian species, although given that frequency-discrimination thresholds measured in animals are generally higher than those in humans [14], this is perhaps not surprising.
Context-Dependent Frequency Selectivity?
Although the cortical neurons tended to be most sensitive to the same frequencies when stimulated with random chord sequences and natural sounds, the STRFs obtained with the ‘sound track’ acoustic stimulus were more broadly tuned. This can be seen by comparing the STRFs in Figure 1B and D, which are from the same neuron but recorded with different stimuli. Just inspecting the spectrograms of the stimuli and their corresponding STRFs, one gets the distinct impression that the STRFs are adapted to match the type of stimulus that the neuron is listening to.
In order to understand what someone is saying, the auditory system must extract both the rapid fluctuations in the sound envelope (note how the amplitude of the sound track stimulus shown in Figure 1C alternates between silent periods and sudden bursts of sound) and the ‘speech formants’ — the broad resonant peaks in the spectrum of the sound waves. Very fine frequency discrimination is not essential for speech comprehension. Broad frequency selectivity coupled with a good sensitivity to amplitude modulations is sufficient, and so the broadly-tuned STRF shown in Figure 1D appears to be well adapted to the processing of speech stimuli. In contrast, the random chord stimulus shown in Figure 1A contains neither informative envelope modulations nor interesting formants, and the neuron appears to re-tune its STRF to extract the only interesting feature of this particular stimulus, namely the countless tone pips occurring randomly across independent, closely spaced frequency bands.
These results suggest that the neural filters used by our auditory system to analyze the sounds around us can adapt rapidly to changing task demands and can become surprisingly sharp if necessary. Task-dependent plasticity has also been observed by measuring STRFs from neurons in the auditory cortex of ferrets, so that the same tone stimulus can evoke quite different responses according to the context in which it is presented [15]. These findings highlight the importance of higher-level processing — even for spectral tasks that were once thought to have their basis in the auditory periphery — and of considering both the types of sound stimuli that are used, and the situation in which they are presented, when exploring the nature of this processing.
References
- 1.Bitterman Y, Mukamel R, Malach R, Fried I, Nelken I. Ultra-fine frequency tuning revealed in single neurons of human auditory cortex. Nature. 2008;451:197–201. doi: 10.1038/nature06476. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Howard MA, 3rd, Volkov IO, Abbas PJ, Damasio H, Ollendieck MC, Granner MA. A chronic microelectrode investigation of the tonotopic organization of human auditory cortex. Brain Res. 1996;724:260–264. doi: 10.1016/0006-8993(96)00315-0. [DOI] [PubMed] [Google Scholar]
- 3.Bidet-Caulet A, Fischer C, Besle J, Aguera PE, Giard MH, Bertrand O. Effects of selective attention on the electrophysiological representation of concurrent sounds in the human auditory cortex. J Neurosci. 2007;27:9252–9261. doi: 10.1523/JNEUROSCI.1402-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Brugge JF, Volkov IO, Oya H, Kawasaki H, Reale RA, Fenoy A, Steinschneider M, Howard MA., 3rd Functional localization of auditory cortical fields of human: Click-train stimulation. Hear Res. 2008;238:12–24. doi: 10.1016/j.heares.2007.11.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Nir Y, Fisch L, Mukamel R, Gelbard-Sagiv H, Arieli A, Fried I, Malach R. Coupling between neuronal firing rate, gamma LFP, and BOLD fMRI is related to interneuronal correlations. Curr Biol. 2007;17:1275–1285. doi: 10.1016/j.cub.2007.06.066. [DOI] [PubMed] [Google Scholar]
- 6.Moore BCJ. Frequency analysis and pitch perception. In: Yost WA, Popper AN, Fay RR, editors. In Human Psychophysics. New York: Springer-Verlag; 1993. pp. 56–115. [Google Scholar]
- 7.Evans EF, Pratt SR, Spenner H, Copper NP. Comparisons of physiological and behavioural properties: auditory frequency selectivity. In: Cazals Y, Horner K, Demany L, editors. In Advances in the Biosciences. Vol. 83. Oxford: Pergamon Press; 1992. pp. 159–169. [Google Scholar]
- 8.Schreiner CE, Langner G. Laminar fine structure of frequency organization in auditory midbrain. Nature. 1997;388:383–386. doi: 10.1038/41106. [DOI] [PubMed] [Google Scholar]
- 9.Wakefield GH, Nelson DA. Extension of a temporal model of frequency discrimination: intensity effects in normal and hearing-impaired listeners. J Acoust Soc Am. 1985;77:613–619. doi: 10.1121/1.391879. [DOI] [PubMed] [Google Scholar]
- 10.Suga N. Tribute to Yasuji Katsuki’s major findings: sharpening of frequency tuning in the central auditory system. Acta Otolaryngol Suppl. 1997;532:9–12. doi: 10.3109/00016489709126138. [DOI] [PubMed] [Google Scholar]
- 11.Riquimaroux H, Gaioni SJ, Suga N. Inactivation of the DSCF area of the auditory cortex with muscimol disrupts frequency discrimination in the mustached bat. J Neurophysiol. 1992;68:1613–1623. doi: 10.1152/jn.1992.68.5.1613. [DOI] [PubMed] [Google Scholar]
- 12.King AJ, Schnupp JWH. Sensory neuroscience: visualizing the auditory cortex. Curr Biol. 1998;8:R784–R787. doi: 10.1016/s0960-9822(07)00499-x. [DOI] [PubMed] [Google Scholar]
- 13.Theunissen FE, David SV, Singh NC, Hsu A, Vinje WE, Gallant JL. Estimating spatio-temporal receptive fields of auditory and visual neurons from their responses to natural stimuli. Network. 2001;12:289–316. [PubMed] [Google Scholar]
- 14.Fay RR. Hearing in Vertebrates: a Psychophysics Databook. Winnetka, IL: Hill-Fay Associates; 1988. [Google Scholar]
- 15.Fritz JB, Elhilali M, Shamma SA. Differential dynamic plasticity of A1 receptive fields during multiple spectral tasks. J Neurosci. 2005;25:7623–7635. doi: 10.1523/JNEUROSCI.1318-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]