Abstract
Neuronal mechanisms of auditory distance perception are poorly understood, largely because contributions of intensity and distance processing are difficult to differentiate. Typically, the received intensity increases when sound sources approach us. However, we can also distinguish between soft-but-nearby and loud-but-distant sounds, indicating that distance processing can also be based on intensity-independent cues. Here, we combined behavioral experiments, fMRI measurements, and computational analyses to identify the neural representation of distance independent of intensity. In a virtual reverberant environment, we simulated sound sources at varying distances (15–100 cm) along the right-side interaural axis. Our acoustic analysis suggested that, of the individual intensity-independent depth cues available for these stimuli, direct-to-reverberant ratio (D/R) is more reliable and robust than interaural level difference (ILD). However, on the basis of our behavioral results, subjects’ discrimination performance was more consistent with complex intensity-independent distance representations, combining both available cues, than with representations on the basis of either D/R or ILD individually. fMRI activations to sounds varying in distance (containing all cues, including intensity), compared with activations to sounds varying in intensity only, were significantly increased in the planum temporale and posterior superior temporal gyrus contralateral to the direction of stimulation. This fMRI result suggests that neurons in posterior nonprimary auditory cortices, in or near the areas processing other auditory spatial features, are sensitive to intensity-independent sound properties relevant for auditory distance perception.
Keywords: computational modeling, psychophysics, spatial hearing, what and where pathways, neuronal adaptation
Determining the distance of objects is of key value in many everyday situations. For objects that fall outside the field of vision, hearing is the only sense that provides such information. For example, consider a person reaching for a ringing phone (1–3) or a listener using distance differences to help focus on one talker in a chattering crowd (4, 5). However, whereas cortical representations of features such as spectral or amplitude modulations have been intensively examined (6, 7), the neural mechanisms of auditory distance perception are poorly understood.
In comparison with the detailed mapping of human visual cortex, knowledge of the subsystems of human auditory cortex is less complete. Only relatively broad anatomical divisions, such as that between the anterior “what” vs. posterior “where” (8–13) pathways, have been shown in human neuroimaging studies. The posterior auditory “where” pathway, which encompasses nonprimary auditory cortex areas including the planum temporale (PT) and posterior superior temporal gyrus (STG), is strongly activated by horizontal sound direction changes (12, 14–16) and movement (17, 18). However, although human (19, 20) (and nonhuman primate, refs. 21, 22) auditory systems have been shown to have neurons preferring sound sources approaching the listener (i.e., “looming”), populations responsible for auditory distance processing have not been previously identified. In fact, auditory distance perception is also relatively poorly understood from the psychophysical and computational points of view (1).
In many situations, the dominant acoustic cue for distance is the overall received stimulus intensity. However, the overall received stimulus intensity is an ambiguous distance cue (e.g., if the emitted stimulus intensity varies independently of the source distance), which can complicate studies on intensity vs. distance processing (23). Acoustically, for sources off the midline, it is impossible to fix the received stimulus intensity at both ears as distance varies, because the rate of change in intensity with distance is different at the two ears. Physiologically, when the stimulus intensity varies, detectors sensitive to a range of stimulus features can get activated. In such cases, neurons tend to respond in a nonspecific way for loud stimuli, even if they are sensitive to a specific feature at low stimulus intensities. These nonspecific responses are most commonly observed when the sounds are broadband noise (24, 25), as typically used in studies of audiospatial perception. However, recent results show that robust intensity-independent distance perception is possible for nearby sources (up to 100 cm from the listener) in simulated reverberant environments. For such sources, two major stimulus-independent and intensity-independent distance cues exist, the interaural level difference (ILD) (26, 27) and direct-to-reverberant energy ratio (D/R) (28–30). Although it is currently not fully known which of the two cues the listeners use and/or how they combine them (31), the cues' availability gives an opportunity to study intensity-independent distance representations using human neuroimaging.
In humans, tuning properties of neuronal populations can be noninvasively studied by examining neuronal adaptation, i.e., suppression of responses to a given stimulus as a function of its similarity and temporal proximity to preceding stimuli (12, 32–34). Adaptation studies can use invasive neurophysiological measurements (33) or noninvasive methods such as electroencephalography (EEG) and magnetoencephalography (MEG) (12, 32, 34), or fMRI (35). Adaptation fMRI compares signals elicited by two sound-sequence conditions, one consisting of identical (i.e., constant) stimuli and the other consisting of sounds varying along the feature dimension of interest. Adaptation fMRI presumably differentiates the tuning properties of neurons within each voxel. Specifically, in a voxel populated by neurons tuned loosely to the feature dimension of interest, such as sound direction (16), fMRI signals get equally adapted during constant and varying stimulation. In contrast, in a voxel with neurons tuned sharply to the feature dimension of interest, a release of adaptation and increased fMRI signal is observed during varied vs. constant stimulus blocks.
Here, we examined auditory distance perception and its neural correlates. Nearby sources located at various distances were simulated along the interaural axis (to the right of the listener) in a reverberant environment (Fig. 1A). Two different adaptation fMRI experiments were conducted to localize brain areas sensitive to intensity-independent auditory distance cues. A behavioral experiment was performed to validate the virtual acoustics used during fMRI, and to examine the perceptual cues and psychophysical principles determining listeners’ sensitivity to changes in distance.
Results
Behavioral Experiment.
We used virtual acoustics to measure perceptual sensitivity to intensity-independent stimulus distance cues (Fig. 1A). The main goal was to confirm that the listeners could use the intensity-independent cues to judge source distance in our simulated auditory environment. Subjects were instructed to indicate whether the second of two noise bursts, simulated from different distances, originated closer or farther than the first burst. These discriminations had to be made irrespectively of perceived sound intensity, which varied randomly for each burst to eliminate its contribution as a distance cue (Fig. 1B). Overall, subjects accurately judged sound source distances despite the confounding information provided by intensity cues. As expected, the listeners’ accuracy improved with increasing distance difference between the two simulated sound sources (Fig. 2 and for further details, SI Results and Fig. S1).
In addition to confirming that our virtual auditory environment was robust, these results demonstrate that distance discrimination sensitivity is largely independent of the baseline distance of the simulated sound sources, as long as the ratio of compared distances is fixed. Specifically, in Fig. 2, bars are grouped such that each group shows the performance for a fixed number of unit log-distance intervals (i.e., the ratio of source distances). Within each group, bars are ordered from the nearest to the farthest pair. No systematic upward or downward trend is visible in performance across simulated source-distance pairs within each group. To confirm this lack of trend, a linear approximation was made to each subject’s performance as a function of the distance pair, separately for each group and each subject. Across subjects, none of the slopes was significantly different from zero (two-tailed Student t test with Bonferroni correction performed on the fitted slopes separately for each source-pair group; P > 0.1), consistent with the predictions of Weber’s law (for more details, see the two sections, Acoustic Analysis of Stimuli and Predictions of Distance Discrimination in SI Results, and Fig. S2). Finally, our acoustical analyses described in SI Results and Fig. S2 show that, of the individual intensity-independent distance cues, D/R is more robust and reliable than ILD.
fMRI Experiment.
To localize intensity-independent distance representations, we compared auditory cortex areas activated during distance vs. intensity changes vs. constant repetitive sound stimulation (Fig. 1C). To control for fluctuations in attention and alertness, subjects were asked to detect occasional changes in sound duration that occurred independently of distance or intensity (across-subject mean hit rate = 93%; mean reaction time = 650 ms, data of one subject failing to conduct this task were excluded from the analyses; see SI Results).
As expected, constant auditory stimulation significantly activated primary (the medial 2/3 of Heschl’s gyrus, HG) and nonprimary (STG, PT, and planum polare, PP) auditory cortex areas (Fig. 3A). Introducing changes to auditory stimulation increased auditory cortex activations, as shown by the contrasts between varying intensity or varying distance vs. constant conditions (Fig. 3B). Note that these contrasts did not reach significance at the crest of HG, consistent with the view that primary auditory cortices are adapted less prominently by high-rate stimulation than the surrounding nonprimary areas (36). In Fig. 3B, the areas activated in the varying distance contrast overlap with the areas activated in the varying intensity contrast, as expected given that the stimuli in varying distance condition also varied in intensity. Finally, Fig. 3C shows the areas that represented more prominent release from adaptation during varying distance than during varying intensity condition. This contrast suggests that the posterior nonprimary auditory cortex areas (posterior STG and PT) are specifically sensitive to intensity-independent auditory distance cues.
In an additional hypothesis-based fMRI analysis (Fig. 4), a region of interest (ROI) was defined in each hemisphere by combining two anatomical FreeSurfer standard-space labels (PT and posterior aspect of STG) that were a priori (8–12) conjectured to encompass areas activated by audiospatial features. Consistent with the whole-brain mapping results (Fig. 3), blood oxygen level dependent (BOLD) percentage signal changes were significantly stronger (F1,10 = 9.3, P = 0.01) during varying distance than varying intensity conditions in the left-hemisphere posterior auditory cortex ROI, but not in the right hemisphere. These data also show the significant increases of auditory activity during varying vs. constant stimulation in the left (F1,10 = 33.4, P < 0.001) and right (F1,10 = 83.6, P < 0.001) hemisphere ROIs.
Additionally, we examined the structures activated when the presentation intensity was normalized such that the overall energy received at the ear closer to the source was constant, independent of the source distance (SI Results and Fig. S3). This experiment revealed more widespread activations than the varying distance vs. varying intensity contrast in the main fMRI experiment (Fig. 3C), most likely because the near-ear intensity normalization did not result in exclusive activation of the distance-sensitive areas.
Finally, to enhance comparability to fMRI studies in volume space, we conducted a whole-brain analysis in a 3D standard brain, using a statistical approach similar to the surface-space analysis (Fig. 5). In the varying distance vs. varying intensity contrast, a significant activation cluster was identified. The strongest voxel was located in the left PT (Montreal Neurological Institute Talairach coordinates {x, y, z} = {−60, −35, 15}, Talairach {x, y, z} = {−59, −33, 16}, cluster volume 912 mm3) with the area extending also to the posterior STG.
Discussion
We studied perceptual and neuronal mechanisms of auditory distance processing by combining computational acoustic analyses with behavioral and adaptation fMRI measurements in a virtual auditory environment. Our results provide evidence for auditory areas specifically sensitive to acoustic distance cues that are independent of sound intensity changes. These activations centered in the posterior STG and PT, near the so-called auditory “where” pathway (10). Although a number of recent studies have associated these areas with perception of sound direction changes (12, 16) and movement (17, 37, 38), our study to our knowledge is unique in revealing neuron populations sensitive to sound-source distance variations.
Auditory-cortex neurons have a strong tendency to stimulus-specific adaptation (33) that suppresses their responsiveness to repetitive sounds. Differential release from this adaptation, as a function of increasing dissimilarity of successive stimuli, is believed to reveal units that are specifically sensitive to the varying feature dimension (12, 34). However, this approach can be confounded by the fact that stimulus intensity changes can increase activations of various feature detectors. In particular, when the baseline stimulus is broadband noise containing all possible features, stimulus intensity is known to interact with feature specificity of auditory neurons (24, 25). A supporting fMRI experiment (SI Results and Fig. S3) was performed, in which the received intensity of varying-distance stimuli was normalized at the near ear to minimize the intensity variation. This experiment showed relatively widespread and nonspecific activation patterns, thus confirming the expectation that various feature detectors can be released from adaptation when intensity varies. However, our main analyses were based on the contrast between conditions with varying distance and received intensity in the listener’s ears (equally loud sounds simulated from different distances) vs. varying only the received intensity (sounds of different intensity from one simulated distance; Fig. 1C). The significant increase in auditory-cortex activation in PT and posterior STG in this contrast presumably revealed neurons sensitive to sound-distance cues, independent of intensity.
The present study concentrated in the near-head range where robust distance cues were presumed to be available. Although larger distances could be represented differently (39), it is possible that the posterior nonprimary auditory cortex areas, near those activated during intensity-independent nearby-source distance processing, also represent distances of sounds originating from farther away. As mentioned above, these areas have been previously shown to be associated with other aspects of spatial hearing (12, 16, 17, 37, 38). For example, a study (16) using an adaptation fMRI paradigm analogous to our approach found bilateral activations to sound-direction changes in PT, near the present distance-dependent activations. It is also noteworthy that in the present study, the strongest activations to intensity-independent sound-distance changes occurred in the left PT/posterior STG, contralateral to the virtual origins of the stimuli. However, given that only one simulated direction was studied, the present data cannot offer conclusive evidence that the distance of sounds originating on one side of the head is always represented by the contralateral hemisphere or that sound-source distance (or direction) is represented in a topographical fashion analogous to other sensory modalities (note also evidence supporting an alternative two-channel model, refs. 40–42).
An important theoretical question is whether the present PT/posterior STG activations reflect neurons encoding auditory distance, per se, or whether they correspond to distinct subpopulations for D/R and/or ILD distance cues. If the latter were true, the present fMRI results could also be interpreted as a partial “byproduct” of ILD neurons that are primarily related to direction discrimination. However, basic binaural cues such as ILD are extracted and extensively processed already at the subcortical level (for a review, see ref. 42). If we understand the cerebral cortex as a self-organizing network, it is more likely that at higher processing levels, beyond the subcortical pathways and primary sensory areas, representations develop for combinations of features that frequently occur together. Previous neurophysiological (8, 11) and human neuroimaging studies (43, 44) generally suggest that nonprimary regions, including PT/posterior STG, process progressively more complex sounds than the primary “core” regions of auditory cortex. These observations are in line with a previous MEG study (45) that systematically compared activations to individual and combined 3D sound features, and showed that ILD alone is not sufficient for producing direction-specific activations at the cortical level. It is also worth noting that in humans, no consistent evidence on ILD-specific fMRI activations in PT/posterior STG exists. The few published ILD–fMRI studies have, instead, compared transient vs. sustained BOLD response differences in more primary regions (46) or showed direction-specific ILD effects only in areas beyond auditory cortices (47). Nevertheless, some MEG (48) and EEG (49) studies have reported nonspecific differences in cortical response patterns to ILD and interaural time differences (ITD). There is also evidence from animal studies suggesting ILD neurons in primary auditory cortices, for example, in the cat (50) (note, however, that the link between animal physiology and human fMRI is relatively weak in the spatial auditory domain) (51). Further, our auditory systems are highly adaptive, and different cues may govern distance perception in different acoustic environments. For example, a recent psychophysical study suggested that listeners do not use the ILD cue for distance perception in natural reverberation (31), even though it is an important distance cue in artificial anechoic spaces where other naturally occurring cues are unavailable (52). Further studies are therefore needed to investigate how individual spatial features are processed across the hierarchy of auditory pathways in humans.
Our behavioral data show that sensitivity to auditory distance changes is constant for sources separated by a constant distance ratio, following Weber’s law. This result holds for nearby distances varying along the interaural axis, as examined here, which could be relevant for general models of auditory distance perception. Specifically, both the rate of distance-dependent changes in the auditory distance cues (ILD and D/R) (SI Results, Acoustic Analysis of Stimuli, and Fig. S2) and/or the perceptual sensitivity to these cue changes have been suggested to vary as a function of sources’ absolute distance (27, 30, 53). One might therefore expect that distance sensitivity would vary with absolute distance if performance were based on either of the two cues alone. Here, we did not find any systematic trends suggesting that distance-discrimination performance depends on the absolute source distance. For example, subjects discriminated distance differences between the smallest relative intervals on average at a 67–71% accuracy, with the best nominal performance observed for the distance pair 50 vs. 75 cm from the head (Fig. 2). On the basis of the previous findings concerning discriminability of individual cues (as discussed in SI Results, Predictions of Distance Discrimination), the D/R-only–based performance would be expected to be poorest for the smallest absolute distances (in the case of the smallest interval, 15 vs. 19 cm) and best for the largest distance (75 vs. 100 cm). Because no such D/R-only (or ILD-only) dependence was observed, our results suggest that distance judgments were not based on a simple mapping of a single distance cue (in contrast to a previous interpretation) (31). However, as shown in SI Results, it may still be most parsimonious to assume that the main cue used by the listeners was the monaural near-ear D/R.
In summary, our results reveal an area in the posterior nonprimary auditory cortex that is sensitive to intensity-independent auditory distance cues.
Materials and Methods
Subjects.
The same set of subjects (n = 12, ages 20–33 y, five females) with self-reported normal hearing ability participated in all three experiments. Human subjects' approval was obtained and voluntary consents were signed before measurements.
Stimuli and Setup.
All stimuli used in this study were generated in virtual auditory space (54, 55), using binaural room impulse responses (BRIR) that included realistic room reverberation. A single set of nonindividualized BRIRs was used, measured on a listener that did not participate in this study. Unless specified otherwise, all details of the measurement procedures, including the microphone, speaker, and the BRIR measurement technique used were identical to our previous study (27).
The BRIRs were measured in a small classroom (3.4 m × 3.6 m × 2.9 m height) using the Bose FreeSpace 3 Series II surface-mount cube speaker (Bose). The room was carpeted, with hard walls and acoustic tiles covering the ceiling. The room reverberation times, T60, in octave bands centered at 500, 1,000, 2,000, and 4,000 Hz ranged from 480 to 610 ms. The BRIRs were measured by placing miniature microphones (FG-3329c; Knowles Electronics) at the blocked entrances of the listener’s ear canals. The loudspeaker was set to face the listener at various distances (15, 19, 25, 38, 50, 75, or 100 cm) from the center of the listener’s head, directly to the right of the listener along the interaural axis, at the level of the listener’s ears (Fig. 1A).
The target stimulus was a 300-ms-long sample of white noise, bandpass-filtered at 100–8,000 Hz. A set of 50 independent noise burst tokens was generated and each token was convolved with each of the BRIRs to create standard stimuli for each source distance (Fig. 1A). An identical set of deviant stimuli was also generated, differing from the standard stimuli only in that the deviant duration was 150 ms. For each experimental trial, either two (behavioral experiment, Fig. 1B) or 14 (imaging experiments, Fig. 1C) noise bursts were randomly selected, scaled depending on the normalization scheme used, and placed in series with stimulus onset asynchrony, to create the stimulus sequence. Finally, the stimuli were filtered to compensate for the headphone transfer functions.
The stimuli used in the experiments mainly differed by the overall stimulus intensity normalization used. For the behavioral experiment, each noise burst was normalized so that its overall intensity received at the right ear (which was close to the simulated source) was fixed and then randomly roved over a 12 dB range so that the monaural overall intensity distance cue was eliminated at both the left and right ears (Fig. 1B).
The imaging experiment stimulus sequences always consisted of 14 noise bursts (Fig. 1C). Each stimulus sequence in trials with varying distance contained two noise bursts for each of the seven distances, ordered pseudorandomly such that each distance was present at least once before the second occurrence of any of the distances. The varying distance stimuli were normalized such that the presentation intensity was fixed (i.e., the overall intensity distance cue was present in these stimuli because the received intensity for the near sources was higher than the received intensity for the far sources). The varying intensity stimuli were simulated from the fixed distance of 38 cm and their presentation intensity was varied such that the received intensity at the near ear varied across the same range as for the varying distance stimuli. The constant stimuli were presented at a fixed distance of 38 cm and at a fixed presentation intensity, equal to the varying distance stimuli.
The stimulus sequence files, generated at a sampling rate of 44.1 kHz, were stored on the hard disk of the control computer (IBM PC-compatible), running Presentation (Neurobehavioral Systems). On each trial, one of the sequences was selected and presented through the Fireface 800 sound processor (RME), Kramer 900XL amplifier (Kramer Electronics), and Sensimetrics S14 (Sensimetrics) MRI-compatible headphones. Average received level was 65 dB(A), measured at the near ear of a KEMAR (Manikin for Acoustic Research; Knowles Electronics) manikin equipped with the DB-100 Zwislocki Coupler and the Etymotic Research ER-11 microphones. Responses were collected via MRI-compatible five-key universal serial bus (USB) keyboard. A video projector was attached to the control computer and projected the instructions to the subject in the scanner.
Experimental Procedure and Data Acquisition.
All experiments were performed during a single 2-h long experimental session using the same experimental setup, equipment, and stimuli. After the initial preparation and practice runs, which were performed outside the scanner, the experiments were performed inside the scanner. First, the imaging experiments were run while the listener was instructed to perform a sound duration deviant detection task. Then, the behavioral experiment was performed. This ordering (behavioral experiment after imaging experiments) was chosen to prevent listeners from focusing on the stimulus distance during the imaging experiments, which could alter the listeners’ concentration level during the different imaging stimulus conditions.
Behavioral experiment.
The behavioral experiment was a distance discrimination experiment, consisting of four runs, each containing 21 randomly ordered trials (1 trial for each combination of two of the seven distances). Two different random noise burst tokens were selected for each trial, one for each distance. A trial started by the word “listen” appearing on the computer screen (Fig. 1B), followed after 200 ms by the two auditory stimuli presented with a 1,000 ms stimulus onset asynchrony. The subject responded by indicating whether the second sound source was closer or farther away than the first source by pressing one of two keys on the keyboard. The experiment was self-paced and the total duration of one trial was, on average, ∼5 s.
Imaging experiment.
The imaging experiment consisted of two runs, each containing 24 different random trials for each of the four stimulus sequence types (Fig. 1C). The order of stimulus trial types was randomized to maximize the sensitivity of the experiment (56). Each trial consisted of 2 s of silence, followed by 7 s of auditory stimulus presentation, 1 s of silence, and 2 s of fMRI image acquisition (Fig. 1C). On 50% of trials, one randomly chosen burst in the stimulus sequence was replaced by a shorter, 150-ms deviant burst. Subjects were instructed to press a key on the keyboard whenever this deviant was detected (i.e., the listeners were instructed to focus on the stimulus duration, even though they were aware that the stimulus distance varied).
Whole-head fMRI was acquired at 3T using a 32-channel coil (Siemens;TimTrio). To circumvent response contamination by scanner noise, we used a sparse-sampling gradient-echo BOLD sequence (TR/TE = 12,000/30 ms, 9.82-s silent period between acquisitions, flip angle = 90°, FOV 192 mm) with 36 axial slices aligned along the anterior–posterior commissure line (3-mm slices, 0.75-mm gap, 3 × 3 mm2 in-plane resolution), with the coolant pump switched off. T1-weighted anatomical images were obtained for combining anatomical and functional data using a multiecho MPRAGE pulse sequence (TR = 2,510 ms; 4 echoes with TEs = 1.64 ms, 3.5 ms, 5.36 ms, 7.22 ms; 176 sagittal slices with 1 × 1 × 1 mm3 voxels, 256 × 256 mm2 matrix; flip angle = 7°).
Data Analysis.
Behavioral data.
In the behavioral experiment, the percentage of correct responses was analyzed for each distance pair, repeat, and subject. Across-subject means and SEMs were computed. To obtain an overall measure of individual subject performance, a simple decision theory model was proposed and fitted to the percentage of correct data (SI Results).
In the imaging experiment, the subjects’ task was to detect a deviant sound in the stimulus sequence. The response was accepted as a correct deviant detection if it occurred within 2.5 s after the deviant onset. Hit rates and reaction times for the correct detections were analyzed (SI Results). Repeated-measures ANOVA was used to test whether these measures varied significantly for different types of stimulus.
fMRI Data.
Cortical surface reconstructions and standard-space coregistrations of the individual anatomical data, as well as functional data analyses, were conducted using Freesurfer 5.0. Individual functional volumes were motion corrected, coregistered with each subject’s structural MRI, intensity normalized, resampled into standard cortical surface space (57, 58), smoothed using a 2-dimensional Gaussian kernel with an FWHM of 5 mm, and entered into a general-linear model (GLM) with the task conditions as explanatory variables. A random-effects GLM was then conducted at the group level. To control for multiple comparisons, a cluster analysis (Monte Carlo simulations with 10,000 iterations; P < 0.01) was used, with specific contrasts between the conditions (varying intensity, varying distance, and constant stimulation) constrained to areas where the main effect (auditory stimulation vs. baseline) was significant (corrected P < 0.01). In addition, to enhance comparability of our main results, data were also analyzed in a 3D standard volume. The statistical procedures were identical to the main analysis.
In the hypothesis-based fMRI analysis, a ROI was defined in each hemisphere by combining two anatomical FreeSurfer standard-space labels (PT and posterior aspect of STG) that were a priori (8–12) conjectured to encompass areas activated by audiospatial features. These ROIs were then resampled through the spherical standard space onto each individual subject’s brain representations. Within each subject, within-ROI voxels showing significant activations (P < 0.01) were used to determine percentage of signal changes vs. the baseline rest condition during different task conditions. Results were entered into a repeated-measures ANOVA with a priori difference contrasts.
Supplementary Material
Acknowledgments
This work was supported by National Institutes of Health (NIH) Awards R21DC010060, R01MH083744, R01HD040712, and R01NS037462, and NIH/National Center for Research Resources (NCRR) Award P41RR14075. N.K. was also supported by the European Community's 7FP/2007-13 Grant PIRSES-GA-2009-247543 and by Slovak Scientific Grant Agency Grant VEGA 1/0492/12. The research environment was supported by NCRR Shared Instrumentation Grants S10RR023401, S10RR019307, and S10RR023043. Acoustic measurements performed at the Boston University Hearing Research Center used resources supported by NIH Award P30DC004663.
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1119496109/-/DCSupplemental.
References
- 1.Zahorik P, Brungart DS, Bronkhorst AW. Auditory distance perception in humans: A summary of past and present research. Acta Acust. United Ac. 2005;91:409–420. [Google Scholar]
- 2.Maier JX, Neuhoff JG, Logothetis NK, Ghazanfar AA. Multisensory integration of looming signals by rhesus monkeys. Neuron. 2004;43:177–181. doi: 10.1016/j.neuron.2004.06.027. [DOI] [PubMed] [Google Scholar]
- 3.Neuhoff JG. Perceptual bias for rising tones. Nature. 1998;395:123–124. doi: 10.1038/25862. [DOI] [PubMed] [Google Scholar]
- 4.Brungart DS, Simpson BD. The effects of spatial separation in distance on the informational and energetic masking of a nearby speech signal. J Acoust Soc Am. 2002;112:664–676. doi: 10.1121/1.1490592. [DOI] [PubMed] [Google Scholar]
- 5.Shinn-Cunningham BG, Schickler J, Kopčo N, Litovsky RY. Spatial unmasking of nearby speech sources in a simulated anechoic environment. J Acoust Soc Am. 2001;110:1118–1129. doi: 10.1121/1.1386633. [DOI] [PubMed] [Google Scholar]
- 6.Adriani M, et al. Sound recognition and localization in man: Specialized cortical networks and effects of acute circumscribed lesions. Exp Brain Res. 2003;153:591–604. doi: 10.1007/s00221-003-1616-0. [DOI] [PubMed] [Google Scholar]
- 7.Nelken I. Processing of complex stimuli and natural scenes in the auditory cortex. Curr Opin Neurobiol. 2004;14:474–480. doi: 10.1016/j.conb.2004.06.005. [DOI] [PubMed] [Google Scholar]
- 8.Rauschecker JP, Tian B, Hauser M. Processing of complex sounds in the macaque nonprimary auditory cortex. Science. 1995;268:111–114. doi: 10.1126/science.7701330. [DOI] [PubMed] [Google Scholar]
- 9.Rauschecker JP. Processing of complex sounds in the auditory cortex of cat, monkey, and man. Acta Otolaryngol Suppl. 1997;532:34–38. doi: 10.3109/00016489709126142. [DOI] [PubMed] [Google Scholar]
- 10.Rauschecker JP, Tian B. Mechanisms and streams for processing of “what” and “where” in auditory cortex. Proc Natl Acad Sci USA. 2000;97:11800–11806. doi: 10.1073/pnas.97.22.11800. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Rauschecker JP. Cortical processing of complex sounds. Curr Opin Neurobiol. 1998;8:516–521. doi: 10.1016/s0959-4388(98)80040-8. [DOI] [PubMed] [Google Scholar]
- 12.Ahveninen J, et al. Task-modulated “what” and “where” pathways in human auditory cortex. Proc Natl Acad Sci USA. 2006;103:14608–14613. doi: 10.1073/pnas.0510480103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Tian B, Reser D, Durham A, Kustov A, Rauschecker JP. Functional specialization in rhesus monkey auditory cortex. Science. 2001;292:290–293. doi: 10.1126/science.1058911. [DOI] [PubMed] [Google Scholar]
- 14.Brunetti M, et al. Human brain activation during passive listening to sounds from different locations: An fMRI and MEG study. Hum Brain Mapp. 2005;26:251–261. doi: 10.1002/hbm.20164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Tata MS, Ward LM. Early phase of spatial mismatch negativity is localized to a posterior “where” auditory pathway. Exp Brain Res. 2005;167:481–486. doi: 10.1007/s00221-005-0183-y. [DOI] [PubMed] [Google Scholar]
- 16.Deouell LY, Heller AS, Malach R, D’Esposito M, Knight RT. Cerebral responses to change in spatial location of unattended sounds. Neuron. 2007;55:985–996. doi: 10.1016/j.neuron.2007.08.019. [DOI] [PubMed] [Google Scholar]
- 17.Warren JD, Zielinski BA, Green GG, Rauschecker JP, Griffiths TD. Perception of sound-source motion by the human brain. Neuron. 2002;34:139–148. doi: 10.1016/s0896-6273(02)00637-2. [DOI] [PubMed] [Google Scholar]
- 18.Krumbholz K, et al. Representation of interaural temporal information from left and right auditory space in the human planum temporale and inferior parietal lobe. Cereb Cortex. 2005;15:317–324. doi: 10.1093/cercor/bhh133. [DOI] [PubMed] [Google Scholar]
- 19.Hall DA, Moore DR. Auditory neuroscience: The salience of looming sounds. Curr Biol. 2003;13:R91–R93. doi: 10.1016/s0960-9822(03)00034-4. [DOI] [PubMed] [Google Scholar]
- 20.Seifritz E, et al. Neural processing of auditory looming in the human brain. Curr Biol. 2002;12:2147–2151. doi: 10.1016/s0960-9822(02)01356-8. [DOI] [PubMed] [Google Scholar]
- 21.Ghazanfar AA, Neuhoff JG, Logothetis NK. Auditory looming perception in rhesus monkeys. Proc Natl Acad Sci USA. 2002;99:15755–15757. doi: 10.1073/pnas.242469699. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Maier JX, Ghazanfar AA. Looming biases in monkey auditory cortex. J Neurosci. 2007;27:4093–4100. doi: 10.1523/JNEUROSCI.0330-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Zahorik P, Wightman FL. Loudness constancy with varying sound source distance. Nat Neurosci. 2001;4:78–83. doi: 10.1038/82931. [DOI] [PubMed] [Google Scholar]
- 24.Sigalovsky IS, Melcher JR. Effects of sound level on fMRI activation in human brainstem, thalamic and cortical centers. Hear Res. 2006;215:67–76. doi: 10.1016/j.heares.2006.03.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Ernst SM, Verhey JL, Uppenkamp S. Spatial dissociation of changes of level and signal-to-noise ratio in auditory cortex for tones in noise. Neuroimage. 2008;43:321–328. doi: 10.1016/j.neuroimage.2008.07.046. [DOI] [PubMed] [Google Scholar]
- 26.Brungart DS. Auditory localization of nearby sources. III. Stimulus effects. J Acoust Soc Am. 1999;106:3589–3602. doi: 10.1121/1.428212. [DOI] [PubMed] [Google Scholar]
- 27.Shinn-Cunningham BG, Kopčo N, Martin TJ. Localizing nearby sound sources in a classroom: Binaural room impulse responses. J Acoust Soc Am. 2005;117:3100–3115. doi: 10.1121/1.1872572. [DOI] [PubMed] [Google Scholar]
- 28.Mershon DH, King LE. Intensity and reverberation as factors in the auditory perception of egocentric distance. Percept Psychophys. 1975;18:409–415. [Google Scholar]
- 29.Hartmann WM. Localization of sound in rooms. J Acoust Soc Am. 1983;74:1380–1391. doi: 10.1121/1.390163. [DOI] [PubMed] [Google Scholar]
- 30.Zahorik P. Direct-to-reverberant energy ratio sensitivity. J Acoust Soc Am. 2002;112:2110–2117. doi: 10.1121/1.1506692. [DOI] [PubMed] [Google Scholar]
- 31.Kopčo N, Shinn-Cunningham BG. Effect of stimulus spectrum on distance perception for nearby sources. J Acoust Soc Am. 2011;130:1530–1541. doi: 10.1121/1.3613705. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Näätänen R, et al. Frequency and location specificity of the human vertex N1 wave. Electroencephalogr Clin Neurophysiol. 1988;69:523–531. doi: 10.1016/0013-4694(88)90164-2. [DOI] [PubMed] [Google Scholar]
- 33.Ulanovsky N, Las L, Nelken I. Processing of low-probability sounds by cortical neurons. Nat Neurosci. 2003;6:391–398. doi: 10.1038/nn1032. [DOI] [PubMed] [Google Scholar]
- 34.Jääskeläinen IP, Ahveninen J, Belliveau JW, Raij T, Sams M. Short-term plasticity in auditory cognition. Trends Neurosci. 2007;30:653–661. doi: 10.1016/j.tins.2007.09.003. [DOI] [PubMed] [Google Scholar]
- 35.Grill-Spector K, Malach R. fMR-adaptation: A tool for studying the functional properties of human cortical neurons. Acta Psychol (Amst) 2001;107:293–321. doi: 10.1016/s0001-6918(01)00019-1. [DOI] [PubMed] [Google Scholar]
- 36.Lü ZL, Williamson SJ, Kaufman L. Human auditory primary and association cortex have differing lifetimes for activation traces. Brain Res. 1992;572:236–241. doi: 10.1016/0006-8993(92)90475-o. [DOI] [PubMed] [Google Scholar]
- 37.Baumgart F, Gaschler-Markefski B, Woldorff MG, Heinze HJ, Scheich H. A movement-sensitive area in auditory cortex. Nature. 1999;400:724–726. doi: 10.1038/23390. [DOI] [PubMed] [Google Scholar]
- 38.Warren JD, Griffiths TD. Distinct mechanisms for processing spatial sequences and pitch sequences in the human auditory brain. J Neurosci. 2003;23:5799–5804. doi: 10.1523/JNEUROSCI.23-13-05799.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Graziano MS, Reiss LA, Gross CG. A neuronal representation of the location of nearby sounds. Nature. 1999;397:428–430. doi: 10.1038/17115. [DOI] [PubMed] [Google Scholar]
- 40.McAlpine D, Jiang D, Palmer AR. A neural code for low-frequency sound localization in mammals. Nat Neurosci. 2001;4:396–401. doi: 10.1038/86049. [DOI] [PubMed] [Google Scholar]
- 41.Salminen NH, May PJ, Alku P, Tiitinen H. A population rate code of auditory space in the human cortex. PLoS ONE. 2009;4:e7600. doi: 10.1371/journal.pone.0007600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Grothe B, Pecka M, McAlpine D. Mechanisms of sound localization in mammals. Physiol Rev. 2010;90:983–1012. doi: 10.1152/physrev.00026.2009. [DOI] [PubMed] [Google Scholar]
- 43.Hall DA, et al. Spectral and temporal processing in human auditory cortex. Cereb Cortex. 2002;12:140–149. doi: 10.1093/cercor/12.2.140. [DOI] [PubMed] [Google Scholar]
- 44.Wessinger CM, et al. Hierarchical organization of the human auditory cortex revealed by functional magnetic resonance imaging. J Cogn Neurosci. 2001;13:1–7. doi: 10.1162/089892901564108. [DOI] [PubMed] [Google Scholar]
- 45.Palomäki KJ, Tiitinen H, Mäkinen V, May PJ, Alku P. Spatial processing in human auditory cortex: The effects of 3D, ITD, and ILD stimulation techniques. Brain Res Cogn Brain Res. 2005;24:364–379. doi: 10.1016/j.cogbrainres.2005.02.013. [DOI] [PubMed] [Google Scholar]
- 46.Lehmann C, et al. Dissociated lateralization of transient and sustained blood oxygen level-dependent signal components in human primary auditory cortex. Neuroimage. 2007;34:1637–1642. doi: 10.1016/j.neuroimage.2006.11.011. [DOI] [PubMed] [Google Scholar]
- 47.Zimmer U, Lewald J, Erb M, Karnath HO. Processing of auditory spatial cues in human cortex: An fMRI study. Neuropsychologia. 2006;44:454–461. doi: 10.1016/j.neuropsychologia.2005.05.021. [DOI] [PubMed] [Google Scholar]
- 48.Johnson BW, Hautus MJ. Processing of binaural spatial information in human auditory cortex: Neuromagnetic responses to interaural timing and level differences. Neuropsychologia. 2010;48:2610–2619. doi: 10.1016/j.neuropsychologia.2010.05.008. [DOI] [PubMed] [Google Scholar]
- 49.Tardif E, Murray MM, Meylan R, Spierer L, Clarke S. The spatio-temporal brain dynamics of processing and integrating sound localization cues in humans. Brain Res. 2006;1092:161–176. doi: 10.1016/j.brainres.2006.03.095. [DOI] [PubMed] [Google Scholar]
- 50.Imig TJ, Irons WA, Samson FR. Single-unit selectivity to azimuthal direction and sound pressure level of noise bursts in cat high-frequency primary auditory cortex. J Neurophysiol. 1990;63:1448–1466. doi: 10.1152/jn.1990.63.6.1448. [DOI] [PubMed] [Google Scholar]
- 51.Werner-Reiss U, Groh JM. A rate code for sound azimuth in monkey auditory cortex: Implications for human neuroimaging studies. J Neurosci. 2008;28:3747–3758. doi: 10.1523/JNEUROSCI.5044-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Brungart DS, Durlach NI, Rabinowitz WM. Auditory localization of nearby sources. II. Localization of a broadband source. J Acoust Soc Am. 1999;106:1956–1968. doi: 10.1121/1.427943. [DOI] [PubMed] [Google Scholar]
- 53.Larsen E, Iyer N, Lansing CR, Feng AS. On the minimum audible difference in direct-to-reverberant energy ratio. J Acoust Soc Am. 2008;124:450–461. doi: 10.1121/1.2936368. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Carlile S. Virtual Auditory Space: Generation and Applications. New York: RG Landes; 1996. [Google Scholar]
- 55.Zahorik P. Assessing auditory distance perception using virtual acoustics. J Acoust Soc Am. 2002;111:1832–1846. doi: 10.1121/1.1458027. [DOI] [PubMed] [Google Scholar]
- 56.Dale AM. Optimal experimental design for event-related fMRI. Hum Brain Mapp. 1999;8:109–114. doi: 10.1002/(SICI)1097-0193(1999)8:2/3<109::AID-HBM7>3.0.CO;2-W. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Fischl B, Sereno MI, Dale AM. Cortical surface-based analysis. II: Inflation, flattening, and a surface-based coordinate system. Neuroimage. 1999;9:195–207. doi: 10.1006/nimg.1998.0396. [DOI] [PubMed] [Google Scholar]
- 58.Fischl B, Sereno MI, Tootell RB, Dale AM. High-resolution intersubject averaging and a coordinate system for the cortical surface. Hum Brain Mapp. 1999;8:272–284. doi: 10.1002/(SICI)1097-0193(1999)8:4<272::AID-HBM10>3.0.CO;2-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.