Abstract
The human superior temporal gyrus (STG) is critical for speech perception, yet the organization of spectrotemporal processing of speech within the STG is not well understood. Here, to characterize the spatial organization of spectrotemporal processing of speech across human STG, we use high-density cortical surface field potential recordings while participants listened to natural continuous speech. While synthetic broad-band stimuli did not yield sustained activation of the STG, spectrotemporal receptive fields could be reconstructed from vigorous responses to speech stimuli. We find that the human STG displays a robust anterior–posterior spatial distribution of spectrotemporal tuning in which the posterior STG is tuned for temporally fast varying speech sounds that have relatively constant energy across the frequency axis (low spectral modulation) while the anterior STG is tuned for temporally slow varying speech sounds that have a high degree of spectral variation across the frequency axis (high spectral modulation). This work illustrates organization of spectrotemporal processing in the human STG, and illuminates processing of ethologically relevant speech signals in a region of the brain specialized for speech perception.
SIGNIFICANCE STATEMENT Considerable evidence has implicated the human superior temporal gyrus (STG) in speech processing. However, the gross organization of spectrotemporal processing of speech within the STG is not well characterized. Here we use natural speech stimuli and advanced receptive field characterization methods to show that spectrotemporal features within speech are well organized along the posterior-to-anterior axis of the human STG. These findings demonstrate robust functional organization based on spectrotemporal modulation content, and illustrate that much of the encoded information in the STG represents the physical acoustic properties of speech stimuli.
Keywords: functional organization, human STG, human superior temporal gyrus, modulation tuning, modulotopic, spectrotemporal processing
Introduction
A remarkable array of imaging, electrophysiological, and functional lesion studies have implicated the human superior temporal gyrus (STG) in speech processing and perception. However, organization of basic spectrotemporal processing during speech perception is not well understood (Boatman et al., 1997; Binder et al., 2000; Boatman, 2004; Hickok and Poeppel, 2007; Rauschecker and Scott, 2009). While tonotopy is a prominent organizing principle in the ascending auditory pathway and has been investigated in STG (Talavage et al., 2004; Striem-Amit et al., 2011; Moerel et al., 2012; Nourski et al., 2014), additional characterization of spectrotemporal processing in the context of speech processing is less well defined. Here, using speech stimuli to assess broad organizational principles during natural speech perception, we examine fundamental spectrotemporal response parameters across the STG.
A major goal of sensory neuroscience is to understand how sensory systems encode natural stimuli and, in particular, how spectrotemporal processing is organized in STG during speech processing. Traditionally sensory encoding has been studied with simple parameterized stimuli. However, recently, there has been increasing use of natural stimuli to study sensory encoding (Theunissen et al., 2000; David et al., 2004; Sharpee et al., 2006; Talebi and Baker, 2012). Part of the motivation for this is increasing evidence that sensory processing is adapted to statistics of behaviorally relevant stimuli. Data suggest receptive fields and response properties of neurons are matched to the statistics of natural inputs to maximize efficiency of information transmission (Dong and Atick, 1995; Rieke et al., 1995; Dan et al., 1996; Olshausen and Field, 1996; Hsu et al., 2004). In higher-order areas, systems become more selective for natural stimuli and less responsive to synthetic or noise stimuli (Theunissen et al., 2000; Wilkinson et al., 2000; Felsen et al., 2005; Talebi and Baker, 2012). In areas that respond broadly to both synthetic and natural stimuli, natural inputs can push neurons into different operating ranges and activate nonlinearities that evoke response properties not present with synthetic stimuli alone (David et al., 2004; Talebi and Baker, 2012). Finally, neurons adapt on a moment-by-moment basis to match their response properties to statistics of the input stimulus, thus making descriptions of sensory encoding inherently specific to the statistics and class of stimuli used for characterization (Smirnakis et al., 1997; Brenner et al., 2000; Fairhall et al., 2001; Sharpee et al., 2006). Thus, in situations where the question of interest is to understand encoding of natural stimuli, it is advantageous to use natural inputs, such as speech, to characterize the system.
Given the relevance of STG in speech processing, we are interested in characterizing spectrotemporal processing in the context of natural speech. However, characterization of spectrotemporal processing based on speech stimuli is difficult due to the statistically biased and highly correlated structure of natural signals. Recently, a number of techniques have been developed to address this bias (Theunissen et al., 2001; Paninski, 2004; Sharpee et al., 2004; David et al., 2007). Here we use maximally informative dimension (MID) analysis, an information-based method designed to lift the requirement of statistically tractable stimuli and allow the use of more complex, but ethologically relevant, natural stimuli for receptive field characterization (Sharpee et al., 2004; Atencio et al., 2008).
By using electrocorticography (ECoG) to record speech-driven activity, we compute spectrotemporal receptive fields (STRFs) using MID analysis and assess the organization of spectrotemporal processing in human STG during natural speech processing. While synthetic broad-band stimuli did not yield strong activation of the STG, STRFs could be reconstructed from vigorous responses to speech stimuli. We find that the human STG displays a robust anterior–posterior spatial distribution of spectrotemporal tuning characterized by tuning for high temporal and low spectral modulation speech features posteriorly and high spectral and low temporal modulation speech features anteriorly. This work further defines organization of spectrotemporal processing of speech in human STG, and illuminates processing of ethologically relevant speech signals in a region of the brain specialized for speech perception.
Materials and Methods
Participants and neural recordings.
Subdural ECoG arrays (interelectrode distance, 4 mm) were placed unilaterally in eight patient volunteers [three right hemisphere (one female/two male), five left hemisphere (three female/two male)] undergoing a neurosurgical procedure for the treatment of medication refractory epilepsy. Seven of eight participants were native English speakers; all were fluent in English. All participants had normal hearing. None had communication deficits. All experimental protocols were approved by the University of California, San Francisco Institutional Review Board and Committee on Human Research. Location of array placement was determined by clinical criteria alone. Participants were asked to passively listen to 15–25 min of natural speech while ECoG signals were recorded simultaneously. Some participants also passively listened to dynamic moving ripple (DMR) or temporally orthogonal ripple combination (TORC) stimuli (see below). Signals were amplified and sampled at 3052 Hz. After rejection of electrodes with excessive noise or artifacts, signals were referenced to a common average and the high-gamma band (70–150 Hz) was extracted as the analytic amplitude of the Hilbert transform (Crone et al., 2001; Chang et al., 2011). Signals were subsequently downsampled to 100 Hz. The resulting signal for each electrode was z-scored based on the mean and SD of activity during the entire block.
Stimuli
Speech stimuli.
Speech stimuli were delivered binaurally through free-field speakers at ∼70 dB average sound pressure level. The frequency power spectrum of stimuli spanned 0–8000 Hz. The stimulus set consisted of prerecorded (2–4 s) sentences from the phonetically transcribed Texas Instruments/Massachusetts Institute of Technology (TIMIT) speech corpus with 1 s silent intervals between each sentence presentation (Garofolo et al., 1993). Each participant was presented 484–499 sentences. The speech corpus included 286 male and 116 female speakers, with 1–3 sentences spoken per speaker, and unique lexical content for each sentence. Spectrogram representations of speech stimuli were generated using a cochlear model of auditory processing (Yang et al., 1992).
Synthetic stimuli.
DMR stimuli were presented to four subjects (EC63, GP30, GP31, and GP33) at 70 dB average sound pressure level. Of these, only GP31 was included in the functional organization analysis. The other three participants had less predictive or noncontiguously spaced MID-based STRFs along STG and thus were less useful for characterizing organization of spectrotemporal processing (but still of sufficient data quality to assess responsiveness). The DMR was composed of a bank of sinusoidal carriers whose amplitude is modulated over time with a spectrotemporal ripple envelope (Depireux et al., 2001; Escabi and Schreiner, 2002). The amplitude distribution of the envelope was Gaussian and had a maximum modulation depth of 40 dB. The DMR is broadband (500–20,000 Hz) and consisted of ∼50 sinusoidal carriers per octave with randomized phase. The spectral and temporal modulation frequency parameters defined the characteristics of the ripple modulation envelope at any given point in time. The spectral modulation parameter defines the number of spectral peaks per octave. The temporal modulation parameter defines the speed and direction of the peak's change. Spectral and temporal modulation parameters were varied randomly and independently during the stimulus. The spectral modulation parameter varied between 0 and 4 cycles/octave (maximum rate of change, 1 Hz) and the temporal modulation rate parameter varied between −35 and +35 Hz (rate of change, ≤3 Hz). TORC stimuli were presented to two subjects (EC2 and EC28) at 70 dB average sound pressure level. TORC stimuli are generated by modulating broadband white noise (frequency content, 250–8000 Hz) with a combination of 12 temporally orthogonal ripples (Klein et al., 2000). The spectral modulation of ripples ranged between 0 and 1.4 cycles/octave and the temporal modulation of ripples ranged between 4 and 48 Hz.
Analysis
STRFs.
STRFs were computed with two different methods designed specifically for use with naturalistic stimuli—MID analysis and normalized reverse correlation (Theunissen et al., 2001; Sharpee et al., 2004). To compute STRFs using MID analysis, a gradient ascent procedure was used to search for the receptive field that maximizes the Kullback–Leibler divergence between the raw distribution of STRF-stimulus projection values and the distribution of STRF-stimulus projection values weighted by the magnitude of the response. STRF estimates based on normalized reverse correlation (normalization based on the stimulus autocorrelation matrix) were computed using ridge regression with open source code available at http://strfpak.berkeley.edu/. Regularization was controlled by fitting a tolerance hyperparameter via cross-validation (David et al., 2007). STRFs were computed with both methods on the same estimation set (90% of the total data) and cross-validated on the same test set, which was withheld from the estimation process (10% of the data).
Modulation tuning.
To characterize modulation tuning-based organization, the modulation transfer function (MTF) for each site was computed by taking the magnitude of the two-dimensional Fourier transform (ℑ2{·}) of each STRF according to the following equation:
Where (t, f) are time and frequency and (ωt, ωs) are temporal and spectral modulation, respectively. The best spectrotemporal modulation (bSTM) is defined as the peak of the MTF (see Fig. 3A). The sign of the best temporal modulation determines the drift direction of spectral content within each ripple (increasing or decreasing; see Fig. 3B). Similar to previous work, we are interested in the magnitude of temporal modulation and therefore take the absolute value of best temporal modulations (Langers et al., 2003; Santoro et al., 2014). The ensemble MTF was computed by normalizing each MTF to have a sum of 1, then computing the average MTF across all sites and all participants.
Spatial analysis.
Permutation tests were used to assess whether STRF-derived parameters, such as bSTM, MTF cluster type, or best frequency (BF), were locally organized. Each test determined whether the average level of similarity between a site and its neighbors would be expected if the true underlying spatial organization were random. For each site, we compute difference between a site and its neighbors (anterior, posterior, dorsal, and ventral), take the absolute value, and then compute the average. This value is computed for each site and then averaged across the whole map to compute the mean neighborhood similarity index. The map is then randomly permuted 10,000 times and the map neighborhood similarity index recomputed on each permutation to generate a distribution of randomized map neighborhood similarity indices. The true neighborhood similarity index is compared with the randomized map neighborhood similarity index distribution to assess the level of significance. To determine significance values for each spectrotemporal modulation map (two parameters at each site: one for best spectral modulation, one for best temporal modulation), the above procedure was repeated with a two-parameter neighborhood similarity metric. The two-parameter neighborhood similarity metric was generated by normalizing best temporal modulation values and best spectral modulation values by their respective maxima so both sets of data had a range of 0–1. For each parameter at each site, we compute difference between a site and its neighbors (anterior, posterior, dorsal, and ventral), take the absolute value, and then compute the average. This value for each parameter was then added together to generate the two-parameter similarity metric. This metric was averaged across sites to generate the two-parameter neighborhood similarity index for the entire map. While there are many metrics that could potentially be used to quantify organization, we choose to use neighborhood similarity because of its generality in that previously described forms of organization (linear gradients, nonlinear gradients, clusters, or modules) have neighborhood similarity as a common feature.
Group analysis was achieved by combining data across subjects in a common coordinate system defined by anatomical landmarks. The x-axis was defined as a line that runs parallel to the long axis of STG along its dorsal–ventral midpoint. The y-axis was defined as a line orthogonal to the x-axis with the origin aligned with the anterior temporal pole (see Fig. 6D, red). After coordinates of ECoG sites were defined, data were binned at 4 mm resolution consistent with the interelectrode distance of the ECoG arrays (4 mm interelectrode distance). Tonotopic and spectrotemporal modulation maps represent topographic distributions of BF and preferred spectrotemporal modulation across the two-dimensional surface of the STG. To compute the gradient of such maps, a two-dimensional plane was fit to the data using linear (planar) regression. The direction of steepest angle of the plane is taken to be the gradient of the topographic map (Baumann et al., 2011). The gradient direction is specified as an angle counterclockwise from the x-axis of the coordinate system that runs in the anterior–posterior direction along the long axis of STG. For spectrotemporal modulation maps, which represent a map of two parameters, spectral and temporal modulation values were normalized by their group maxima so each set of data has a range of 0–1. Gradients were calculated for the spectral modulation map and temporal modulation map independently and then averaged to determine the gradient for the joint spectrotemporal modulation map.
Results
STRF maps
Given the role of STG in speech processing, we are interested in characterizing functional organization in STG based on responses to natural speech stimuli. After ECoG array placement, participants were asked to passively listen to 15–25 min of natural speech, which consisted of prerecorded (2–4 s) sentences from the phonetically transcribed TIMIT speech corpus (Garofolo et al., 1993). After data collection, an STRF was computed off-line from the local field potential signal at each cortical site to generate an STG STRF map in each subject (Fig. 1A). STRFs were computed using the high-gamma (70–150 Hz) band of ECoG recordings (Crone et al., 2001), which correlates with spiking activity (Ray and Maunsell, 2011) and spike-based tuning properties in the midlaminar auditory cortex (Steinschneider et al., 2008). To compute STRFs, two different methods designed for use with natural signals were used: MID analysis (Sharpee et al., 2004) and normalized reverse correlation (Theunissen et al., 2001). Using these two methods, we computed MID and normalized reverse correlation-based STRFs using responses to speech stimuli and evaluated their performance through cross-validation. While MID-based STRFs were generally similar to normalized reverse correlation-based STRFs, MID-based STRFs produced higher prediction values and were therefore used in the remainder of the analysis (Fig. 1C; p < 0.001, Wilcoxon signed-rank test, mean percentage increase in prediction: 19.0 ± 1.9% SEM).
For purposes of studying organization of spectrotemporal processing, it is necessary to restrict analysis to STRFs that characterize the underlying spectrotemporal processing. Similar to previous work, only sites that predict ≥5% of the variance in the response were included in the analysis (Fig. 1D,E; Kim and Doupe, 2011). This set of STRFs showed relatively high prediction performance (mean r = 0.48 ± 0.12 SD; Fig. 1D), comparable to STRF prediction values from lower-order areas (Calabrese et al., 2011; Kim and Doupe, 2011). These highly predictive STRFs are based on the acoustic properties of speech alone and the model does not take into account other aspects of speech, such as semantic meaning. This indicates that the physical acoustic properties of speech are a major component of the encoded information in STG and the computed speech-based STRFs provide a good characterization of the underlying spectrotemporal processing of speech. Similar to STRFs found in lower-order auditory areas, STRFs within STG showed clear contiguous excitatory and inhibitory regions and structure characteristic of “temporal” and “spectral” STRFs (Fig. 2; Nagel and Doupe, 2008; Atencio et al., 2008). Temporal STRF aspects exhibit short excitatory regions followed by short inhibitory regions (Fig. 2, •). These types of STRFs are characteristic of sites tuned for rapid temporal modulations in sound energy that occur at the onset or offset of sound and within many consonants. Spectral STRF aspects are characterized by temporally long excitatory regions flanked by inhibition on one or both sides (Fig. 2, ■). These types of STRFs are characteristic of sites tuned to sound energy that fluctuates across frequency, but is relatively constant over time, such as the formant structure prevalent in vowel sounds. These data show that STG STRFs have characteristics similar to STRFs found in lower-order auditory areas and are predictive of responses to speech stimuli, demonstrating their ability to characterize spectrotemporal processing of speech in STG.
Organization of spectrotemporal processing of speech in human STG
Given that the STRF characterizes spectrotemporal processing at each ECoG site, the STRF map characterizes how processing varies across STG. However, as Figure 2 shows, the spectrotemporal structure within each STRF is relatively complex, making it difficult to empirically visualize or quantify organizational aspects from the raw STRF maps. One alternative representation of the STRF that contains identical spectrotemporal information (minus phase), but which exhibits visually less complex structure, is the MTF (Fig. 3). The MTF characterizes spectrotemporal modulation tuning of each STRF with a structure that is often more localized in modulation space than the corresponding STRF structure within the time–frequency domain (Fig. 3A, compare STRF, MTF). Additionally, STRF features that are shifted in frequency or time, but have the same overall spectrotemporal structure, have the same MTF representation because phase information is discarded. Using this alternative, but less complex and equivalent representation of the STRF, we examined the distribution of MTFs within the STG. A map of MTFs for one participant (EC6) is shown in Figure 4. The “temporal” and “spectral” STRFs shown in Figure 2 have corresponding temporal and spectral MTFs in Figure 4. Temporal transfer functions (•) are characterized by tuning shifted away from the vertical midline toward high temporal modulations. Spectral transfer functions (■) are characterized by tuning shifted away from the base of the MTF toward high spectral modulations. In this representation it becomes more apparent, on visual inspection, that sites in the posterior STG are tuned for high temporal modulations (energy shifted away from the vertical axis) and sites in the anterior STG are tuned for slow temporal modulations, but high spectral modulations (energy falling along the vertical axis and shifted upward).
To more thoroughly characterize this apparent nonrandom distribution of spectrotemporal processing, representative MTFs of STG were first identified by k-means clustering of all MTFs from all participants. Silhouette criterion values were used to identify the number of data clusters that maximize intracluster similarity compared with similarity between neighboring clusters. The centroid MTFs from this analysis are shown in Figure 5A, along with their 50% energy contours. Compared with the average MTF for STG (Fig. 5E), centroid MTFs occupy discrete regions of modulation tuning space in an orderly fashion from high spectral/low temporal modulation tuning (Fig. 5A, left MTF with red contour) to high temporal/low spectral modulation tuning (Fig. 5A, right MTF with yellow contour). To examine the distribution of overall spectrotemporal modulation tuning across STG, a group MTF map was computed and each MTF within the map was classified according to its cluster identity (Fig. 5B). Cluster identities from the group map are plotted in Figure 5C. This map shows significant local organization of the MTF type (p < 1.0 × 10−5, neighborhood similarity permutation test) and illustrates anterior-to-posterior organization of modulation tuning along the STG in which high spectral modulation/low temporal modulation tuned MTFs are located anteriorly (red). By contrast, high temporal modulation/low spectral modulation tuned MTFs are located posteriorly (yellow). This is further quantified by the average location of each MTF type along the anterior-to-posterior extent of the STG (Fig. 5D). Overall, this shows that the STG is tuned for temporally fast-changing speech sounds that have relatively constant energy across the frequency axis (low spectral modulation) in the posterior STG while the anterior STG is tuned for temporally slow-changing speech sound that have a high degree of spectral variation across the frequency axis (high spectral modulation).
To further illustrate the organized distribution of spectrotemporal processing along the anterior–posterior extent of the STG, we plot the distribution of MTF peak values across the STG. The peak of the MTF defines the bSTM, and in this context represents the dominant spectrotemporal modulation in the feature each site is tuned to (as captured by the STRF). It should be noted that examining the distribution of MTF peaks (bSTM values) derived for natural speech does not imply that the STG will respond to individual spectrotemporal modulations. Rather, the localized nature of modulation energy in MTFs makes the peak a good descriptor of the overall MTF tuning. The distribution of bSTMs can, therefore, be used to further characterize overall organization of spectrotemporal modulation tuning in the STG. The distribution of bSTMs for all sites and from all participants is shown in Figure 6A. The relationship between best temporal modulation and best spectral modulation shows a trade-off pattern in which spectral modulation decreases with increasing temporal modulation. This relationship between spectral and temporal modulation tuning has been observed previously in other auditory areas, including the human primary auditory cortex (Schönwiesner and Zatorre, 2009), the cat primary auditory cortex, the cat medial geniculate body, the cat inferior colliculus (Miller et al., 2002; Rodríguez et al., 2010; Atencio and Schreiner, 2012), and in models of auditory processing that predict receptive field structure based on efficient coding hypotheses (Carlson et al., 2012).
To examine the spatial distribution of bSTM tuning across the human STG, we plot bSTM values at their corresponding cortical position within each subject (Fig. 6A, color scale, B). Most participants show a significant degree of nonrandom local organization that is consistent across participants (EC6, p < 1.0 × 10−5; GP31, p = 5.0 × 10−4; EC36, p = 0.029; EC28, p = 4.5 × 10−5; EC53, p = 0.015; EC58, p = 0.10; EC56, p = 0.06; EC2, p = 0.42; two-parameter neighborhood similarity permutation test). Posterior regions of the STG show primarily high temporal modulation/low spectral modulation tuning (blue), while more anterior regions show primarily high spectral modulation/low temporal modulation tuning (green). With respect to interhemispheric differences, there was no significant difference in best temporal modulation tuning (left: mean, 0.89 ± 0.8 Hz; right: mean, 0.72 ± 0.8 Hz, p = 0.17, Wilcoxon rank-sum test) and a subtle difference in best spectral modulation tuning (left: mean, 0.14 ± 0.17 cycles/octave; right: mean, 0.08 ± 0.13 cycles/octave, p = 0.015, Wilcoxon rank-sum test). To examine data across subjects, we computed a group modulation tuning map, which is shown in Figure 6D. Similar to individual subject maps, the group map shows primarily high temporal modulation/low spectral modulation tuning posteriorly (blue) and high spectral modulation/low temporal modulation tuning anteriorly (green, p < 1.0 × 10−5, neighborhood similarity permutation test). The dominant gradient within the group modulation map runs in the posterior-to-anterior direction (+176° counterclockwise from the 3 o'clock position; Fig. 6D) nearly parallel with the anterior–posterior axis of the STG. Again, these data represent organized spectrotemporal processing along the STG, in which posterior regions are tuned to temporally fast sounds with relatively constant energy along the spectral axis (low spectral modulation) and more anterior regions are tuned for temporally slower sound with high variation in energy along the spectral axis (high spectral modulation). To further characterize modulation organization, average best temporal modulation and best spectral modulation tuning are plotted as a function of distance along the anterior–posterior extent of the STG (Fig. 6E,F, red). Additionally, modulation tuning was plotted as a function of map distance after individual maps have been aligned by their individual gradients (Fig. 6E,F, blue and green lines). This shows a transition from high temporal modulation to high spectral modulation along the posterior–anterior aspect of the STG. A similar but less robust distribution of modulation tuning was also seen for STRFs based on normalized reverse correlation (Fig. 6E,F, insets). Further analysis showed no compelling additional organization within the maps based on inclusion of the temporal modulation sign (positive vs negative), no organized distribution of prediction performance that paralleled the distribution of spectrotemporal modulation tuning, and no difference in prediction performance between “spectral” and “temporal” STRFs (r value mean, 0.50 ± 0.13 vs 0.48 ± 0.11, respectively; p = 0.23, Wilcoxon rank-sum test; “spectral” and “temporal” STRFs defined as having bSTM values above or below the diagonal in Fig. 5A, respectively). Collectively, these data reveal spatially organized spectrotemporal processing of speech in human STG in which tuning varies from high temporal/low spectral modulation tuning in the posterior STG to high spectral/low temporal modulation tuning in the middle STG. This organization is reminiscent of the posterior-to-anterior distribution of temporal modulation tuning found at the lateral aspect of the superior temporal plane in macaques as may have been predicted for directly adjacent areas (Baumann et al., 2015).
To determine whether organization of spectrotemporal modulation tuning within the STG is present during processing of more traditional stimuli used to characterize modulation tuning, DMR stimuli (Depireux et al., 2001; Escabi and Schreiner, 2002) were presented to four subjects while recording from the STG. Consistent with previous work (Schönwiesner and Zatorre, 2009), these synthetic stimuli did not sufficiently activate the human STG (Fig. 7). Additionally, TORC stimuli, a variant of moving ripple stimuli (Klein et al., 2006), were presented in two participants without significant sustained activation of the STG (data not shown). These data are consistent with only minimal temporal lobe activation outside of the superior temporal plane by nonspeech stimuli as seen by Overath et al. (2015) and Schönwiesner and Zatorre (2009) and implies a degree of specificity of the organization demonstrated here for speech in that more traditional nonspeech synthetic stimuli do not drive robust activity in the STG.
Topography of spectral tuning
Last, we analyzed the spatial distribution of frequency tuning in human STG to examine the relationship between spectral frequency and modulation tuning in this region. To examine the distribution of spectral tuning, two metrics derived from the STRF were used to characterize BF. The first metric is defined as the peak excitatory value of the STRF (STRF-BF) and represents the frequency with the largest gain at any time point within the STRF. The second metric is defined as the peak of the spectral receptive field (SRF). The SRF is obtained by summing across the temporal dimension of the STRF. The peak of the SRF (SRF-BF) takes into account all temporal lags of the STRF and identifies the frequency associated with the largest net gain. Both metrics of BF showed similar tuning across ECoG sites (Pearson correlation coefficient, 0.693) and the average difference in BF between an ECoG site and its directly adjacent neighbors (anterior, posterior, dorsal, ventral) was not significantly different between the two metrics, reflecting equal degrees of neighborhood similarity (mean difference in BF between adjacent sites: STRF-BF, 0.7 ± 0.65 octaves SE; SRF-BF, 0.84 ± 0.72 octaves SE, p = 0.065, Wilcoxon rank-sum test). Individual maps for the topography of STRF-BF values are shown in Figure 8B. For both metrics, only two of eight participants showed a significant degree of local organization based on neighborhood similarity (STRF-BF p/SRF-BF p: EC6, p = 0.12/0.23; GP31, p = 0.004/0.007; EC36, p = 0.10/0.04; EC28, p = 0.02/0.75; EC53, p = 0.29/0.13; EC58, p = 0.27/0.07; EC56, p = 0.39/0.20; EC2, p = 0.11/0.25; neighborhood similarity permutation test). Examination of the group map also showed a low degree of local organization based on neighborhood similarity (Fig. 8A, group: STRF-BF, p = 0.1; SRF-BF, p = 0.06; neighborhood similarity permutation test). Despite the general lack of significant tonotopic organization within individual subjects based on neighborhood similarity, mean BF in the group data decreased as a function of distance along the STG (Fig. 8C). One aspect of spectral tuning that increases the complexity of characterizing spectral tuning organization is the widespread presence of multipeaked SRFs in the STG. Figure 8D shows an example of a multipeaked SRF with four peaks at the 50% maximum level (red line). We quantified the degree of multipeaked tuning in the STG by measuring the average number of spectral tuning peaks as a function of percentage maximum level (Fig. 8E). At 80% maximum, the average number of spectral peaks was 1.4 ± 0.7 with the number of peaks increasing to 2.6 ± 1.1 at 50% maximum. We also quantified the distribution of SRF types (classified by peak number) for each level of percentage maximum. Figure 8F shows the distribution of SRF types at each percentage maximum level. At 90% maximum, 72% of the SRFs were single peaked, 25% were double peaked, and 3% had three peaks. By 70% maximum, there are more multipeaked SRFs than single-peaked SRFs (52% multipeaked versus 48% single peaked). Last, we examined the overall distribution of BFs and found the majority of BFs are concentrated below 1000 Hz (Fig. 8G). This low-frequency tuning is consistent with previous work characterizing tonotopic organization across the human auditory cortex in which the STG occupies a low-frequency reversal region between putative cochleotopic maps that span the planum temporale, the STG, and the superior temporal sulcus (Striem-Amit et al., 2011; Moerel et al., 2012). Collectively, these data show that the STG has a substantial proportion of multipeaked spectral tuning, with low-to-moderate tonotopic organization within the STG, and a high concentration of low BF frequencies consistent with its placement as a low-frequency region within larger-scale cochleotopic maps (Striem-Amit et al., 2011; Moerel et al., 2012).
Discussion
In this study we used natural speech stimuli in conjunction with MID analysis and ECoG from human participants to investigate the organization of spectrotemporal processing of speech in the human STG. Based on receptive field maps, we found tuning for temporally fast sound with relatively constant energy across the frequency axis (low spectral modulation) in the posterior STG and a transition to tuning for temporally slower sound with higher variation in spectral energy across the frequency axis (high spectral modulation) in the anterior STG. Additionally, we found that the STG shows BF tuning <1000 Hz, which is consistent with its placement as a low-frequency reversal region in larger-scale cochleotopic maps (Striem-Amit et al., 2011; Moerel et al., 2012). These data expand our view of how the spectrotemporal processing of speech in the STG is organized and demonstrate organized tuning for different acoustic aspects of speech sounds along the human STG.
Our data reveal topographic organization of spectrotemporal modulation tuning in the STG derived from speech. A previous study examined modulation tuning in the human auditory cortex using fMRI and spectrotemporal ripple stimuli (Schönwiesner and Zatorre, 2009). They observed cortical areas responsive to ripple stimuli in primary and secondary auditory cortex within superior temporal plane, but did not find evidence of organized modulation tuning. Consistent with our study, individual spectrotemporal ripple stimuli did not drive activity in the STG. Taken alone, this may indicate organized spectrotemporal modulation processing emerges in the STG. However, recent work has shown topographic organization of temporal modulation tuning alone in the human primary auditory cortex using fMRI and amplitude modulated white noise stimuli (Herdener et al., 2013). Consistent with this, Schönwiesner et al. (2009) endorse a tendency for temporal modulation tuning to increase from the medial Heschl's gyrus (HG) to the lateral HG in a manner similar to that supported by data from Herdener et al. (2013). Additionally, organized temporal modulation tuning has also been found in subcortical areas, such as the cat and the primate midbrain (Schreiner and Langner, 1988; Baumann et al., 2011). Organized joint spectrotemporal modulation tuning has been characterized in the cat inferior colliculus (Rodríguez et al., 2010) and the cat auditory cortex (Atencio and Schreiner, 2012). Collectively, these studies provide evidence for organized spectrotemporal modulation tuning in various subcortical and cortical auditory areas and indicate that such processing may be a prevalent form of functional organization within the auditory system.
In contrast to robust activation of STG by speech stimuli, our data with individual ripples (DMR stimulus) and TORCs indicate that these stimuli do not robustly activate STG, which is consistent with previous work (Schönwiesner and Zatorre, 2009). The two primary differences of speech from ripple stimuli are acoustic complexity and behavioral relevance. Although spectrotemporal modulations represent fundamental elements of complex sounds, such that a linear combination of these features can generate any segment of speech, speech is more acoustically complex due to such features as increased temporal envelope variability from frequent onsets in words or syllables, harmonic and formant structure, and formant transitions. These additional acoustic characteristics of speech may be necessary to drive speech activation and may underlie selectivity seen in areas of the human auditory cortex (Overath et al., 2015). Future work, akin to the analysis-by-synthesis approach by McDermott and Simoncelli (2011), may be able to determine which acoustic properties of sound are most important for activation of the STG. Our work shows that in the context of speech processing, the STG shows organized tuning for temporal and spectral modulation properties of speech.
Acoustic properties may not be the only etiology of speech selectivity in the STG. Along the auditory pathway, acoustic information is eventually transformed to semantic information, which confers behavioral relevance to the stimulus. Synthetic modulation stimuli lack semantic information and thus lack an important component of speech stimuli that may be responsible for activation of the STG. It is difficult to quantify the contribution of acoustic versus semantic information for driving activity in the STG in our dataset. The robust ability of STRFs to predict responses to novel speech stimuli relies on the acoustic content of sound alone. This suggests a significant component of the encoded information represents acoustic rather than semantic information and is consistent with recent work in the superior temporal sulcus showing significant selectivity for speech based on acoustic properties alone (Overath et al., 2015).
The range of temporal tuning within the human auditory cortex gives insight into the timescale of analysis being performed. Fast temporal tuning reflects processing on the phonemic timescale while slower temporal tuning reflects processing on syllabic or prosodic timescales. Currently there is considerable variation in reported temporal tuning and shape of the temporal MTF for both HG and the STG (HG, range of average best temporal modulation: 1.5–10 Hz; STG, range of average best temporal modulation: 2–8 Hz; Binder et al., 1994; Tanaka et al., 2000; Harms and Melcher, 2002; Langers et al., 2003; Liégeois-Chauvel et al., 2004; Rinne et al., 2005; Schönwiesner and Zatorre, 2009; Overath et al., 2012; Pasley et al., 2012; Wang et al., 2012; Gross et al., 2013). There are a number of possible contributors to this variability. Within these studies, a wide range of stimuli have been used to characterize temporal tuning, including sinusoidally amplitude-modulated (SAM) tones, SAM noise, moving spectrotemporal ripples, tone pips, harmonic complexes, noise bursts, consonant–vowel pairs, and continuous speech. This variation is relevant because temporal tuning is not invariant to stimulus type (Eggermont, 2002; Malone et al., 2007, 2013; Zheng and Escabí, 2008) and differences in temporal tuning in the various studies may be partially attributable to the vast array of stimulus types used to characterize the system. In this study, we use speech stimuli to determine temporal tuning in the STG and thus characterize temporal processing relevant for speech processing in the STG.
Variability in temporal tuning within the human auditory cortex may also be attributable to differences in the range and spacing of temporal modulation content used to characterize the system (Edwards and Chang, 2013). The ideal stimulus set would extend from 0 Hz to temporal modulations well beyond the limits of the system and have resolution fine enough to resolve peaks and troughs within the temporal MTF. However, variable resolution and differences in the upper and lower limits of temporal modulations used to test temporal tuning have led to variable interpretations of the shape of the temporal MTF (low pass vs bandpass) and average temporal tuning for HG and the STG (Edwards and Chang, 2013). In this study, we used speech stimuli to define the temporal tuning relevant for speech processing. Speech contains a continuous distribution of temporal modulation content rather than discretely spaced temporally modulated stimuli (ripples, amplitude-modulated white noise, click trains) that may miss peaks and troughs in temporal tuning functions. Speech also spans the range of temporal modulations of interest for speech processing. Thus, an analysis using speech will not miss upper or lower ranges of tuning that may be missed using a predefined and narrower range of temporal modulations.
Finally, differences in metrics used to characterize temporal tuning have led to variable characterizations of temporal tuning within the STG and other areas. Nonsynchronized metrics of temporal tuning (mean activity as a function of temporal modulation) and synchronized metrics of temporal tuning (vector strength and STRF-based MTFs) are in prevalent use within the literature, but can show different tuning properties (Eggermont, 2002; Zheng and Escabí, 2008). Consistent with this, recent work in the STG has shown that a nonsynchronized metric and a synchronized metric show low-pass and bandpass tuning, respectively (Pasley et al., 2012). In this study we characterized the synchronized component of temporal tuning, the dominant form of temporal tuning for the range of temporal modulations most important for speech perception (Elliott and Theunissen, 2009; Pasley et al., 2012), and find a low-pass distribution of temporal tuning in the STG most consistent with processing on prosodic and syllabic timescales.
Footnotes
This work was supported by National Institutes of Health Grants DP2-OD00862 (to E.F.C.), R01-DC012379 (to E.F.C.), R01-DC02260 (to C.E.S.), and F32-DC014192 (to L.S.H.), and the McKnight Foundation (to E.F.C.). E.F.C. is a New York Stem Cell Foundation—Robertson Investigator. This research was also supported by The New York Stem Cell Foundation. The authors declare no competing financial interests. We thank Craig Atencio, Brian Malone, Bryan Seybold, and Erik Edwards for their insightful comments on the manuscript.
References
- Atencio CA, Schreiner CE. Spectrotemporal processing in spectral tuning modules of cat primary auditory cortex. PloS One. 2012;7:e31537. doi: 10.1371/journal.pone.0031537. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Atencio CA, Sharpee TO, Schreiner CE. Cooperative nonlinearities in auditory cortical neurons. Neuron. 2008;58:956–966. doi: 10.1016/j.neuron.2008.04.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baumann S, Griffiths TD, Sun L, Petkov CI, Thiele A, Rees A. Orthogonal representation of sound dimensions in the primate midbrain. Nat Neurosci. 2011;14:423–425. doi: 10.1038/nn.2771. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baumann S, Joly O, Rees A, Petkov CI, Sun L, Thiele A, Griffiths TD. The topography of frequency and time representation in primate auditory cortices. Elife. 2015:4. doi: 10.7554/eLife.03256. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Binder JR, Rao SM, Hammeke TA, Frost JA, Bandettini PA, Hyde JS. Effects of stimulus rate on signal response during functional magnetic resonance imaging of auditory cortex. Brain Res Cogn Brain Res. 1994;2:31–38. doi: 10.1016/0926-6410(94)90018-3. [DOI] [PubMed] [Google Scholar]
- Binder JR, Frost JA, Hammeke TA, Bellgowan PS, Springer JA, Kaufman JN, Possing ET. Human temporal lobe activation by speech and nonspeech sounds. Cereb Cortex. 2000;10:512–528. doi: 10.1093/cercor/10.5.512. [DOI] [PubMed] [Google Scholar]
- Boatman D. Cortical bases of speech perception: evidence from functional lesion studies. Cognition. 2004;92:47–65. doi: 10.1016/j.cognition.2003.09.010. [DOI] [PubMed] [Google Scholar]
- Boatman D, Hall C, Goldstein MH, Lesser R, Gordon B. Neuroperceptual differences in consonant and vowel discrimination: as revealed by direct cortical electrical interference. Cortex. 1997;33:83–98. doi: 10.1016/S0010-9452(97)80006-8. [DOI] [PubMed] [Google Scholar]
- Brenner N, Bialek W, de Ruyter van Steveninck R. Adaptive rescaling maximizes information transmission. Neuron. 2000;26:695–702. doi: 10.1016/S0896-6273(00)81205-2. [DOI] [PubMed] [Google Scholar]
- Calabrese A, Schumacher JW, Schneider DM, Paninski L, Woolley SM. A generalized linear model for estimating spectrotemporal receptive fields from responses to natural sounds. PLoS One. 2011;6:e16104. doi: 10.1371/journal.pone.0016104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carlson NL, Ming VL, Deweese MR. Sparse codes for speech predict spectrotemporal receptive fields in the inferior colliculus. PLoS Comput Biol. 2012;8:e1002594. doi: 10.1371/journal.pcbi.1002594. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chang EF, Edwards E, Nagarajan SS, Fogelson N, Dalal SS, Canolty RT, Kirsch HE, Barbaro NM, Knight RT. Cortical spatio-temporal dynamics underlying phonological target detection in humans. J Cogn Neurosci. 2011;23:1437–1446. doi: 10.1162/jocn.2010.21466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Crone NE, Boatman D, Gordon B, Hao L. Induced electrocorticographic gamma activity during auditory perception. Brazier Award-winning article, 2001. Clin Neurophys. 2001;112:565–582. doi: 10.1016/S1388-2457(00)00545-9. [DOI] [PubMed] [Google Scholar]
- Dan Y, Atick JJ, Reid RC. Efficient coding of natural scenes in the lateral geniculate nucleus: experimental test of a computational theory. J Neurosci. 1996;16:3351–3362. doi: 10.1523/JNEUROSCI.16-10-03351.1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
- David SV, Vinje WE, Gallant JL. Natural stimulus statistics alter the receptive field structure of v1 neurons. J Neurosci. 2004;24:6991–7006. doi: 10.1523/JNEUROSCI.1422-04.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- David SV, Mesgarani N, Shamma SA. Estimating sparse spectro-temporal receptive fields with natural stimuli. Network. 2007;18:191–212. doi: 10.1080/09548980701609235. [DOI] [PubMed] [Google Scholar]
- Depireux DA, Simon JZ, Klein DJ, Shamma SA. Spectro-temporal response field characterization with dynamic ripples in ferret primary auditory cortex. J Neurophysiol. 2001;85:1220–1234. doi: 10.1152/jn.2001.85.3.1220. [DOI] [PubMed] [Google Scholar]
- Dong D, Atick J. Temporal decorrelation: a theory of lagged and nonlagged responses in the lateral geniculate nucleus. Netw Comput Neural Syst. 1995;6:159–178. doi: 10.1088/0954-898X_6_2_003. [DOI] [Google Scholar]
- Edwards E, Chang EF. Syllabic (∼2–5 Hz) and fluctuation (∼1–10 Hz) ranges in speech and auditory processing. Hear Res. 2013;305:113–134. doi: 10.1016/j.heares.2013.08.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eggermont JJ. Temporal modulation transfer functions in cat primary auditory cortex: separating stimulus effects from neural mechanisms. J Neurophysiol. 2002;87:305–321. doi: 10.1152/jn.00490.2001. [DOI] [PubMed] [Google Scholar]
- Elliott TM, Theunissen FE. The modulation transfer function for speech intelligibility. PLoS Comput Biol. 2009;5:e1000302. doi: 10.1371/journal.pcbi.1000302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Escabi MA, Schreiner CE. Nonlinear spectrotemporal sound analysis by neurons in the auditory midbrain. J Neurosci. 2002;22:4114–4131. doi: 10.1523/JNEUROSCI.22-10-04114.2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fairhall AL, Lewen GD, Bialek W, de Ruyter Van Steveninck RR. Efficiency and ambiguity in an adaptive neural code. Nature. 2001;412:787–792. doi: 10.1038/35090500. [DOI] [PubMed] [Google Scholar]
- Felsen G, Touryan J, Han F, Dan Y. Cortical sensitivity to visual features in natural scenes. PLoS Biol. 2005;3:e342. doi: 10.1371/journal.pbio.0030342. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Garofolo J, Lamel LF, Fisher WM, Fiscus J, Pallett D, Dahlgren N, Zue V. TIMIT acoustic-phonetic continuous speech corpus. Philadelphia: Linguistic Data Consortium; 1993. [Google Scholar]
- Gross J, Hoogenboom N, Thut G, Schyns P, Panzeri S, Belin P, Garrod S. Speech rhythms and multiplexed oscillatory sensory coding in the human brain. PLoS Biol. 2013;11:e1001752. doi: 10.1371/journal.pbio.1001752. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Harms MP, Melcher JR. Sound repetition rate in the human auditory pathway: representations in the waveshape and amplitude of fMRI activation. J Neurophysiol. 2002;88:1433–1450. doi: 10.1152/jn.2002.88.3.1433. [DOI] [PubMed] [Google Scholar]
- Herdener M, Esposito F, Scheffler K, Schneider P, Logothetis NK, Uludag K, Kayser C. Spatial representations of temporal and spectral sound cues in human auditory cortex. Cortex. 2013;49:2822–2833. doi: 10.1016/j.cortex.2013.04.003. [DOI] [PubMed] [Google Scholar]
- Hickok G, Poeppel D. The cortical organization of speech processing. Nat Rev Neurosci. 2007;8:393–402. doi: 10.1038/nrn2113. [DOI] [PubMed] [Google Scholar]
- Hsu A, Woolley SM, Fremouw TE, Theunissen FE. Modulation power and phase spectrum of natural sounds enhance neural encoding performed by single auditory neurons. J Neurosci. 2004;24:9201–9211. doi: 10.1523/JNEUROSCI.2449-04.2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kim G, Doupe A. Organized representation of spectrotemporal features in songbird auditory forebrain. J Neurosci. 2011;31:16977–16990. doi: 10.1523/JNEUROSCI.2003-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Klein DJ, Depireux DA, Simon JZ, Shamma SA. Robust spectrotemporal reverse correlation for the auditory system: optimizing stimulus design. J Comput Neurosci. 2000;9:85–111. doi: 10.1023/A:1008990412183. [DOI] [PubMed] [Google Scholar]
- Klein DJ, Simon JZ, Depireux DA, Shamma SA. Stimulus-invariant processing and spectrotemporal reverse correlation in primary auditory cortex. J Comput Neurosci. 2006;20:111–136. doi: 10.1007/s10827-005-3589-4. [DOI] [PubMed] [Google Scholar]
- Langers DR, Backes WH, van Dijk P. Spectrotemporal features of the auditory cortex: the activation in response to dynamic ripples. Neuroimage. 2003;20:265–275. doi: 10.1016/S1053-8119(03)00258-1. [DOI] [PubMed] [Google Scholar]
- Liégeois-Chauvel C, Lorenzi C, Trébuchon A, Régis J, Chauvel P. Temporal envelope processing in the human left and right auditory cortices. Cereb Cortex. 2004;14:731–740. doi: 10.1093/cercor/bhh033. [DOI] [PubMed] [Google Scholar]
- Malone BJ, Scott BH, Semple MN. Dynamic amplitude coding in the auditory cortex of awake rhesus macaques. J Neurophysiol. 2007;98:1451–1474. doi: 10.1152/jn.01203.2006. [DOI] [PubMed] [Google Scholar]
- Malone BJ, Beitel RE, Vollmer M, Heiser MA, Schreiner CE. Spectral context affects temporal processing in awake auditory cortex. J Neurosci. 2013;33:9431–9450. doi: 10.1523/JNEUROSCI.3073-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McDermott JH, Simoncelli EP. Sound texture perception via statistics of the auditory periphery: evidence from sound synthesis. Neuron. 2011;71:926–940. doi: 10.1016/j.neuron.2011.06.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miller LM, Escabí MA, Read HL, Schreiner CE. Spectrotemporal receptive fields in the lemniscal auditory thalamus and cortex. J Neurophysiol. 2002;87:516–527. doi: 10.1152/jn.00395.2001. [DOI] [PubMed] [Google Scholar]
- Moerel M, De Martino F, Formisano E. Processing of natural sounds in human auditory cortex: tonotopy, spectral tuning, and relation to voice sensitivity. J Neurosci. 2012;32:14205–14216. doi: 10.1523/JNEUROSCI.1388-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nagel KI, Doupe AJ. Organizing principles of spectro-temporal encoding in the avian primary auditory area field L. Neuron. 2008;58:938–955. doi: 10.1016/j.neuron.2008.04.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nourski KV, Steinschneider M, Oya H, Kawasaki H, Jones RD, Howard MA. Spectral organization of the human lateral superior temporal gyrus revealed by intracranial recordings. Cereb Cortex. 2014;24:340–352. doi: 10.1093/cercor/bhs314. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Olshausen BA, Field DJ. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature. 1996;381:607–609. doi: 10.1038/381607a0. [DOI] [PubMed] [Google Scholar]
- Overath T, Zhang Y, Sanes DH, Poeppel D. Sensitivity to temporal modulation rate and spectral bandwidth in the human auditory system: fMRI evidence. J Neurophysiol. 2012;107:2042–2056. doi: 10.1152/jn.00308.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Overath T, McDermott JH, Zarate JM, Poeppel D. The cortical analysis of speech-specific temporal structure revealed by responses to sound quilts. Nat Neurosci. 2015;18:903–911. doi: 10.1038/nn.4021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paninski L. Maximum likelihood estimation of cascade point-process neural encoding models. Network. 2004;15:243–262. doi: 10.1088/0954-898X_15_4_002. [DOI] [PubMed] [Google Scholar]
- Pasley BN, David SV, Mesgarani N, Flinker A, Shamma SA, Crone NE, Knight RT, Chang EF. Reconstructing speech from human auditory cortex. PLoS Biol. 2012;10:e1001251. doi: 10.1371/journal.pbio.1001251. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rauschecker JP, Scott SK. Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat Neurosci. 2009;12:718–724. doi: 10.1038/nn.2331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ray S, Maunsell JH. Different origins of gamma rhythm and high-gamma activity in macaque visual cortex. PLoS Biol. 2011;9:e1000610. doi: 10.1371/journal.pbio.1000610. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rieke F, Bodnar DA, Bialek W. Naturalistic stimuli increase the rate and efficiency of information transmission by primary auditory afferents. Proc Biol Sci. 1995;262:259–265. doi: 10.1098/rspb.1995.0204. [DOI] [PubMed] [Google Scholar]
- Rinne T, Pekkola J, Degerman A, Autti T, Jääskeläinen IP, Sams M, Alho K. Modulation of auditory cortex activation by sound presentation rate and attention. Hum Brain Mapp. 2005;26:94–99. doi: 10.1002/hbm.20123. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rodríguez FA, Read HL, Escabí MA. Spectral and temporal modulation tradeoff in the inferior colliculus. J Neurophysiol. 2010;103:887–903. doi: 10.1152/jn.00813.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Santoro R, Moerel M, De Martino F, Goebel R, Ugurbil K, Yacoub E, Formisano E. Encoding of natural sounds at multiple spectral and temporal resolutions in the human auditory cortex. PLoS Comp Biol. 2014;10:e1003412. doi: 10.1371/journal.pcbi.1003412. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schönwiesner M, Zatorre RJ. Spectro-temporal modulation transfer function of single voxels in the human auditory cortex measured with high-resolution fMRI. Proc Nat Acad Sci USA. 2009;106:14611–14616. doi: 10.1073/pnas.0907682106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schreiner CE, Langner G. Periodicity coding in the inferior colliculus of the cat. II. Topographical organization. J Neurophysiol. 1988;60:1823–1840. doi: 10.1152/jn.1988.60.6.1823. [DOI] [PubMed] [Google Scholar]
- Sharpee T, Rust NC, Bialek W. Analyzing neural responses to natural signals: maximally informative dimensions. Neural Comput. 2004;16:223–250. doi: 10.1162/089976604322742010. [DOI] [PubMed] [Google Scholar]
- Sharpee TO, Sugihara H, Kurgansky AV, Rebrik SP, Stryker MP, Miller KD. Adaptive filtering enhances information transmission in visual cortex. Nature. 2006;439:936–942. doi: 10.1038/nature04519. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Smirnakis SM, Berry MJ, Warland DK, Bialek W, Meister M. Adaptation of retinal processing to image contrast and spatial scale. Nature. 1997;386:69–73. doi: 10.1038/386069a0. [DOI] [PubMed] [Google Scholar]
- Steinschneider M, Fishman YI, Arezzo JC. Spectrotemporal analysis of evoked and induced electroencephalographic responses in primary auditory cortex (A1) of the awake monkey. Cereb Cortex. 2008;18:610–625. doi: 10.1093/cercor/bhm094. [DOI] [PubMed] [Google Scholar]
- Striem-Amit E, Hertz U, Amedi A. Extensive cochleotopic mapping of human auditory cortical fields obtained with phase-encoding FMRI. PloS One. 2011;6:e17832. doi: 10.1371/journal.pone.0017832. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Talavage TM, Sereno MI, Melcher JR, Ledden PJ, Rosen BR, Dale AM. Tonotopic organization in human auditory cortex revealed by progressions of frequency sensitivity. J Neurophysiol. 2004;91:1282–1296. doi: 10.1152/jn.01125.2002. [DOI] [PubMed] [Google Scholar]
- Talebi V, Baker CL., Jr Natural versus synthetic stimuli for estimating receptive field models: a comparison of predictive robustness. J Neurosci. 2012;32:1560–1576. doi: 10.1523/JNEUROSCI.4661-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tanaka H, Fujita N, Watanabe Y, Hirabuki N, Takanashi M, Oshiro Y, Nakamura H. Effects of stimulus rate on the auditory cortex using fMRI with “sparse” temporal sampling. Neuroreport. 2000;11:2045–2049. doi: 10.1097/00001756-200006260-00047. [DOI] [PubMed] [Google Scholar]
- Theunissen FE, Sen K, Doupe AJ. Spectral-temporal receptive fields of nonlinear auditory neurons obtained using natural sounds. J Neurosci. 2000;20:2315–2331. doi: 10.1523/JNEUROSCI.20-06-02315.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Theunissen FE, David SV, Singh NC, Hsu A, Vinje WE, Gallant JL. Estimating spatio-temporal receptive fields of auditory and visual neurons from their responses to natural stimuli. Network. 2001;12:289–316. doi: 10.1080/net.12.3.289.316. [DOI] [PubMed] [Google Scholar]
- Wang Y, Ding N, Ahmar N, Xiang J, Poeppel D, Simon JZ. Sensitivity to temporal modulation rate and spectral bandwidth in the human auditory system: MEG evidence. J Neurophysiol. 2012;107:2033–2041. doi: 10.1152/jn.00310.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wilkinson F, James TW, Wilson HR, Gati JS, Menon RS, Goodale MA. An fMRI study of the selective activation of human extrastriate form vision areas by radial and concentric gratings. Curr Biol. 2000;10:1455–1458. doi: 10.1016/S0960-9822(00)00800-9. [DOI] [PubMed] [Google Scholar]
- Yang X, Wang K, Shamma SA. Auditory representations of acoustic signals. IEEE Trans Information Theory. 1992;38:824–839. doi: 10.1109/18.119739. [DOI] [Google Scholar]
- Zheng Y, Escabí MA. Distinct roles for onset and sustained activity in the neuronal code for temporal periodicity and acoustic envelope shape. J Neurosci. 2008;28:14230–14244. doi: 10.1523/JNEUROSCI.2882-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]