Abstract
Echoplanar MRI is associated with significant acoustic noise, which can interfere with the presentation of auditory stimuli, create a more challenging listening environment, and increase discomfort felt by participants. Here we investigate a scanning sequence that significantly reduces the amplitude of acoustic noise associated with echoplanar imaging (EPI). This is accomplished using a constant phase encoding gradient and a sinusoidal readout echo train to produce a narrow-band acoustic frequency spectrum, which is adapted to the scanner's frequency response function by choosing an optimum gradient switching frequency. To evaluate the effect of these nonstandard parameters we conducted a speech experiment comparing four different EPI sequences: Quiet, Sparse, Standard, and Matched Standard (using the same readout duration as Quiet). For each sequence participants listened to sentences and signal-correlated noise (SCN), which provides an unintelligible amplitude-matched control condition. We used BOLD sensitivity maps to quantify sensitivity loss caused by the longer EPI readout duration used in the Quiet and Matched Standard EPI sequences. We found that the Quiet sequence provided more robust activation for SCN in primary auditory areas and comparable activation in frontal and temporal regions for Sentences > SCN, but less sentence-related activity in inferotemporal cortex. The increased listening effort associated with the louder Standard sequence relative to the Quiet sequence resulted in increased activation in the left temporal and inferior parietal cortices. Together, these results suggest that the Quiet sequence is suitable, and perhaps preferable, for many auditory studies. However, its applicability depends on the specific brain regions of interest.
Keywords: Sparse imaging, Quiet EPI, BOLD sensitivity, Listening effort
Introduction
During echoplanar imaging (EPI), rapid switching causes the gradient coils to oscillate, resulting in the significant acoustic noise typically associated with fMRI scanning (Price et al., 2001; Ravicz et al., 2000). This acoustic noise, which can exceed 100 dBA, presents several serious challenges for studying the processing of auditory stimuli (Moelker and Pattynama, 2003). First, and perhaps most obviously, the high sound levels generated can render experimental stimuli unintelligible. Second, even if it is possible to hear the stimuli, listeners must separate them from the noise of the scanner, adding additional perceptual and cognitive demands to the task—that is, increases in listening effort (Davis and Johnsrude, 2003). Such task demands are likely to differentially affect participants with difficulties in auditory processing due to hearing impairment or normal aging (Grimault et al., 2001; Peelle and Wingfield, 2005; Wingfield et al., 2006). Finally, the acoustic noise of the scanner itself will activate auditory cortex (Bandettini et al., 1998), which may diminish effects induced by experimental manipulations (Elliott et al., 1999; Gaab et al., 2007). Thus, standard EPI sequences are sub-optimal for auditory tasks, and may make any results difficult to interpret. In addition, the reduction of acoustic noise may also be desirable for participant comfort, particularly when dealing with children or other special populations.
One common solution to the challenge posed by acoustic scanner noise is to use a sparse imaging procedure in which the repetition time (TR) of a sequence is longer than its acquisition time (TA), clustering slice acquisition in time in order to provide a silent period between the acquisition of consecutive volumes (Edmister et al., 1999; Hall et al., 1999; Scheffler et al., 1998). Auditory stimuli can then be presented during these silent periods without disruption from echoplanar scanner noise; the delay in the hemodynamic response to a stimulus enables the BOLD signal changes associated with these stimuli to be measured by the next volume of data acquired. Because of the longer TR, for a constant amount of scanning time, fewer images are acquired in a sparse imaging paradigm than in a continuous paradigm. This approach therefore reduces the temporal resolution of the data and, due to there being fewer observations, potentially reduces the accuracy of the parameter estimates (although this may be offset by higher overall levels of signal due to the absence of spin history effects). Some modifications to sparse imaging paradigms have been developed to compensate for these shortfalls by collecting multiple volumes following a silent period (e.g., Schwarzbauer et al., 2006). Nonetheless, for a given amount of scanning time, sparse imaging approaches are fundamentally limited in the number of volumes that can be acquired relative to continuous sequences, and the extent to which differently-timed responses to a stimulus can be measured.
Another important consideration is the fact that responses in auditory regions are not only influenced by the amplitude of the scanner noise, but by other parameters, such as its perceived continuity: auditory cortex responds strongly to pulsed noises in the frequency ranges associated with typical gradient switching, and thus typical EPI sequences make for particularly effective stimulation (Giraud et al., 2000; Harms and Melcher, 2002; Seifritz et al., 2003; Tanaka et al., 2000). Thus, a second approach for addressing some of the issues faced in auditory fMRI studies is to change the qualitative nature of the acoustic noise. This approach was taken by Seifritz et al. (2006), who developed an EPI sequence that emits continuous noise (rather than pulsed) by implementing a quasi-continuous gradient switching pattern. The authors presented audio recordings of the noise generated by both types of sequences to participants and recorded neural responses using a sparse imaging paradigm. They found that conventional EPI produced stronger responses than the continuous noise EPI. Additionally, responses to pure tones in auditory cortex were greater when measured with continuous noise EPI relative to conventional EPI. These results emphasize that the nature or quality of acoustic stimulation from the scanner, and not just its average loudness, must be considered in auditory fMRI studies.
Although changing the acoustic characteristics of the scanner noise effectively boosts BOLD responses in auditory areas for some stimuli, it still leaves open the possibility of interference by energetic masking, and may also lead to extra challenges of listening effort, especially for more complex (e.g., linguistic) stimuli. One way to mitigate these effects is to use active noise control to minimize the effects of scanner noise (Chambers et al., 2007; Hall et al., 2009). Here we adopt an alternate approach to reducing the impact of acoustic noise by using an EPI sequence that is sufficiently quiet to allow participants to easily perceive auditory stimuli, even in the presence of pulsed scanner noise (Schmitter et al., 2008). In this sequence, acoustic noise is minimized by using a constant phase encoding gradient and a sinusoidal readout echo train to produce a narrow-band acoustic frequency spectrum. The scanner-specific frequency response function can be measured using an MR-compatible microphone placed inside the magnet bore. It is then possible to choose a readout gradient switching frequency (within the limits imposed by BOLD fMRI) that results in a lower acoustic response based on this frequency response function. In addition, the clicking noise of the slice-selection gradient is reduced by choosing a lower slew rate.
This modified gradient switching scheme can reduce the acoustic noise of EPI by up to 20–30 dB compared to trapezoidal EPI using the same imaging parameters. However, it may also influence data quality. For example, the longer EPI readout duration required by using a slower gradient switching frequency would be expected to exacerbate susceptibility effects near tissue boundaries, such as in inferior temporal and orbital frontal regions (Devlin et al., 2000; Ojemann et al., 1997). In addition, the nonuniform sampling of k-space requires an adaptation of standard image reconstruction software, and because of the sinusoidal readout gradient, the resulting images are also smoother than those from a standard sequence.
The primary aim of the current study is to evaluate the data provided by this new sequence relative to existing EPI sequences. We chose to do so using auditory stimuli that result in robust and replicable patterns of activation in well-known regions of cortex based on several previous studies (Davis et al., 2007; Mummery et al., 1999; Rodd et al., 2005).
Method
Participants
Six healthy right-handed adults aged 20–26 years (3 females) participated in this study. All were native English speakers with self-reported normal hearing and no history of neurological problems. Written consent was obtained from all participants on a protocol approved by the local ethics committee.
Materials
The materials consisted of a set of unambiguous sentences created for Rodd et al. (2005). Sentences ranged in duration from 1.14 to 3.58 s and contained simple declarative statements (e.g., “The police returned to the museum.”). The 120 sentences used were divided into 4 groups of 30, matched for duration, naturalness, imageability, and number of words using Match software (Van Casteren and Davis 2007), available from http://www.mrc-cbu.cam.ac.uk/people/maarten.van-casteren/mixandmatch.html. For each sentence, a probe word was generated for use in a behavioral task. Half of these probe words were semantically related to the sentence, and half were unrelated.
For a baseline condition, sentences matched in duration to the experimental sentences were used to create signal-correlated noise (SCN) (Schroeder, 1968). These stimuli have the same overall amplitude envelope and spectral profile as the original sentences but are lacking spectral detail, and are therefore entirely unintelligible. They thus provide a good control for the acoustic stimulation and temporal pattern of the sentences without conveying any linguistic information.
Procedure
Participants were instructed to attend to each stimulus, and after each sentence respond to the probe word on the screen. In the case of the real sentences, the probe word had a 50% probability of being semantically related to the just-heard sentence; participants indicated whether or not the probe word was semantically related to the sentence using a button-press response. For example, for the sentence “There were beer and cider on the kitchen shelf,” a related probe word might have been “drink”. For the SCN trials, the word “left” or “right” appeared on the screen, and participants were instructed to press the appropriate button. Participants were situated in the magnet and familiarized with this procedure through a short practice session, during which we also ensured the sentences were being presented at a comfortable listening level. Participants were informed that they would be performing the same task four times, but that due to different imaging parameters the scanner noise would differ. In all cases they were instructed to ignore the scanner noise as much as possible and concentrate on listening to the auditory stimuli.
Image acquisition
All images were acquired on a Siemens 3 T Tim Trio scanner (Siemens Medical Systems, Erlangen, Germany). The four EPI sequences used to collect data on the experimental paradigm are listed in Table 1 and described below. In addition to EPI data, we cquired a T1-weighted structural image for each participant using an MPRAGE sequence (TR = 2250 ms, TE = 2.99 ms, TI = 900 ms, flip angle = 9°, FOV = 256 mm × 240 mm × 160 mm, voxel size = 1 mm × 1 mm × 1 mm). In addition, field map data (2D structural and phase difference images) were acquired using a standard double echo GE sequence (TE1/TE2 = 5.19/7.65 ms; TR = 400 ms; flip angle = 60°, slice thickness = 3 mm; matrix size = 64 × 64; in-plane resolution = 3 × 3 mm; total acquisition time = 54 s). The phase difference images were unwrapped and converted into magnetic field maps (Jenkinson, 2003).
Table 1.
Sequence | TR (s) | TA (s) | TE (ms) | Bandwidth (Hz/Px) | # of scans | dBA |
---|---|---|---|---|---|---|
Standard | 2.8 | 2.8 | 30 | 2230 | 284 | 90.2 |
Sparse | 11.2 | 2.8 | 30 | 2230 | 71 | n/a |
Matched Standard | 2.8 | 2.8 | 44 | 1220 | 284 | n/a |
Quiet | 2.8 | 2.8 | 44 | 1220 | 284 | 67.4 |
Note: Sound levels were measured for the Standard and Quiet sequences with an MR-compatible microphone placed inside an empty scanner. Measurements were not taken for the Matched Standard sequence but it was uniformly judged to be louder than the Quiet sequence, and softer than the Standard sequence. During the Sparse sequence the only noise came from ambient scanner noise (e.g., helium pumps).
Standard EPI sequence
We acquired 32 slices in axial oblique orientation (i.e., angled to ensure the eyeballs were not on the same plane as auditory cortex to avoid possible Nyquist ghost artifacts), collected in a descending order, with a .75 mm gap between slices (TR = 2.8 s, TA = 2.8 s, TE = 30 ms, flip angle = 90°, FOV = 192 mm × 192 mm, matrix = 64 × 64 mm, voxel size = 3 × 3 × 3 mm, bandwidth = 2230 Hz/Px). We ensured full temporal lobe coverage in all participants, but did not cover portions of superior parietal lobe in participants with larger heads. Unless noted below these parameters were consistent across all sequences tested.
Sparse EPI sequence
The Sparse sequence used a sparse or clustered acquisition sequence commonly employed in fMRI studies of auditory processing. This approach relies on the fact that, due to the lag inherent in the hemodynamic response, the peak BOLD response to a stimulus occurs several seconds after the stimulus, and can thus be captured by a scan occurring later in time. The imaging parameters were identical to the standard sequence, with the exception that we used a TR of 11.2 s. Given the TA of 2.8 s, this left 8.4 s of silence in which we presented a single sentence or SCN, allowing the item to be presented in the absence of echoplanar scanner noise. The presentation of the sentences and SCN was timed so that the midpoint of each stimulus occurred 5 s before the midpoint of data acquisition in order to maximize detection of the peak hemodynamic response.
Quiet EPI sequence
This sequence is described in more detail in Schmitter et al. (2008). The principle behind this sequence is that because the acoustic response of gradient coils changes as a function of the gradient coil switching frequency, changing the gradient switching frequency to one with a low acoustic response can significantly reduce acoustic noise. The quiet sequence therefore involved gradient switching g(t) which is characterized by a narrow-band input spectrum G(f), located in a frequency interval where the frequency response function showed a local minimum, judged using the individual frequency response profile measured for the empty scanner using an MR-safe optical microphone (Sennheiser MO 2000, Sennheiser GmbH & Co. KG, Wedemark, Germany). Based on these measurements a readout bandwidth of 1220 Hz/Px was chosen, along with a TE of 44 ms. Finally, the quiet sequence also used sinusoidally switched readout gradients, with the phase encoding gradient switched constant. This combination results in an S-shaped trajectory through k-space, which was regridded using a custom software implementation. The sinusoidal readout gradient leads to smoother image reconstruction in the readout direction. Note that because the regridding procedure is identical for all image volumes, it does not result in additional signal variance over time. Sound levels measured inside the scanner indicated a considerable sound level reduction of approximately 20 dBA compared to the Standard sequence. Finally, as noted previously, the subjective quality of the scanner noise can influence the measured activity in auditory cortex. In our study, the perceived continuity of the Quiet sequence was similar to that of the Standard and Matched Standard (see below) sequences; that is, the reduction in overall acoustic amplitude would be expected to reduce the energetic masking, but responses that relate to the nature of the background noise (Seifritz et al., 2006) are unlikely to be significantly affected. This is an advantage in that, having equated the pulsatile nature of the scanning sequence across sequences, we can more confidently attribute changes in activity to the differences in acoustic noise levels. At the same time, there is a possibility that an EPI sequence that was not only acoustically quiet but also continuous would afford even greater sensitivity.
Matched Standard EPI sequence
There are two significant differences between the Standard and Quiet sequences: (1) the Quiet sequence has a longer TE, and a narrower bandwidth (1220 Hz/Px as opposed to 2230 Hz/Px); and (2) the Quiet sequence uses a different type of readout gradient, requiring custom reconstruction. Thus, if differences were found between the Quiet and Standard sequences, it would be difficult to know to which parameter we could attribute the difference. To address this issue we also collected data using the Matched Standard sequence, which is identical to the Standard sequence, with the exception that the duration of the EPI readout was matched to the Quiet sequence by choosing the same bandwidth of 1220 Hz/Px; thus, the frequency of the readout gradient was identical to that of the Quiet sequence.
Signal-to-noise ratio measurements
To estimate the temporal signal-to-noise ratio (SNR) for each sequence, 6 white matter voxels (MNI coordinates: [−27 −18 32], [−22 39 4], and [−20 18 26] in the left hemisphere, along with their right hemisphere counterparts) were selected from which to sample the signal. Each voxel had at least a 97% probability of being white matter according to the tissue probability maps distributed with SPM, and their location in white matter was verified on the structural images for individual participants in the current study. For each sequence, the data from each voxel were extracted from the unsmoothed normalized images, and baseline corrected using a 5th order cosine basis set to remove low-frequency drifts. The SNR was then calculated by dividing the mean of the timeseries by its standard deviation (Krüger and Glover, 2001). For each sequence the SNR was calculated for each of the 6 voxels for each participant, and then the SNR values were averaged to provide a single SNR value for each sequence for each participant.
BOLD sensitivity maps
BOLD sensitivity maps were calculated from the field maps and the specific acquisition parameters of the Standard and Quiet sequences using the theoretical framework described by De Panfilis and Schwarzbauer (2005) with the extension proposed by Weiskopf et al. (2007) to account for the effect of susceptibility gradients in readout direction. Because the Standard and Sparse sequences differ only in TR, the BOLD sensitivity maps are identical. In addition, the estimated sensitivity maps for the Quiet and Matched Standard sequences can be considered identical, as the relevant imaging parameters were identical and the effect of the nonuniform k-space sampling used in the Quiet sequence is negligible.
fMRI analysis
Image preprocessing and statistical analyses were performed using SPM5 (Wellcome Trust Centre for Neuroimaging, London, UK). The first 9 images for the three continuous sequences and the first 3 images for the sparse sequence were discarded to allow for T2 saturation effects. Low-frequency drifts were removed with highpass filtering (with a cutoff period of 90 s) and autocorrelations were modeled using a first-order autoregressive model. Images for each participant were realigned to the first image in the series (Friston et al., 1995), corrected for effects of magnetic field inhomogeneity (Cusack et al., 2003) and coregistered with the structural image (Ashburner and Friston, 1997). The transformation required to bring a participant's images into standard MNI152 space was calculated using MNI-space tissue probability maps (2 × 2 × 2 mm) included with SPM5 (Ashburner and Friston, 2005), and these warping parameters were then applied to all functional images for that participant. During normalization the data were interpolated to 2 mm isotropic voxels using trilinear interpolation. Prior to statistical analysis the data were spatially smoothed with a 10 mm FWHM isotropic Gaussian kernel.
Data were initially analyzed separately for each participant, using a separate general linear model for each scanning sequence. Each event of interest (sentence or SCN) was convolved with a standard hemodynamic response function to create the repressors used in the model. This was specified as an event with no duration occurring at the middle of each stimulus. The 6 motion parameters obtained during realignment were included in the model as additional regressors. Following analysis on the subject level, images containing the contrasts of parameter estimates for each subject were entered into second-level group analyses. To ensure fair comparison for results across sequences despite differences in sensitivity and signal dropout, we used the same explicit brain mask (distributed with SPM) for all analyses. We did not use a proportional threshold for the fMRI data but analyzed all voxels included in the mask.
For comparisons among the four sequences we examined both parameter estimates and t-statistics derived from the first-level (single-subject) analyses. We examined t-statistics in addition to parameter estimates because they give an indication as to the effect size relative to the variability present (which we expected might differ by sequence). For example, due to the lack of spin history effects, images acquired with a sparse sequence tend to exhibit higher overall levels of signal, but also more variability due to the reduced number of observations.
Images are displayed on an MNI-space template brain included with MRIcron (Rorden and Brett, 2000), available from http://www.cabiatl.com/mricro/mricron/. Coordinates listed for maxima are in MNI152 space.
Results
Behavioral performance
We first examined accuracy on the behavioral task in all four scanning sequences, which was ≥ 93% correct in all cases. Accuracy did not differ as a function of scanning sequence, evidenced by a Friedman nonparametric test [χ2(3) = 2.32, n.s.], indicating that participants were able to accurately perceive the speech stimuli present during each scanning sequence. However, we note that equal intelligibility can still be achieved through different amounts of listening effort, which we expect to be the case when auditory stimuli are presented in the context of background noise.
Signal-to-noise ratio
The sinusoidal readout gradient used in the Quiet sequence leads to an increase of the effective voxel size in readout direction, which translates into a linear increase in the SNR. Therefore, we first estimated the expected change in SNR based on this difference, relative to the Matched Standard sequence, which in all other ways is identical. The relative SNR gain, compared to the trapezoidal gradient switching schemes used in the other sequences, can be calculated from the areas under the respective gradient waveforms to give , where Ts is the gradient ramp time and Tg the total duration of a trapezoidal gradient lobe. Compared to a Ts/Tg = .15 (Matched Standard sequence), the expected SNR gain of the Quiet sequence is 34%.
Signal-to-noise ratio (SNR) values calculated based on extracted data for each sequence are plotted in Fig. 1. As expected, due to increased intrinsic smoothness, the Quiet sequence showed significantly higher SNR values. To assess whether the differences in values were significant, we submitted these data to a within-subjects one-way analysis of variance (ANOVA) to test for a main effect of sequence, which was significant, F(3,15) = 52.6, p < .001, MSE = 701.9. We then conducted post-hoc t-tests between the Quiet sequence and all other sequences; these confirmed that, the Quiet sequence had a higher SNR than all of the other sequences, all ts > 8.9 and all ps < .001. The relative change in SNR between the Matched Standard and Quiet sequences was 27%, which is in reasonably good agreement with the theoretical gain described above. Although the increase in SNR was expected due to the increased smoothness inherent in the Quiet sequence, it should be noted that this SNR difference in the unsmoothed data becomes negligible in the smoothed data used in the fMRI analyses, as the increase in the effective voxel size is small compared to the dimension of the Gaussian smoothing kernel (10 mm FWHM).
BOLD sensitivity
We calculated BOLD sensitivity maps for each subject based on spatially normalized field maps, and then created group average maps, shown in Fig. 2a. This was done for the Standard and Quiet sequence (as noted previously, these are identical to those for the Sparse and Matched Standard, respectively). Although both sequences exhibit lower signal around the temporal lobes, this was most pronounced in the Quiet sequence. Shown in Fig. 2b is the comparison of the BOLD sensitivity maps across sequence. The absolute difference in sensitivity is shown in the top panel; there were no regions that showed more sensitivity for the quiet sequence, but several regions that showed less sensitivity (cool colors). To quantify this difference we performed a pairwise t-test to compare the two sets of BOLD sensitivity maps across participants. The maps were smoothed with a 10 mm FWHM isotropic Gaussian kernel prior to analysis to match the fMRI data. The results in the bottom portion of Fig. 2b, and listed in Table 2, show significant differences in sensitivity using a voxelwise threshold of p < .0001 (uncorrected) and a minimum cluster extent of 50 voxels which in all cases led to clusters which were whole-brain FWE-corrected for significance using cluster extent, p < .05 [FWHM = 14.7 mm × 12.9 mm × 11.5 mm, resel count = 1216.4]. Differences in BOLD sensitivity were widespread, and included inferior temporal and orbital frontal regions where the sensitivity was lowest. Whole-brain maps of BOLD sensitivity and differences are shown in Fig. S1–S4.
Table 2.
Region | # of voxels | Coordinates |
Z score | ||
---|---|---|---|---|---|
x | y | z | |||
Sinus | 2297 | 2 | 16 | −26 | 6.08 |
L orbital frontal cortex | −12 | 22 | −24 | 5.15 | |
L ventral frontal (eyeball) | −26 | 42 | −26 | 5.12 | |
Ventromedial frontal cortex | −6 | 24 | −24 | 5.06 | |
Orbital frontal cortex | −2 | 66 | −4 | 5.06 | |
R inferior temporal cortex | 1332 | 68 | −26 | −34 | 5.57 |
R fusiform gyrus | 36 | −22 | −26 | 5.44 | |
R medial inferior medial temporal | 18 | −26 | −32 | 5.29 | |
Supplemental motor area | 4766 | 4 | −2 | 66 | 5.30 |
R supplemental motor area | 12 | 10 | 74 | 4.98 | |
L supplemental motor area | −10 | −4 | 68 | 4.94 | |
R supplemental motor area | 10 | −10 | 62 | 4.64 | |
L supplemental motor area | −18 | −10 | 64 | 4.42 | |
Lateral ventricles | 2 | 0 | 26 | 4.38 | |
R orbital frontal cortex | 163 | 22 | 50 | −24 | 5.28 |
L posterior inferior temporal/fusiform | 1114 | −48 | −40 | −26 | 5.09 |
L anterior inferior temporal/fusiform | −46 | −12 | −30 | 4.90 | |
L posterior inferior temporal/fusiform | −50 | −30 | −20 | 4.58 | |
L supramarginal gyrus | 711 | −60 | −46 | 30 | 5.02 |
L inferior parietal lobe/angular gyrus | −50 | −62 | 44 | 5.01 | |
L superior parietal lobe | −30 | −64 | 50 | 4.69 | |
R posterior temporal lobe | 297 | 56 | −58 | 20 | 4.84 |
R inferior parietal sulcus | 87 | 24 | −70 | 50 | 4.18 |
L medial inferior temporal lobe | 120 | −22 | −26 | −38 | 4.07 |
L medial inferior temporal lobe | −18 | −18 | −30 | 4.04 | |
L occipital gyrus | 263 | −44 | −86 | 16 | 4.03 |
Anterior cingulate | 60 | 4 | 40 | 10 | 3.86 |
Note: When areas of significant difference occur outside the brain (i.e. near brain/air boundaries), these are described by the nearest cortical landmark.
Acoustic activity
We used the SPM anatomy toolbox (Eickhoff et al., 2005) to create a binary mask encompassing regions TE1.0 and TE1.1 of bilateral primary auditory cortex (Morosan et al., 2001), outlined in blue in Fig. 3a. Regions included in the mask had at least an 80% chance of belonging to these subdivisions based on cytoarchitectonic characteristics. We then conducted a within-subjects ANOVA to identify voxels in which the main effect of SCN relative to the mean over scans differed by sequence type, shown in Fig. 3a. We used a cluster-defining voxelwise threshold of p < .005 (uncorrected), and FWE-corrected for set-level significance within the search volume using cluster extent p < .05 (Worsley et al., 1992). For the peak voxel in each cluster, we then extracted the parameter estimates for each of the four sequences, shown in the top of Fig. 3a, to illustrate the nature of the difference between sequences. We also extracted the t-statistics from the first-level single-subject analysis, plotted along the bottom. These results indicate that in primary auditory regions the Quiet sequence gave the most robust response to SCN stimuli.
Speech activity
A second comparison of interest was the increase in activity when participants heard sentences compared to when they heard SCN, a contrast which results in robust language-related activation when data are collected using a sparse scanning sequence (Davis et al., 2007; Rodd et al., 2005; Schwarzbauer et al., 2006). Results for the Sentences > SCN contrast averaged across the four scanning sequences are shown in Fig. 3b and listed in Table 3, using a cluster-defining voxelwise threshold of p < .005 (uncorrected), and whole-brain FWE-corrected for significance using cluster extent, p < .05 [FWHM = 9.4 mm × 9.7 mm × 9.6 mm, resel count = 2653.5].
Table 3.
Region | # of voxels | Coordinates |
Z score | ||
---|---|---|---|---|---|
x | y | z | |||
L anterior temporal cortex | 5051 | −62 | 18 | −20 | 5.17 |
L posterior superior temporal gyrus | −68 | −22 | 10 | 4.94 | |
L inferior frontal gyrus pars opercularis and triangularis | −52 | 16 | 18 | 4.74 | |
L ventral inferior frontal gyrus | −58 | 32 | −6 | 4.24 | |
R middle occipital gyrus | 2657 | 44 | −84 | 28 | 4.26 |
Calcarine fissure | 6 | −96 | 16 | 4.26 | |
R posterior superior temporal gyrus | 1104 | 66 | −14 | 6 | 4.11 |
R superior temporal gyrus | 68 | −12 | −6 | 4.02 | |
R superior temporal gyrus | 56 | −26 | 6 | 3.07 | |
L precuneus | 483 | −6 | −50 | 10 | 3.68 |
R precuneus | 6 | −50 | 10 | 3.52 | |
R anterior superior temporal gyrus | 197 | 60 | 26 | −18 | 3.65 |
R anterior superior temporal gyrus | 54 | 18 | −18 | 2.98 | |
L middle occipital gyrus | 202 | −30 | −84 | 34 | 3.58 |
L precentral gyrus | 467 | −60 | 0 | 50 | 3.35 |
L precentral gyrus | −50 | 0 | 50 | 3.16 | |
L basal ganglia/thalamus | 158 | −16 | −16 | −12 | 3.34 |
−14 | −22 | −4 | 3.10 | ||
−22 | −6 | −4 | 2.85 |
To examine whether the Sentences > SCN contrast differed as a function of scanning sequence we extracted parameter estimates and first-level t-statistics for each participant from five of the peak voxels identified by the Sentences > SCN contrast, and subjected each to a one-way within-subjects ANOVA, Bonferroni-corrected to control for multiple comparisons within each dependent measure (i.e. separately for parameter estimates and t-statistics). For the parameter estimates, there were significant effects of sequence in the left posterior superior temporal gyrus (STG) [−68 −22 10: F(3,15) = 14.4, p < .005] and right posterior STG [68 −12 −6: F(3,15) = 16.2, p < .005]. None of the regions showed a significant effect of sequence on first-level t-statistics (all other ps > .05).
As noted previously, even sentences in which word report is equivalent can differ in the amount of effort required to achieve this level of intelligibility. To look for neural correlates of listening effort, we conducted a second-level paired-samples t-test comparing the Sentences > SCN parameter estimates for the Standard and Quiet sequences, under the assumption that the acoustically louder Standard sequence would be more effortful than the Quiet sequence, and thus result in increased neural activity. This comparison is shown in Fig. 4, with maxima listed in Table 4. We used a cluster-defining voxelwise threshold of p < .005 (uncorrected), and whole-brain FWE-corrected for significance using cluster extent, p ≤ .052 [FWHM = 8.9 mm × 9.0 mm × 8.9 mm, resel count = 3305.9]. This analysis revealed several significant peaks along left superior temporal cortex, as well as left inferior parietal cortex (extending into angular gyrus).
Table 4.
Region | # of voxels | Coordinates |
Z score | ||
---|---|---|---|---|---|
x | y | z | |||
L inferior parietal cortex | 339 | −36 | −74 | 44 | 3.97 |
L angular gyrus | −40 | −66 | 44 | 3.72 | |
L inferior parietal cortex | −48 | −60 | 48 | 3.40 | |
L posterior middle temporal gyrus | 209 | −56 | −46 | 8 | 3.71 |
L posterior middle temporal gyrus | −66 | −44 | 0 | 3.18 | |
L anterior superior temporal sulcus | 148 | −68 | −14 | 2 | 3.27 |
L anterior superior temporal sulcus | −68 | 2 | −8 | 3.21 | |
L anterior superior temporal sulcus | −60 | 4 | −14 | 2.78 |
Finally, to assess the qualitative effect of different sequences on the Sentences > SCN contrast at a group level, we performed group (random effects) analyses for each of the four sequences for this contrast. For each sequence the significant effect of Sentences > SCN is shown in Fig. 5a, with a cluster-defining voxelwise threshold of p < .005 (uncorrected), and whole-brain FWE-corrected for significance using cluster extent, p < .05. Fig. 5b shows the arbitrarily-thresholded subtraction of group t-statistics, demonstrating differences in group t-statistics for the Quiet sequence relative to the other three sequences. Of particular interest are the higher t-statistics near primary auditory areas (green circles) and lower t-statistics in inferior temporal regions (magenta circles) in the Quiet sequence relative to all other sequences. These qualitative observations are consistent with our quantitative comparisons, and support increased sensitivity of the Quiet sequence near primary auditory areas, and reduced sensitivity in inferior temporal regions.
Discussion
There are multiple challenges associated with using conventional echoplanar fMRI sequences to study neural responses to auditory stimuli due to acoustic scanner noise. Although for many years sparse imaging has proved to be a valuable approach, information regarding the timecourse of the BOLD response is lacking in these studies. An ideal sequence would be sufficiently quiet so as not to challenge auditory processing, and yet provide continuous information regarding brain activity. In the current report we set out to see whether the quiet EPI sequence introduced by Schmitter et al. (2008) might fulfill this role. Previously it has been shown that this sequence increases sensitivity in primary auditory areas in response to pure tone stimulation (Schmitter et al., 2008); in the current report we extended these results by including a greater number of comparison sequences, and by assessing the sensitivity of the Quiet sequence to both nonlinguistic auditory stimuli (i.e., SCN) and sentences. These are discussed in turn below.
Responses to SCN in primary auditory cortex
The neural response of auditory cortex to acoustic scanner noise has been a subject of interest for some time. It is clear that the considerable sound produced by echoplanar imaging produces robust responses in primary auditory cortex and surrounding areas (Bandettini et al., 1998; Hall et al., 2000; Talavage et al., 1999), and that this elevated response can reduce task-related activation (Elliott et al., 1999; Gaab et al., 2007; Haller et al., 2005). We hypothesized, therefore, that the Quiet sequence would show a greater response to basic task-related acoustic stimulation than the louder continuous sequences. This prediction was borne out: the Quiet sequence showed the largest response to SCN in cytoarchitectonically-defined primary auditory cortex.
Perhaps somewhat surprising is that the response was larger in these auditory regions in the Quiet sequence than the Sparse sequence, in which there is no scanner noise during stimulus presentation. One possible reason for this is that, for a given period of scanning time, the Quiet sequence collects more scans compared to the Sparse sequence. Even if all other factors were held constant, this substantial increase in degrees of freedom would result in a more reliable response measurement. Although these additional degrees of freedom are reflected in the single-subject t-statistics, they do not affect the parameter estimates, in which the Quiet sequence also appears to show a significant advantage. This may be due to variability in the timecourse of the BOLD response, either across participants (Aguirre et al., 1998) or across stimuli. That is, in a sparse imaging paradigm, the time of each stimulus relative to a scan is typically chosen so as to capture the BOLD response approximately 4–6 s after the event, and thus hopefully measure the peak response. However, because the timecourse of the response is not identical across stimuli or participants, it is possible for the single scan acquired in sparse imaging to miss the peak response. The continuous data acquisition afforded by the Quiet sequence may enhance the ability to detect the BOLD response from auditory regions. In principle, this may be especially true when longer stimuli, such as sentences, are used, as the estimation of when a discrete “event” occurs is not straightforward. However, if this was the case, we would expect the Quiet sequence to consistently outperform the Sparse sequence (i.e., in all conditions), as it should always be able to give a better estimation of the hemodynamic response. Because of the lack of differences in the Sentences > SCN comparison (discussed below), then, this seems to be an unlikely explanation.
A third possible reason why the Quiet sequence showed greater auditory activation than the Sparse sequence is that, although the Sparse sequence does not involve scanner noise concurrent with the stimuli, it is still relatively loud (∼ 90 dBA). This may induce other phasic changes in the response of auditory cortex which our TR of 11.2 s (which in the context of sparse imaging is fairly moderate) would not permit to completely return to baseline. This may be particularly relevant for sparse imaging as with this type of acoustic stimulation significant onset responses (i.e. edge detection) might be expected for each scan (Chait et al., 2007; Herdener et al., 2007).
Responses to sentences in frontal, temporal, and parietal cortices
To identify a sentence processing network, we first looked for the effect of Sentences > SCN averaged over scanning sequence. This analysis identified regions in frontal and temporal cortex that are in good agreement with a number of previous studies using the same comparison (Davis et al., 2007; Mummery et al., 1999; Rodd et al., 2005; Schwarzbauer et al., 2006). Within these regions we selected 5 maxima at which to examine the difference in responses between sequences. Only the left and right posterior STG regions showed a significant difference in parameter estimates among sequences, with the Sparse sequence providing the largest parameter estimates. All other regions showed statistically equivalent parameter estimates across sequence, which was also the case for all regions with the first-level t-statistics. These results suggest that, in contrast to what was observed in primary auditory areas, the response in these sentence processing regions is largely insensitive to the scanning sequence used. The higher parameter estimate for the Sparse sequence may be explained by the decay of spin history effects, associated with the pause between scans, which typically provide larger yet more variable signal. This conclusion is supported by the lack of a clear advantage for the Sparse sequence in the first-level t-statistics (in which scan-by-scan variance is taken into account).
One core sentence processing area not identified by our whole-brain analysis of Sentences > SCN is the left posterior inferior temporal gyrus and/or nearby fusiform gyrus (hereafter referred to together as inferotemporal cortex), a region that has been gaining in prominence in theories of speech processing (Hickok and Poeppel, 2007). In a voxel-based lesion-symptom mapping study, Bates et al. (2003) showed that damage to this region in stroke patients is associated with reduced auditory sentence comprehension. In neuroimaging studies activity in inferotemporal regions has been reported for spoken words > pseudowords (Binder et al., 2000; Orfanidou et al., 2006; see also the meta-analysis in Davis and Gaskell (2009)) and during comprehension of connected speech (Awad et al., 2007; Crinion et al., 2003). Perhaps most relevant for the current discussion, robust left inferotemporal activity in response to sentences has been observed in previous studies using materials similar to the ones in the current study (Davis et al., 2007; Rodd et al., 2005). As reviewed by Price (2000), activity in left inferotemporal cortex is common in word processing studies of both written and spoken language, including silent reading (words > false font), auditory and visual word retrieval, reading Braille, and several types of naming. The common activity across different tasks and modalities suggests a critical role in language processing. Based primarily on studies of single word processing, Price (2000) concluded that left inferotemporal cortex played a role in semantic processing. This is consistent with more recent studies of sentence processing in which sentences containing increased semantic ambiguity result in increased inferotemporal recruitment (Davis et al., 2007; Rodd et al., 2005). Together the available evidence makes a compelling case for left posterior inferotemporal cortex playing a critical role in semantic processing, and we thus expect its ubiquitous involvement in determining the meaning of connected speech.
The absence of inferotemporal activation in response to sentences in our current analysis is best explained by the significant signal dropout in this region in the Quiet and Matched Standard sequences caused by the longer duration of the EPI readout used in these sequences. In fact, the comparison of group results show that the Quiet sequence detected less sentence-related activity in this region than any of the other three sequences. Thus, the Quiet sequence appears to do as well as other sequences within most of these speech-processing regions, but is not sensitive to activations in inferotemporal cortex.
The lack of posterior inferotemporal activation due to susceptibility effects in our study highlights the difficulty of obtaining adequate signal in this region, even with standard EPI sequences, which may contribute to the lack of consensus regarding speech-related responses in posterior inferior temporal cortex in fMRI studies. Even with standard TEs (i.e., 30 ms for a 3 T scanner), inadequate active shimming may lead to substantial signal loss, which may go unnoticed. We reiterate the recommendation of Poldrack et al. (2008) that fMRI researchers check internally, and also report, average BOLD sensitivity and group coverage maps (e.g., the mask used in analysis) to provide an indication of the likelihood of detecting activations throughout the brain, particularly in regions prone to susceptibility artifact. An indication of the coverage provided by a mask (implicit, explicit, or a combination) is especially important if a proportional threshold is used in analysis. Given the increasing realization of the critical role played by inferior temporal regions in speech processing, this step would be extremely beneficial in interpreting the sometimes inconsistent fMRI results from this region.
Finally, we have suggested above that even if stimuli can be perceived in the presence of scanner noise, the additional effort required to achieve this perception will result in increased cognitive demand. For example, Davis and Johnsrude (2003) used three different acoustic manipulations to degrade sentence stimuli, and found that increased activity in large portions of left temporal, prefrontal and premotor cortices was associated with listening to the more degraded sentences. In the current study we hypothesized that due to increased energetic masking, additional neural activity would be associated with successful sentences comprehension during the Standard sequence relative to the Quiet sequence. Consistent with this prediction, we found significantly more speech-related activity in several areas of left temporal cortex, as well as left inferior parietal cortex. Although this finding must be interpreted with some caution due to the difference in scanning sequence, these areas did not show large differences in BOLD sensitivity, and thus we think it likely that these increases can be attributed to the different acoustic environments of the Standard and Quiet sequences. This finding underscores the importance of listening effort in interpreting data from auditory fMRI studies, and in particular the effects of standard EPI noise on required effort. Interestingly, contrary to Davis and Johnsrude (2003), we did not see significant increases in frontal regions associated with listening effort. This might be due to the relatively small sample size in the current study, or may instead reflect different processes required when acoustic degradation is intrinsic to the speech signal and hence originates from the same spatial location as the noise (Davis and Johnsrude, 2003) or spatially separate (as in the current study).
Conclusions
Relative to standard continuous EPI sequences or popular sparse imaging approaches, the Quiet scanning sequence we evaluated (Schmitter et al., 2008) provides increased sensitivity to acoustic response in primary auditory regions, and a comparable response to sentences within a fronto-temporal speech network. The most notable disadvantage of the Quiet sequence was an increase in signal dropout in inferior temporal cortex associated with its longer EPI readout duration, and the resulting lack of sensitivity to language-related activation in this region. As a future improvement, parallel imaging acquisition techniques could be used to reduce the duration of the EPI readout, which would considerably improve its sensitivity of the Quiet EPI in regions affected by macroscopic magnetic field inhomogeneities. Future work comparing the effect of sound level reductions (as in the current study) and changes in perceived continuity (Seifritz et al., 2006) or active noise cancelling (Hall et al., 2009) would also be of interest.
It is important to emphasize that various types of stimuli may interact with acoustic scanner noise differently; one example of this being the different effects we have reported for SCN and Sentences > SCN. Thus, care should be taken when generalizing our findings to novel stimuli. However, the current results suggest that technical developments in acoustically quiet echoplanar imaging have the potential to increase sensitivity to the BOLD response for a wide variety of auditory stimuli. These approaches may be most helpful when acoustic-related activity in and near primary auditory cortex is of primary interest. In addition, they may have applicability in non-auditory tasks in service of increasing participant comfort and removing potential task confounds, provided areas prone to susceptibility effects are not of interest.
Acknowledgments
This research was supported by the Medical Research Council (U.1055.04.013.01.01). We are grateful to Steve Eldridge and Helen Lloyd for their assistance with data collection, Stefan Hetzer for helpful discussions, and to our volunteers for their participation. Portions of this work were presented at the Annual Meeting of the Organization for Human Brain Mapping, June 2009.
Footnotes
Supplementary data associated with this article can be found, in the online version, at doi:10.1016/j.neuroimage.2010.05.015.
Appendix A. Supplementary data
References
- Aguirre G.K., Zarahn E., D'Esposito M. The variability of human, BOLD hemodynamic responses. NeuroImage. 1998;8:360–369. doi: 10.1006/nimg.1998.0369. [DOI] [PubMed] [Google Scholar]
- Ashburner J., Friston K.J. Multimodal image coregistration and partitioning — a unified framework. NeuroImage. 1997;6:209–217. doi: 10.1006/nimg.1997.0290. [DOI] [PubMed] [Google Scholar]
- Ashburner J., Friston K.J. Unified segmentation. NeuroImage. 2005;26:839–851. doi: 10.1016/j.neuroimage.2005.02.018. [DOI] [PubMed] [Google Scholar]
- Awad M., Warren J.E., Scott S.K., Turkheimer F.E., Wise R.J.S. A common system for the comprehension and production of narrative speech. J. Neurosci. 2007;27:11455–11464. doi: 10.1523/JNEUROSCI.5257-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bandettini P.A., Jesmanowicz A., Van Kylen J., Birn R.M., Hyde J.S. Functional MRI of brain activation induced by scanner acoustic noise. Magn. Reson. Med. 1998;39:410–416. doi: 10.1002/mrm.1910390311. [DOI] [PubMed] [Google Scholar]
- Bates E., Wilson S.M., Saygin A.P., Dick F., Sereno M.I., Knight R.T., Dronkers N.F. Voxel-based lesion-symptom mapping. Nat. Neurosci. 2003;6:448–450. doi: 10.1038/nn1050. [DOI] [PubMed] [Google Scholar]
- Binder J.R., Frost J.A., Hammeke T.A., Bellgowan P.S., Springer J.A., Kaufman J.N., Possing E.T. Human temporal lobe activation by speech and nonspeech sounds. Cereb. Cortex. 2000;10:512–528. doi: 10.1093/cercor/10.5.512. [DOI] [PubMed] [Google Scholar]
- Chait M., Poeppel D., de Cheveigné A., Simon J.Z. Processing asymmetry of transitions between order and disorder in human auditory cortex. J. Neurosci. 2007;27:5207–5214. doi: 10.1523/JNEUROSCI.0318-07.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chambers J., Bullock D., Kahana Y., Kots A., Palmer A. Developments in active noise control sound system for magnetic resonance imaging. Appl Acoust. 2007;68:281–295. [Google Scholar]
- Crinion J.T., Lambon Ralph M.A., Warburton E.A, Howard D., Wise R.J.S, Hyde J.S. Temporal lobe regions engaged during normal speech comprehension. Brain. 2003;126:1193–1201. doi: 10.1093/brain/awg104. [DOI] [PubMed] [Google Scholar]
- Cusack R., Brett M., Osswald K. An evaluation of the use of magnetic field maps to undistort echo-planar images. NeuroImage. 2003;18:127–142. doi: 10.1006/nimg.2002.1281. [DOI] [PubMed] [Google Scholar]
- Davis M.H., Gaskell M.G. A complementary systems account of word learning: neural and behavioural evidence. Philos. Roy. Soc. B. 2009;364:3773–3800. doi: 10.1098/rstb.2009.0111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Davis M.H., Johnsrude I.S. Hierarchical processing in spoken language comprehension. J. Neurosci. 2003;23:3423–3431. doi: 10.1523/JNEUROSCI.23-08-03423.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Davis M.H., Coleman M.R., Absalom A.R., Rodd J.M., Johnsrude I.S., Matta B.F., Owen A.M., Menon D.K. Dissociating speech perception and comprehension at reduced levels of awareness. Proc. Natl. Acad. Sci. U. S. A. 2007;104:16032–16037. doi: 10.1073/pnas.0701309104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- De Panfilis C., Schwarzbauer C. Positive or negative blips? The effect of phase encoding scheme on susceptibility-induced signal losses in EPI. NeuroImage. 2005;25:112–121. doi: 10.1016/j.neuroimage.2004.11.014. [DOI] [PubMed] [Google Scholar]
- Devlin J.T., Russell R.P., Davis M.H., Price C.J., Wilson J., Moss H.E., Matthews P.M., Tyler L.K. Susceptibility-induced loss of signal: comparing PET and fMRI on a semantic task. NeuroImage. 2000;11:589–600. doi: 10.1006/nimg.2000.0595. [DOI] [PubMed] [Google Scholar]
- Edmister W.B., Talavage T.M., Ledden P.J., Weisskoff R.M. Improved auditory cortex imaging using clustered volume acquisitions. Hum. Brain Mapp. 1999;7:89–97. doi: 10.1002/(SICI)1097-0193(1999)7:2<89::AID-HBM2>3.0.CO;2-N. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eickhoff S., Stephan K., Mohlberg H., Grefkes C., Fink G., Amunts K., Zilles K. A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. NeuroImage. 2005;25:1325–1335. doi: 10.1016/j.neuroimage.2004.12.034. [DOI] [PubMed] [Google Scholar]
- Elliott M.R., Bowtell R.W., Morris P.G. The effect of scanner sound in visual, motor, and auditory functional MRI. Magn. Reson. Med. 1999;41:1230–1235. doi: 10.1002/(sici)1522-2594(199906)41:6<1230::aid-mrm20>3.0.co;2-1. [DOI] [PubMed] [Google Scholar]
- Friston K.J., Ashburner J., Frith C.D., Poline J.-B., Heather J.D., Frackowiak R.S.J. Spatial registration and normalization of images. Hum. Brain Mapp. 1995;2:165–189. [Google Scholar]
- Gaab N., Gabrieli J.D.E., Glover G.H. Assessing the influence of scanner background noise on auditory processing. II. An fMRI study comparing auditory processing in the absence and presence of recorded scanner noise using a sparse design. Hum. Brain Mapp. 2007;28:721–732. doi: 10.1002/hbm.20299. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Giraud A.-L., Lorenzi C., Ashburner J., Wable J., Johnsrude I., Frackowiak R., Kleinschmidt A. Representation of the temporal envelope of sounds in the human brain. J. Neurophysiol. 2000;84:1588–1598. doi: 10.1152/jn.2000.84.3.1588. [DOI] [PubMed] [Google Scholar]
- Grimault N., Micheyl C., Carlyon R.P., Arthaud P., Collet L. Perceptual auditory stream segregation of sequences of complex sounds in subjects with normal and impaired hearing. Br. J. Audiol. 2001;35:173–182. doi: 10.1080/00305364.2001.11745235. [DOI] [PubMed] [Google Scholar]
- Hall D.A., Haggard M.P., Akeroyd M.A., Palmer A.R., Summerfield A.Q., Elliott M.R., Gurney E.M., Bowtell R.W. “Sparse” temporal sampling in auditory fMRI. Hum. Brain Mapp. 1999;7:213–223. doi: 10.1002/(SICI)1097-0193(1999)7:3<213::AID-HBM5>3.0.CO;2-N. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hall D.A., Summerfield A.Q., Gonçalves M.S., Foster J.R., Palmer A.R., Bowtell R.W. Time-course of the auditory BOLD response to scanner noise. Magn. Reson. Med. 2000;43:601–606. doi: 10.1002/(sici)1522-2594(200004)43:4<601::aid-mrm16>3.0.co;2-r. [DOI] [PubMed] [Google Scholar]
- Hall D.A., Chambers J., Akeroyd M.A., Foster J.R. Acoustic, psychophysical, and neuroimaging measurements of the effectiveness of active cancellation during auditory functional magnetic resonance imaging. J. Acoust. Soc. Am. 2009;125:347–359. doi: 10.1121/1.3021437. [DOI] [PubMed] [Google Scholar]
- Haller S., Bartsch A.J., Radue E.W., Klarhöfer M., Seifritz E., Scheffler K. Effect of fMRI acoustic noise on non-auditory working memory task: comparison between continuous and pulsed sound emitting EPI. Magn. Reson. Mater. Phys. 2005;18:263–271. doi: 10.1007/s10334-005-0010-2. [DOI] [PubMed] [Google Scholar]
- Harms M.P., Melcher J.R. Sound repetition rate in the human auditory pathway: representations in the waveshape and amplitude of fMRI activation. J. Neurophysiol. 2002;88:1433–1450. doi: 10.1152/jn.2002.88.3.1433. [DOI] [PubMed] [Google Scholar]
- Herdener M., Esposito F., Di Salle F., Lehmann C., Bach D.R., Scheffler K., Seifritz E. BOLD correlates of edge detection in human auditory cortex. NeuroImage. 2007;36:194–201. doi: 10.1016/j.neuroimage.2007.01.050. [DOI] [PubMed] [Google Scholar]
- Hickok G., Poeppel D. The cortical organization of speech processing. Nat. Rev. Neurosci. 2007;8:393–402. doi: 10.1038/nrn2113. [DOI] [PubMed] [Google Scholar]
- Jenkinson M. Fast, automated N-dimensional phase-unwrapping algorithm. Magn. Reson. Med. 2003;49:193–197. doi: 10.1002/mrm.10354. [DOI] [PubMed] [Google Scholar]
- Krüger G., Glover G.H. Physiological noise in oxygenation-sensitive magnetic resonance imaging. Magn. Reson. Med. 2001;46:631–637. doi: 10.1002/mrm.1240. [DOI] [PubMed] [Google Scholar]
- Loftus G.R., Masson M.E.J. Using confidence intervals in within-subject designs. Psychon. Bull. Rev. 1994;1:476–490. doi: 10.3758/BF03210951. [DOI] [PubMed] [Google Scholar]
- Moelker A., Pattynama P.M.T. Acoustic noise concerns in functional magnetic resonance imaging. Hum. Brain Mapp. 2003;20:123–141. doi: 10.1002/hbm.10134. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Morosan P., Rademacher J., Schleicher A., Amunts K., Schormann T., Zilles K. Human primary auditory cortex: cytoarchitectonic subdivisions and mapping into a spatial reference system. NeuroImage. 2001;13:684–701. doi: 10.1006/nimg.2000.0715. [DOI] [PubMed] [Google Scholar]
- Mummery C.J., Ashburner J., Scott S.K., Wise R.J. Functional neuroimaging of speech perception in six normal and two aphasic subjects. J. Acoust. Soc. Am. 1999;106:449–457. doi: 10.1121/1.427068. [DOI] [PubMed] [Google Scholar]
- Ojemann J.G., Akbudak E., Snyder A.Z., McKinstry R.C., Raichle M.E., Conturo T.E. Anatomic localization and quantitative analysis of gradient refocused echo-planar fMRI susceptibility artifacts. NeuroImage. 1997;6:156–167. doi: 10.1006/nimg.1997.0289. [DOI] [PubMed] [Google Scholar]
- Orfanidou E., Marslen-Wilson W.D., Davis M.H. Neural response suppression predicts repetition priming of spoken words and pseudowords. J. Cogn. Neurosci. 2006;18:1237–1252. doi: 10.1162/jocn.2006.18.8.1237. [DOI] [PubMed] [Google Scholar]
- Peelle J.E., Wingfield A. Dissociations in perceptual learning revealed by adult age differences in adaptation to time-compressed speech. J. Exp. Psychol. Hum. Percept. Perform. 2005;31:1315–1330. doi: 10.1037/0096-1523.31.6.1315. [DOI] [PubMed] [Google Scholar]
- Poldrack R.A., Fletcher P.C., Henson R.N., Worsley K.J., Brett M., Nichols T.E. Guidelines for reporting an fMRI study. NeuroImage. 2008;40:409–414. doi: 10.1016/j.neuroimage.2007.11.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Price C.J. The anatomy of language: contributions from functional neuroimaging. J. Anat. 2000;197:335–359. doi: 10.1046/j.1469-7580.2000.19730335.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Price D.L., De Wilde J.P., Papadaki A.M., Curran J.S., Kitney R.I. Investigation of acoustic noise on 15 MRI scanners from 0.2T to 3T. J. Magn. Reson. Imaging. 2001;13:288–293. doi: 10.1002/1522-2586(200102)13:2<288::aid-jmri1041>3.0.co;2-p. [DOI] [PubMed] [Google Scholar]
- Ravicz M.E., Melcher J.R., Kiang N.Y.-S. Acoustic noise during functional magnetic resonance imaging. J. Acoust. Soc. Am. 2000;108:1683–1696. doi: 10.1121/1.1310190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rodd J.M., Davis M.H., Johnsrude I.S. The neural mechanisms of speech comprehension: fMRI studies of semantic ambiguity. Cereb. Cortex. 2005;15:1261–1269. doi: 10.1093/cercor/bhi009. [DOI] [PubMed] [Google Scholar]
- Rorden C., Brett M. Stereotaxic display of brain lesions. Behav. Neurol. 2000;12:191–2000. doi: 10.1155/2000/421719. [DOI] [PubMed] [Google Scholar]
- Scheffler K., Bilecen D., Schmid N., Tschopp K., Seelig J. Auditory cortical responses in hearing subjects and unilateral deaf patients as detected by functional magnetic resonance imaging. Cereb. Cortex. 1998;8:156–163. doi: 10.1093/cercor/8.2.156. [DOI] [PubMed] [Google Scholar]
- Schmitter S., Diesch E., Amann M., Kroll A., Moayer M., Schad L.R. Silent echo-planar imaging for auditory FMRI. Magn. Reson. Mater. Phys. 2008;21:317–325. doi: 10.1007/s10334-008-0132-4. [DOI] [PubMed] [Google Scholar]
- Schroeder M.R. Reference signal for signal quality studies. J. Acoust. Soc. Am. 1968;44:1735–1736. [Google Scholar]
- Schwarzbauer C., Davis M.H., Rodd J.M., Johnsrude I. Interleaved silent steady state (ISSS) imaging: a new sparse imaging method applied to auditory fMRI. NeuroImage. 2006;29:774–782. doi: 10.1016/j.neuroimage.2005.08.025. [DOI] [PubMed] [Google Scholar]
- Seifritz E., Di Salle F., Esposito F., Bilecen D., Neuhoff J.G., Scheffler K. Sustained blood oxygenation and volume response to repetition rate-modulated sound in human auditory cortex. NeuroImage. 2003;20:1365–1370. doi: 10.1016/S1053-8119(03)00421-X. [DOI] [PubMed] [Google Scholar]
- Seifritz E., Di Salle F., Esposito F., Herdener M., Neuhoff J.G., Scheffler K. Enhancing BOLD response in the auditory system by neurophysiologically tuned fMRI sequence. NeuroImage. 2006;29:1013–1022. doi: 10.1016/j.neuroimage.2005.08.029. [DOI] [PubMed] [Google Scholar]
- Talavage T.M., Edmister W.B., Ledden P.J., Weisskoff R.M. Quantitative assessment of auditory cortex responses induced by imager acoustic noise. Hum. Brain Mapp. 1999;7:79–88. doi: 10.1002/(SICI)1097-0193(1999)7:2<79::AID-HBM1>3.0.CO;2-R. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tanaka H., Fujita N., Watanabe Y., Hirabuki N., Takanashi M., Oshiro Y., Nakamura H. Effects of stimulus rate on the auditory cortex using fMRI with ‘sparse’ temporal sampling. NeuroReport. 2000;11:2045–2049. doi: 10.1097/00001756-200006260-00047. [DOI] [PubMed] [Google Scholar]
- Van Casteren M., Davis M.H. Match: a program to assist in matching the conditions of factorial experiments. Behav. Res. Meth. 2007;39:973–978. doi: 10.3758/bf03192992. [DOI] [PubMed] [Google Scholar]
- Weiskopf N., Hutton C., Josephs O., Turner R., Deichmann R. Optimized EPI for fMRI studies of the orbitofrontal cortex: compensation of susceptibility-induced gradients in the readout direction. Magn. Reson. Mater. Phys. 2007;20:39–49. doi: 10.1007/s10334-006-0067-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wingfield A., McCoy S.L., Peelle J.E., Tun P.A., Cox C. Effects of adult aging and hearing loss on comprehension of rapid speech varying in syntactic complexity. J. Am. Acad. Audiol. 2006;17:487–497. doi: 10.3766/jaaa.17.7.4. [DOI] [PubMed] [Google Scholar]
- Worsley K.J., Evans A.C., Marrett S., Neelin P. A three-dimensional statistical analysis for CBF activation studies in human brain. J. Cereb. Blood Flow Metab. 1992;12:900–918. doi: 10.1038/jcbfm.1992.127. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.