Abstract
Verbal communication in noisy backgrounds is challenging. Understanding speech in background noise that fluctuates in intensity over time is particularly difficult for hearing-impaired listeners with a sensorineural hearing loss (SNHL). The reduction in fast-acting cochlear compression associated with SNHL exaggerates the perceived fluctuations in intensity in amplitude-modulated sounds. SNHL-induced changes in the coding of amplitude-modulated sounds may have a detrimental effect on the ability of SNHL listeners to understand speech in the presence of modulated background noise. To date, direct evidence for a link between magnified envelope coding and deficits in speech identification in modulated noise has been absent. Here, magnetoencephalography was used to quantify the effects of SNHL on phase locking to the temporal envelope of modulated noise (envelope coding) in human auditory cortex. Our results show that SNHL enhances the amplitude of envelope coding in posteromedial auditory cortex, whereas it enhances the fidelity of envelope coding in posteromedial and posterolateral auditory cortex. This dissociation was more evident in the right hemisphere, demonstrating functional lateralization in enhanced envelope coding in SNHL listeners. However, enhanced envelope coding was not perceptually beneficial. Our results also show that both hearing thresholds and, to a lesser extent, magnified cortical envelope coding in left posteromedial auditory cortex predict speech identification in modulated background noise. We propose a framework in which magnified envelope coding in posteromedial auditory cortex disrupts the segregation of speech from background noise, leading to deficits in speech perception in modulated background noise.
SIGNIFICANCE STATEMENT People with hearing loss struggle to follow conversations in noisy environments. Background noise that fluctuates in intensity over time poses a particular challenge. Using magnetoencephalography, we demonstrate anatomically distinct cortical representations of modulated noise in normal-hearing and hearing-impaired listeners. This work provides the first link among hearing thresholds, the amplitude of cortical representations of modulated sounds, and the ability to understand speech in modulated background noise. In light of previous work, we propose that magnified cortical representations of modulated sounds disrupt the separation of speech from modulated background noise in auditory cortex.
Keywords: envelope coding, magnetoencephalography, sensorineural hearing loss, speech perception
Introduction
Hearing loss is a major health issue that affects >40% of the population who are 60 years of age or older (Agrawal et al., 2008). The most common form of hearing loss, sensorineural hearing loss (SNHL), is associated with damage to the hair cells in the cochlea. In addition to elevating audiometric thresholds, SNHL alters the perception and neural representations of sounds. For example, the reduction in fast-acting compression associated with outer hair cell dysfunction exaggerates the perceived fluctuations in the amplitude of modulated sounds (Moore et al., 1996). However, magnified neural coding of the temporal envelope of modulated sounds may not be beneficial for real-world listening situations. Simulations of SNHL in normal-hearing listeners suggest that magnified neural coding of sound envelopes has a detrimental effect on the ability to understand speech in the presence of modulated background noise (Moore and Glasberg, 1993). Magnified envelope coding may distract hearing-impaired listeners from using other auditory cues to aid speech perception in noise (Kale and Heinz, 2010; Henry et al., 2014; Zhong et al., 2014).
Direct evidence for a link between magnified envelope coding and deficits in the ability to understand speech in modulated noise backgrounds is lacking. To address this issue, we used magnetoencephalography (MEG) to measure cortical phase-locking to the temporal envelope of modulated noise (envelope coding) in normal-hearing (NH) and hearing-impaired listeners with bilateral SNHL. We considered both groups because NH and SNHL listeners are known to differ in their ability to “listen in the dips” of modulated noise to aid speech perception; that is, to take advantage of the higher signal-to-noise ratio (SNR) during amplitude minima in modulated background noise (Duquesnoy, 1983; Festen and Plomp, 1990; Moore et al., 1999; Moore, 2008).
We tested the hypothesis that magnified cortical envelope coding is associated with deficits in speech perception in modulated noise backgrounds and, in addition, we aimed to quantify the effects of SNHL on the fidelity of cortical envelope coding. Indeed, speech perception in modulated noise may partially rely on accurate coding of the amplitude minima in the envelope shape of a fluctuating background noise (Grose et al., 2009). The fidelity of envelope coding in human auditory cortex may provide a measure of temporal processing that is directly related to the ability to benefit from the temporal dips in modulated maskers. The amplitude of the phase-locked response to the modulated noise was measured using a general linear model (GLM) approach. Cross-correlation analyses provided a measure of the fidelity of envelope coding, that is, a measure of the accuracy with which the temporal structure of the predictor is represented by the measured cortical activity (see Materials and Methods: “MEG analysis”).
In the present study, we found that both the amplitude and the fidelity of cortical envelope coding of square-wave-modulated noise were enhanced in SNHL listeners compared with NH listeners. Cortical envelope coding of modulated noise was lateralized toward right auditory cortex within both listeners groups, consistent with asymmetric sampling in time (AST) theory (Poeppel, 2003). AST theory predicts that the cortical coding of slowly fluctuating sounds (∼4 Hz) is lateralized toward the right hemisphere because neurons in right auditory cortex integrate preferentially over time windows of ∼250 ms. Enhanced envelope coding was more evident in right auditory cortex despite the symmetrical hearing loss of the SNHL listeners. These results permit new insights into the functional lateralization of enhanced cortical envelope coding in SNHL listeners. Critically, both hearing thresholds and, to a lesser extent, the amplitude of envelope coding in left posteromedial auditory cortex were predictive of deficits in the identification of speech sentences presented against a background of modulated noise.
Materials and Methods
Participants
Seventeen (4 male) NH listeners (mean age = 57 years, SD = 5 years) and 17 (8 male) SNHL listeners (mean age = 61 years, SD = 11 years), all right-handed native English speakers, participated in the study. A one-way ANOVA showed that there was no significant age difference between the NH and SNHL listeners (F(1,32) = 1.42, p = 0.24). All listeners provided written informed consent in accordance with the Declaration of Helsinki and were paid for their participation in the study. The study was approved by the Research Governance Committee at the University of York.
Audiological assessment
Audiometric thresholds were measured in accordance with the procedures recommended by the British Society of Audiology (British Society of Audiology, 2004). Pure tone air conduction thresholds were measured for all listeners and bone conduction thresholds were also measured for hearing-impaired listeners. Audiometric thresholds were measured for frequencies of 0.5, 1, 2, and 4 kHz only because stimuli were band-pass filtered with cutoff frequencies of 0.5 and 4 kHz (linear-phase FIR digital filter followed by sixth-order Butterworth filter). The NH listeners had clinically normal hearing in both ears, defined here as pure tone audiometric thresholds of no more than 20 dB HL for octave frequencies between 0.5 and 4 kHz. All SNHL listeners had a bilateral mild to moderate hearing loss, defined as audiometric thresholds >20 dB HL in both ears for at least one test frequency.
Hearing thresholds for NH and SNHL listeners were analyzed in a 2 (hearing status) × 2 (ear) × 4 (hearing test frequency) mixed repeated-measures ANOVA. Greenhouse–Geisser corrected p-values are reported where necessary.
Behavioral measures of speech perception in noise
Listeners were seated in a double-walled sound-attenuating booth. Stimuli were delivered diotically through Sennheiser HD 650 headphones. Stimuli were played through an E-MU soundcard using custom MATLAB (The MathWorks) routines. Linear frequency-dependent amplification (Moore and Glasberg, 1998) was used to increase the audibility of the stimuli for SNHL listeners (see “Audibility of auditory stimuli” section).
The ability to benefit from the temporal dips in modulated maskers was measured using target sentences from the IEEE corpus (Rothauser et al., 1969). The entire list of IEEE sentences was sorted based on their duration: 120 sentences of the IEEE shortest sentences (mean duration = 2.22 s, SD = 0.06 s) were used for the behavioral testing and a further 30 sentences were used for training before behavioral testing. Sentences contained four or five key words (mean number of key words = 4.92, SD = 0.27); for example, “The lake sparkled in the red hot sun” (keywords underlined). At the end of each sentence, listeners were instructed to type any words that they could understand using a computer keyboard.
Keyword identification was measured in a masking noise that was spectrally matched to the long-term power spectrum of the speech sentences. The masker was either unmodulated (see Fig. 3A) or 100% modulated with a 2 Hz square wave (50% duty cycle) (see Fig. 3B). The “on” and “off” periods of the modulated masker were ∼250 ms in duration because the on/off slopes of the square-wave modulator were shaped with 5 ms cosine squared ramps. The most common rise time in the temporal envelopes of IEEE speech sentences is 12 ms (79 modulations/s) across auditory 128 filters, but many envelope fluctuations with a rise time of 4–6 ms (61 modulations/s) are also present (Prendergast et al., 2011).
Different noise samples were selected for each presentation. When the noise was modulated, the first period of the noise was always “on” and therefore the first 250 ms portion of each sentence was masked. Therefore the modulated noise was always in phase with the sentence onset, which would make the amplitude minima in the modulated noise predictable given that only one modulation rate was tested. The predictability of the temporal dips in the modulated masker should improve absolute performance in the modulated masker condition and increase the amount of speech MR relative to a modulated masker with a random starting phase. The speech-in-noise stimuli were ramped on and off using a 25 ms raised cosine function. SNRs were fixed at −4, −8, and −12 dB. Twenty sentences were randomly assigned to each of three SNR conditions (−4, −8, and −12 dB) and two masker conditions (unmodulated, modulated). Because the lower SNRs (−8 and −12 dB) resulted in floor performance in the unmodulated masker condition, speech MR was only measured for −4 dB SNR.
Speech masking release, the ability to benefit from the temporal dips in modulated maskers, was defined as the difference in performance (percentage of sentence keywords correctly identified) between the unmodulated and modulated maskers presented at a fixed SNR of −4 dB (Bernstein and Grant, 2009; Gregan et al., 2013). Individual scores for correct identification of keywords were transformed into rationalized arcsine units (Studebaker, 1985) for statistical analyses.
The percentage of sentence keywords correctly identified were analyzed in a 2 (hearing status) × 2 (masker type) mixed repeated-measures ANOVA. The number of correctly identified keywords in each keyword position within a sentence was analyzed in a 2 (hearing status) × 5 (keyword position) mixed repeated-measures ANOVA. Greenhouse–Geisser corrected p-values are reported where necessary.
Audibility of auditory stimuli
The Cambridge formula (Moore and Glasberg, 1998) was used to improve audibility for the SNHL listeners. Based on the audiometric threshold of each listener, gains specified at audiometric frequencies between 0.5–4 kHz were prescribed by the Cambridge formula (CAMEQ):
where IG(f) is the insertion gain in dB at frequency f, HL(f) is the hearing loss in dB at frequency f, and INT(f) is a frequency-dependent intercept. The CAMEQ was applied after speech and noise were mixed at −4 dB SNR. The prescribed gains were applied to the processed sounds using a linear-phase FIR digital filter (Vickers et al., 2001; Hopkins, 2009); that is, the MATLAB FIR2 function with 443 taps (Hopkins, 2009). For SNHL listeners, the CAMEQ was used to apply frequency-dependent amplification based on individual audiometric thresholds (0.5–4 kHz), calculated, and applied separately for each ear before stimulus presentation. The stimuli for NH listeners were also subjected to the signal processing pipeline required to apply the CAMEQ, but the frequency-dependent amplification was not applied. The stimulus level was 65 dB SPL for NH listeners.
MEG procedure
Listeners who had taken part in the behavioral experiment were invited to take part in a separate MEG session at a later date. The modulated noise, which was used to measure the perceptual benefit of the modulated masker for speech perception in noise, was also used to measure cortical envelope coding. One hundred different samples of modulated noise were played to the listeners, with one sample of modulated noise presented per trial. The duration of each sample of modulated noise was 2000 ms. One hundred epochs of silence, 100 epochs of spoken sentences mixed with unmodulated noise (SNR = −4 dB), and 100 epochs of spoken sentences mixed with modulated noise (SNR = −4 dB) were also presented during the MEG recording. The speech sentences used in the MEG session were from the IEEE corpus (Rothauser et al., 1969), but listeners had not been exposed to these sentences before the MEG recording. The duration of each epoch was increased to 3000 ms through the addition of silence to the end of each stimulus. Stimuli were presented to both ears through ER30 insert earphones (Etymotic Research). Linear-frequency-dependent amplification was applied to stimuli for SNHL listeners (see “Audibility of auditory stimuli” section). Before linear amplification, the stimulus level was 65 dB SPL.
Data were collected using a Magnes 3600 whole-head 248-channel magnetometer (originally manufactured by 4-D Neuroimaging). The data were recorded at a sample rate of 678.17 Hz and digitally filtered between 1 and 200 Hz online. Participants were asked to close their eyes during the MEG recording. Catch trials (10% of the total number of trials) were used to maintain a constant level of alertness. During a catch trial, participants were presented with an auditory cue (the word “rate”), which required a button press on a response box. The auditory cue prompted participants to indicate, via the button press, whether the previous trial contained an intelligible (masked) speech sentence, an unintelligible (masked) speech sentence, or a sound containing no speech sentence, that is, modulated noise alone.
MEG data were coregistered with the anatomical magnetic resonance (MR) scans of individual listeners. Before MEG data acquisition, individual facial and scalp landmarks were spatially coregistered using a Polhemus Fastrak System. The landmark locations in relation to the sensor positions were derived on the basis of a precise localization signal provided by five spatially distributed head coils with a fixed spatial relation to the landmarks. These head coils provided a measurement of the listeners' head movement at the beginning and end of each recording. The landmark locations were matched with the individual listeners' anatomical MR scans using a surface-matching technique adapted from Kozinska et al. (2001). T1-weighted MR images were acquired with a 3.0 T Signa Excite HDx system (General Electric) using an 8-channel head coil and a 3-D Fast Spoiled Gradient Recall Sequence (TR/TE/flip angle = 8.03 ms/3.07 ms/20°, spatial resolution of 1.13 mm × 1.13 mm × 1.0 mm, with an in-plane acquisition matrix of 256 × 256 and >176 contiguous slices).
MEG analysis
The raw data from each epoch were inspected visually. Epochs containing physiological or nonphysiological artifacts were removed.
Derivation of spatial filters.
In this study, a vectorized, linearly constrained minimum variance beamformer (Van Veen et al., 1997; Huang et al., 2004) was used to obtain the spatial filters with a multiple-spheres head model (Huang et al., 1999). The beamformer grid size was 5 mm. The three orthogonal spatial filters were implemented as a single 3-D system (Johnson et al., 2011) (Fig. 1A). Spatial filters were generated with a time window of 2500 ms, including 500 ms before stimulus presentation, and a 1–10 Hz band-pass filter.
Analysis of spatial filter outputs.
The amplitudes of the MEG responses were analyzed using a GLM approach as follows:
where each observed data point (Y) is the sum of a constant term (β0), the β of interest (β1), the stimulus (X1), and residual error (ε). GLM analyses were performed using the regstats function in MATLAB (The MathWorks). The outputs of the beamformer spatial filters were the observed data. The temporal structure of the square-wave-gated noise was the predictor after the predictor was smoothed (Brookes et al., 2004) using a low-pass filter with a cutoff frequency of 10 Hz (Fig. 1B). The β1 of the GLM gives the amplitude of envelope coding. The r2 of the GLM was used to optimize the orientation of the spatial filters (Millman et al., 2015) (Fig. 1B). The r2 of the model fit, rather than β1, was chosen as the optimizing metric to increase the likelihood that the spatial filter output resembled a square wave. Table 1 shows the mean GLM r2 for both listener groups in each cortical area of interest. The overall GLM fit was significant in all individual listeners (p < 0.001).
Table 1.
GLM r2 |
||||
---|---|---|---|---|
Right HG | Left HG | Right STG | Left STG | |
NH | 0.46 (0.14) | 0.34 (0.14) | 0.42 (0.13) | 0.37 (0.17) |
SNHL | 0.56 (0.14) | 0.35 (0.13) | 0.59 (0.10) | 0.4 (0.12) |
Means and SD (in parentheses) of the GLM r2 are shown for NH and SNHL listeners in each cortical LOI.
The fidelity of envelope coding was measured using cross-correlation in the time domain using the MATLAB xcov function (Abrams et al., 2008; Nourski et al., 2009; Millman et al., 2013). Cross-correlation analyses were performed between the smoothed square-wave predictor and the phase-locked MEG responses. Cross-correlation coefficients (c) were Fisher transformed before statistical analyses (Abrams et al., 2008; Millman et al., 2013, 2015).
Measures of cortical envelope coding were restricted to the “envelope following period” (Abrams et al., 2008) from 250–2000 ms after stimulus presentation. A variable lag of 0–150 ms (Nourski et al., 2009; Millman et al., 2013) was introduced between the onset of stimulus presentation and the MEG signal to identify the lag that resulted in the best r2 of the GLM fit or maximum c for cross-correlation analyses (Fig. 1B). These lags were optimized for individual participants.
The MEG data were analyzed in a 2 (hearing status) × 2 (hemisphere) × 2 (location) repeated-measures ANOVA. The optimal stimulus–response lags were analyzed in a 2 (hearing status) × 2 (analysis method) × 2 (hemisphere) × 2 (location) repeated-measures ANOVA.
Locations of spatial filters.
The MEG data were analyzed using a location of interest (LOI) approach (Millman et al., 2013, 2015). The LOIs were left and right posteromedial Heschl's Gyrus (HG) and the posterolateral portion of left and right superior temporal gyrus (STG) (see Fig. 4B). These LOIs were chosen because they are known to phase lock to temporal modulations in sounds (Nourski et al., 2009, 2013; Ding and Simon, 2012; Millman et al., 2015). Posterolateral STG has been identified as playing a significant role in the perception of speech in noise: This brain region may be used when increased effort is required to understand speech in noise (Wong et al., 2009), during listening in the dips of modulated maskers (Scott et al., 2009), and for object-based neural representations (Ding and Simon, 2012).
LOIs were seeded manually in HG (Millman et al., 2013, 2015) because of the interlistener variability in the anatomy of HG (Rademacher et al., 2001). LOIs in STG were based on the region identified in left posterolateral STG (MNI −64 −30 8) by Wong et al. (2009) and its right hemisphere homolog (MNI 64 −30 8). These MNI coordinates for the LOI in posterolateral STG fall within the Harvard–Oxford probabilistic cortical atlas for Brodmann's area 42. MNI coordinates for left and right STG were transformed back into individuals' anatomical space and spatial filters were generated from these locations.
LOIs in left and right auditory cortices were selected to test the predictions of the AST model regarding right hemisphere lateralization for low-frequency envelope coding (Poeppel, 2003). LOIs in both early (HG) and associative (STG) auditory cortex were analyzed because: (1) the locus of functional asymmetry predicted by AST (Poeppel, 2003) remains unclear and (2) there is evidence that speech is at least partially separated from background noise in HG, whereas complete segregation of speech and noise occurs in posterior STG (Ding and Simon, 2012; Simon, 2015).
Relationship between speech perception in noise and envelope coding metrics
Hierarchical multiple regression analyses were used to assess the predictive value of the pure-tone average (PTA) hearing thresholds, age, and the envelope coding metrics (amplitude and fidelity) on speech identification in the presence of the modulated masker. Both NH and SNHL listeners were entered into the regression analysis. Known predictors of speech perception in noise, namely age (Dubno et al., 2002) and PTA (Bacon et al., 1998), were entered in the first step of the model. The envelope coding metrics were included in a stepwise fashion in the second step of the model because the relative contributions of the amplitude and fidelity of cortical envelope coding to speech perception in noise are unknown. Collinearity diagnostics did not suggest that the regression model was influenced by multicollinearity: The mean variance inflation factor (VIF) was not substantially >1 (mean VIF = 1.25, SD = 0.44) and the maximum VIF was <10 (maximum VIF = 2.55) (Bowerman and O'Connell, 1990).
Results
Symmetrical hearing thresholds in NH and SNHL listeners
Figure 2 shows the individual and mean hearing thresholds for the NH and SNHL listeners. A mixed repeated-measures ANOVA was used to analyze the hearing thresholds. The between-participants factor was hearing status (NH or SNHL). The within-participants factors were ear (right or left) and test frequency (0.5, 1, 2, or 4 kHz). For the comparisons of test frequency, the assumption of sphericity was violated (Mauchly's test: χ2(5) = 35.03, p < 0.001); therefore, the degrees of freedom were corrected (Greenhouse–Geisser; ε = 0.72). The ANOVA revealed main effects of hearing status (F(1,32) = 80.38, p < 0.001, partial η2 = 0.72) and test frequency (F(2.15,68.68) = 37.32, p < 0.001, partial η2 = 0.54). There was also a significant interaction between hearing status and test frequency ((F(2.15,68.68) = 26.8, p < 0.001, partial η2 = 0.46). Figure 2 shows that SNHL listeners had higher (i.e., worse) hearing thresholds than NH listeners, that hearing thresholds increased as a function of test frequency, and that the difference in hearing thresholds of NH and SNHL listeners was increased for higher test frequencies. Importantly, the ANOVA showed that there was no main effect of ear (F(1,32) = 0.28, p = 0.60, partial η2 = 0.01), demonstrating that hearing thresholds were symmetrical across ears for both listener groups.
Speech perception in noise
The percentage of correctly identified sentence keywords in the presence of unmodulated and modulated noise maskers are shown in Figure 3C. Speech perception in the unmodulated masker was poor for both NH and SNHL listeners. Noise modulation improved speech perception for both groups. An ANOVA was used to assess speech perception performance with the between-participants factor hearing status (NH or SNHL) and the within-participants factor masker type (unmodulated or modulated). There were significant effects of hearing status (F(1,32) = 49.23, p < 0.001, partial η2 = 0.61) and masker type (F(1,32) = 394.1, p < 0.001, partial η2 = 0.93) and an interaction between the two (F(1,32) = 24.24, p < 0.001, partial η2 = 0.43). The interaction demonstrates that the NH listeners were more able to benefit more from the temporal dips in the modulated masker than SNHL listeners (Bacon et al., 1998; Gregan et al., 2013). Figure 3D shows the difference in speech identification in the modulated and unmodulated masker conditions for NH and SNHL listeners.
Table 2 shows the mean number of correctly identified keywords in NH and SNHL listeners as a function of the keyword position (first, second, third, fourth, or fifth) in the IEEE sentences used in the present study. To test for a predictability gain in speech comprehension for final key words (Hartwigsen et al., 2015), a mixed repeated-measures ANOVA was used to measure a change in correct keyword identification as a function of keyword position. The between-participants factor was hearing status (NH or SNHL) and the within-participants factor was keyword position (first, second, third, fourth, or fifth). For the comparisons of keyword position, the assumption of sphericity was violated (Mauchly's test: χ2(9) = 24.67, p = 0.003); therefore, the degrees of freedom were corrected (Greenhouse–Geisser; ε = 0.72). There was an effect of hearing status (F(1,32) = 62.28, p < 0.001, partial η2 = 0.66) because NH listeners identified more keywords correctly. There was also an effect of keyword position (F(2.89, 92.43) = 47.74, p < 0.001; partial η2 = 0.60) because the number of keywords correctly identified increased as a function of keyword position in both listener groups. There was also a significant interaction between keyword position and hearing status (F(2.89, 92.43) = 83.67, p < 0.001; partial η2 = 0.18), suggesting that NH listeners were more able to identify the later keywords correctly.
Table 2.
Keyword position (no. identified correctly) |
|||||
---|---|---|---|---|---|
First | Second | Third | Fourth | Fifth | |
NH | 17.41 (6.09) | 18.29 (5.10) | 23.29 (6.96) | 25.94 (4.99) | 28.41 (4.70) |
SNHL | 7.24 (4.88) | 6.35 (5.77) | 8.17 (6.44) | 10.41 (5.71) | 11.94 (6.46) |
Means and SD (in parentheses) of the number of keywords correctly identified in each keyword position are shown for NH and SNHL listeners.
Post hoc comparisons (Bonferroni corrected) were used to compare the number of correctly identified keywords in the final (fifth) position with the other keyword positions within listener groups. Consistent with a predictability gain for the final keyword (Hartwigsen et al., 2015), the number of final keywords correctly identified was significantly greater than the number of keywords in other positions in both listener groups: first versus fifth (NH p < 0.001, d = 2.83; SNHL p = 0.007, d = 0.88); second versus fifth (NH p < 0.001, d = 2.20; SNHL p < 0.001, d = 1.14); third versus fifth (NH p = 0.001, d = 0.94; SNHL p = 0.005, d = 0.80); fourth versus fifth (NH p = 0.007, d = 0.76; SNHL p = 0.011, d = 0.70).
SNHL enhances cortical coding of modulated noise
The MEG responses to the modulated noise (Fig. 4A) from the LOIs in HG and STG (Fig. 4B) were analyzed to obtain the amplitude and fidelity of cortical envelope coding (Fig. 4C).
The mean amplitudes of cortical envelope coding in HG (Fig. 4D) and STG (Fig. 4E) were compared in NH and SNHL listeners. Mixed repeated-measures ANOVA were used to assess the amplitudes of envelope coding in NH and SNHL listeners. The between-participants factor was hearing status (NH or SNHL). The within-participants factors were hemisphere (left or right) and location (HG or STG). A main effect of hearing status (F(1,32) = 7.45, p = 0.01, partial η2 = 0.19) showed that envelope coding was magnified in SNHL listeners (Fig. 4D,E). Likewise, a main effect of hemisphere (F(1,32) = 11.21, p = 0.002, partial η2 = 0.26) showed that envelope coding was lateralized toward the right hemisphere (Fig. 4D,E). There was also a main effect of location (F(1,32) = 219.46, p < 0.001, partial η2 = 0.87) because the amplitude of envelope coding was greater in HG (Fig. 4D) than in STG (Fig. 4E). Finally, there was also a significant interaction between hemisphere and location (F(1,32) = 33.44, p < 0.001, partial η2 = 0.51). Post hoc comparisons (Bonferroni corrected) revealed that this interaction between hemisphere and location was driven by significantly magnified envelope coding in right HG versus left HG (NH p = 0.002, d = 0.88; SNHL p = 0.002, d = 0.93). There was no significant hemispheric difference for STG (NH p = 0.52; SNHL p = 0.94).
The mean fidelities of cortical envelope coding are shown in HG (Fig. 4F) and STG (Fig. 4G). An ANOVA with the between-participants factor of hearing status (NH or SNHL) and the within-participants factors of hemisphere (left or right) and location (HG or STG) showed a main effect of hearing status (F(1,32) = 8.51, p = 0.006, partial η2 = 0.21), indicating enhanced fidelity of envelope coding in SNHL listeners. There was also a main effect of hemisphere (F(1,32) = 21.26, p < 0.001, partial η2 = 0.40): Fidelity of envelope coding was greater in the right hemisphere than the left hemisphere. A significant interaction between hemisphere and location (F(1,32) = 5.72, p = 0.02, partial η2 = 0.15) revealed a significant difference (post hoc comparisons, Bonferroni corrected) in the fidelity of envelope coding in right HG versus left HG in both listener groups (NH p = 0.01, d = 0.70; SNHL p > 0.001, d = 1.11). The fidelity of envelope coding was also greater for SNHL listeners in right STG versus left STG (p = 0.001, d = 1.0) but not for NH listeners (p = 0.30).
The mean optimal stimulus–response lags obtained through either GLM or cross-correlation analyses are shown in Table 3. Note that the lags reported in Table 3 were calculated during the “envelope following period” from 250–2000 ms after stimulus presentation. A mixed repeated-measures ANOVA with a between-participants factor of hearing status (NH or SNHL) and the within-participants factors of analysis type (GLM or cross-correlation), hemisphere (left or right), and location (HG or STG) was performed to analyze the optimal stimulus–response lags. This ANOVA showed no effects of hearing status (F(1,32) = 3.04, p = 0.09, partial η2 = 0.09), analysis type (F(1,32) = 0.05, p = 0.82, partial η2 = 0.002), hemisphere (F(1,32) = 0.35, p = 0.56, partial η2 = 0.01), or location (F(1,32) = 2.69, p = 0.11, partial η2 = 0.08), nor any significant interactions between ANOVA terms.
Table 3.
Optimal stimulus–response lag (ms) |
||||
---|---|---|---|---|
Right HG | Left HG | Right STG | Left STG | |
GLM | ||||
NH | 55.9 (45.0) | 81.5 (51.1) | 79.7 (54.1) | 89.2 (54.4) |
SNHL | 50.4 (42.6) | 58 (47.6) | 71.7 (51.8) | 60.2 (46.6) |
Cross-correlation | ||||
NH | 55.7 (44.9) | 84.8 (47.7) | 82.7 (58.8) | 73 (56.6) |
SNHL | 54.5 (44.6) | 65 (52.5) | 73.6 (52.4) | 62.7 (51.3) |
Means and SD (in parentheses) of the stimulus–response lags are shown for each LOI for NH and SNHL listeners.
Magnified envelope coding in left HG predicts deficits in speech perception in modulated noise
Table 4 shows the results of the hierarchical regression analyses used to assess predictors of speech perception in modulated noise. Speech identification in modulated noise was reliably predicted by the first step of the regression model (R2 = 0.65, p < 0.001), which included age and PTA. Consistent with previous reports, PTA contributed significantly to the model fit (p < 0.001) (Gregan et al., 2013), but age did not make a unique contribution (p = 0.46) (Füllgrabe et al., 2014). Figure 5A illustrates that listeners with a higher PTA, that is, greater hearing loss, did not perform as well on the speech identification task (r = −0.84, p < 0.001; Pearson's correlation coefficient, 2-tailed). The envelope coding metrics added in the second step further improved the model fit (ΔR2 = 0.09, p = 0.003). However, the amplitude of envelope coding in left HG was the only unique contributor to this improvement (p = 0.003), with magnified envelope coding associated with poor speech identification in modulated noise (r = −0.52, p = 0.002; Pearson's correlation coefficient, 2-tailed) (Fig. 5B).
Table 4.
Model | Predictor | β | p |
---|---|---|---|
Step 1 | Age | 0.09 | 0.46 |
PTA | −0.84 | <0.001 | |
R2 = 0.65, adjusted R2 = 0.63 for Step 1 (p < 0.001) | |||
Step 2 | Age | 0.12 | 0.25 |
PTA | −0.85 | <0.001 | |
Left HG β | −0.31 | 0.003 | |
Left HG c | −0.07 | 0.63 | |
Right HG β | 0.03 | 0.81 | |
Right HG c | 0.001 | 0.99 | |
Left STG β | −0.10 | 0.43 | |
Left STG c | 0.09 | 0.39 | |
Right STG β | 0.04 | 0.67 | |
Right STG c | 0.11 | 0.39 | |
ΔR2 = 0.09, adjusted R2 = 0.72 for Step 2 (p = 0.003) |
Model parameters include standardized beta coefficients (β) and the significance value (p). The R2 for the initial model step and the change in R2 (ΔR2) for the second step of the model are also reported. Values in bold indicate results at p < 0.05.
The significance of the link between magnified envelope coding in left HG and speech perception in noise can be explained by examining the relationships between PTA and the envelope coding metrics entered into the regression model (Table 5). Table 5 shows that scores for correct identification of keywords in modulated noise (speech) were correlated with several parameters: PTA (r = −0.80, p < 0.001), fidelity of envelope coding in right HG (r = −0.41, p = 0.02), and fidelity of envelope coding in right STG (r = −0.49, p = 0.004). Speech was also marginally correlated with the amplitude of envelope coding in both left HG (r = −0.31, p = 0.07) and right HG (r = −0.33, p = 0.06). However, the bottom row in Table 5 shows that PTA was also correlated with the same envelope coding metrics in the right hemisphere that were correlated with speech: right HG amplitude (r = 0.40, p = 0.02), right HG fidelity (r = 0.46, p = 0.006), and right STG fidelity (r = 0.56, p < 0.001). Importantly, PTA was not correlated with the amplitude of envelope coding in left HG (r = 0.20, p = 0.91).
Table 5.
Left HG β | Left HG c | Right HG β | Right HG c | Left STG β | Left STG c | Right STG β | Right STG c | |
---|---|---|---|---|---|---|---|---|
Speech | −0.31* | −0.19 | −0.33* | −0.41** | −0.20 | −0.06 | −0.10 | −0.49*** |
Age | 0.10 | 0.11 | 0.17 | 0.22 | 0.15 | 0.10 | −0.19 | 0.21 |
PTA | 0.20 | 0.009 | 0.40** | 0.46** | −0.04 | −0.008 | 0.15 | 0.56*** |
Correlation coefficients in boldface indicate significant results (**p ≤ 0.05; ***p ≤ 0.005). Correlation coefficients in italics are marginally significant (*p ≤ 0.07).
This robust relationship between PTA and the envelope coding metrics in the right hemisphere can account for the overall right hemisphere lateralization in the comparisons of envelope coding in NH and SNHL listeners (Fig. 4D–G). When the effects of PTA were controlled for in the regression analysis, envelope coding metrics in the right hemisphere were not linked with speech identification scores. However, the relationship between speech identification scores and the amplitude of envelope of coding in left HG persisted because PTA was not strongly related to the amplitude of envelope coding in left HG.
Discussion
The present study establishes a link between the amplitude of the cortical phase-locked response to modulated sounds (envelope coding) and speech identification in the presence of modulated background noise. Both the amplitude (Wilding et al., 2012) and the fidelity (Presacco et al., 2016) of cortical envelope coding were enhanced in listeners with sensorineural hearing loss (SNHL). Our results show that SNHL exerts differential effects on envelope coding in left and right auditory cortices and posteromedial and posterolateral auditory cortices. Specifically, there was an anatomical dissociation in the effects of SNHL on envelope coding metrics in HG and STG: the amplitude of envelope coding was enhanced in HG and the fidelity of envelope coding was enhanced in HG and STG in SNHL listeners compared with NH listeners. Enhanced envelope coding was more evident in right auditory cortex in hearing-impaired listeners with a symmetrical SNHL.
Magnified envelope coding disrupts segregation of speech and modulated noise
The results reported here link deficits in speech identification in modulated background noise with magnified cortical envelope coding. This relationship is consistent with previous work suggesting that “enhanced” envelope coding is not perceptually beneficial for speech identification in a modulated background noise (Moore and Glasberg, 1993). Heinz and colleagues proposed that magnified envelope coding distracts from other auditory cues that could be used to aid speech perception in noise (Kale and Heinz, 2010; Henry et al., 2014; Zhong et al., 2014).
Our results suggest that the envelope coding of both speech and modulated background noise are magnified in SNHL listeners. Speech identification in background noise is partly determined by interactions between the target and masker temporal envelopes (Stone et al., 2012). Magnified envelope coding of both speech and modulated background noise may impair the perception of speech envelope cues, either within the same frequency region as the modulated masker, or in a different frequency region to the modulated masker through the process of modulation discrimination interference (MDI) (Yost et al., 1989; Yost and Sheft, 1989; Bacon and Moore, 1993). MDI may partly arise from perceptual grouping of the target and masker envelopes (Yost and Sheft, 1989; Moore and Shailer, 1992), making it difficult to distinguish the target envelope from the masker envelope.
Sek et al. (2015) measured MDI in NH and SNHL listeners and found that modulation detection thresholds in hearing-impaired listeners were more susceptible to MDI when the masker modulation frequency was the same as the target modulation frequency. Sek et al. (2015) argued that SNHL-induced magnified envelope coding of highly modulated signals, such as those used in the present study, may “saturate” the sensation of fluctuations in the level of modulated sounds. This saturation in the coding of magnified envelopes could facilitate MDI between speech and modulated noise envelopes, thereby disrupting the segregation of speech cues from modulated background noise. We propose that deficits in speech identification in modulated noise in SNHL listeners could arise because magnified envelope coding interrupts the separation of speech from modulated background noise. Our results implicate left HG as an important cortical location for the segregation of speech and modulated background noise.
Effects of SNHL on cortical envelope coding
The effects of SNHL on the magnitude of envelope coding can be linked with physiological changes at the level of the cochlea, that is, a reduction or loss of fast-acting compression that magnifies the perceived fluctuations in the amplitude of modulated sounds in comparison with an NH ear (Moore and Glasberg, 1993; Moore et al., 1996). Therefore, the magnified cortical envelope coding identified here could be inherited from the SNHL-induced changes in the auditory periphery or auditory nerve (Kale and Heinz, 2010, 2012). However, the anatomical dissociations in the effects of SNHL on cortical envelope coding metrics identified in the present study suggest some influence of SNHL on central auditory processing. Specifically, the functional asymmetry in the right-lateralized enhanced envelope coding shown here presumably occurs at the cortical level (Poeppel, 2003; Luo and Poeppel, 2007), adding to the increasing evidence that auditory pathology results in changes in the central auditory system (Morita et al., 2003; Wienbruch et al., 2006; Wilding et al., 2012; Alain et al., 2014; Auerbach et al., 2014).
The present results reveal that the fidelity of cortical envelope coding was also enhanced in SNHL listeners compared with NH listeners. Fast-acting compression present in the basilar membrane of NH listeners distorts the shape of envelope coding (Stone and Moore, 2007). A reduction or loss of fast-acting compression in SNHL listeners may reduce the distortions introduced into the coding of the temporal envelope of sounds at the level of the basilar membrane, resulting in increased fidelity of envelope coding in SNHL listeners. Alternatively, the broadened auditory filters of SNHL listeners may reduce the influence of the inherent fluctuations present in noise maskers (Oxenham and Kreft, 2014; Stone and Moore, 2014), leading to the improved fidelity of envelope coding measured in SNHL listeners in the present study.
Summary
Understanding speech in modulated noise is particularly difficult for hearing-impaired listeners with SNHL. We used MEG to investigate how modulated noise is represented in auditory cortex of NH and hearing-impaired listeners. Enhanced envelope coding was associated with a perceptual deficit: The amplitude of cortical envelope coding of modulated noise was linked with speech identification in the presence of modulated noise. Magnified cortical envelope coding may cause deficits in speech perception in modulated background noise because it disrupts the segregation of speech from a modulated background noise.
Footnotes
This work was supported by the Wellcome Trust (Grant C2D2 ISSF 097829 to S.L.M.) through the Centre for Chronic Diseases and Disorders (C2D2) at the University of York, UK.
The authors declare no competing financial interests.
References
- Abrams DA, Nicol T, Zecker S, Kraus N (2008) Right-hemisphere auditory cortex is dominant for coding syllable patterns in speech. J Neurosci 28:3958–3965. 10.1523/JNEUROSCI.0187-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Agrawal Y, Platz EA, Niparko JK (2008) Prevalence of hearing loss and differences by demographic characteristics among US adults: data from the National Health and Nutrition Examination Survey. Arch Intern Med 168:1522–1530. 10.1001/archinte.168.14.1522 [DOI] [PubMed] [Google Scholar]
- Alain C, Roye A, Salloum C (2014) Effects of age-related hearing loss and background noise on neuromagnetic activity from auditory cortex. Front Syst Neurosci 8:8. 10.3389/fnsys.2014.00008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Auerbach BD, Rodrigues PV, Salvi RJ (2014) Central gain control in tinnitus and hyperacusis. Front Neurol 5:206. 10.3389/fneur.2014.00206 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bacon SP, Moore BC (1993) Modulation detection interference: Some spectral effects. J Acoust Soc Am 93:3442–3453. 10.1121/1.405674 [DOI] [PubMed] [Google Scholar]
- Bacon SP, Opie JM, Montoya DY (1998) The effects of hearing loss and noise masking on the masking release for speech in temporally complex backgrounds. J Speech Lang Hear Res 41:549–563. 10.1044/jslhr.4103.549 [DOI] [PubMed] [Google Scholar]
- Bernstein JG, Grant KW (2009) Auditory and auditory-visual intelligibility of speech in fluctuating maskers for normal-hearing and hearing-impaired listeners. J Acoust Soc Am 125:3358–3372. 10.1121/1.3110132 [DOI] [PubMed] [Google Scholar]
- Bowerman BL, O'Connell RT (1990) Linear statistical models: an applied approach, Ed 2 Belmont, CA: Duxbury. [Google Scholar]
- British Society of Audiology (2004) Recommended procedure: Pure tone air and bone conduction thresholds audiometry with and without masking and determination of uncomfortable loudness levels. Reading, UK: British Society of Audiology. [Google Scholar]
- Brookes MJ, Gibson AM, Hall SD, Furlong PL, Barnes GR, Hillebrand A, Singh KD, Holliday IE, Francis ST, Morris PG (2004) A general liner model for MEG beamformer imaging. Neuroimage 23:936–946. 10.1016/j.neuroimage.2004.06.031 [DOI] [PubMed] [Google Scholar]
- Ding N, Simon JZ (2012) Emergence of neural encoding of auditory objects while listening to competing speakers. Proc Natl Acad Sci U S A 109:11854–11859. 10.1073/pnas.1205381109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dubno JR, Horwitz AR, Ahlstrom JB (2002) Benefit of modulated maskers for speech recognition by younger and older adults with normal hearing. J Acoust Soc Am 111:2897–2907. 10.1121/1.1480421 [DOI] [PubMed] [Google Scholar]
- Duquesnoy AJ. (1983) Effect of a single interfering noise or speech source on the binaural sentence intelligibility of aged persons. J Acoust Soc Am 74:739–743. 10.1121/1.389859 [DOI] [PubMed] [Google Scholar]
- Festen JM, Plomp R (1990) Effects of fluctuating noise and interfering speech on the speech-reception threshold for impaired and normal hearing. J Acoust Soc Am 88:1725–1736. 10.1121/1.400247 [DOI] [PubMed] [Google Scholar]
- Füllgrabe C, Moore BC, Stone MA (2014) Age-group differences in speech identification despite matched audiometrically normal hearing: contributions from auditory temporal processing and cognition. Front Aging Neurosci 6:347. 10.3389/fnagi.2014.00347 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gregan MJ, Nelson PB, Oxenham AJ (2013) Behavioural measures of cochlear compression and temporal resolution as predictors of speech masking release in hearing-impaired listeners. J Acoust Soc Am 134:2895–2912. 10.1121/1.4818773 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grose JH, Mamo SK, Hall JW 3rd (2009) Age effects in temporal envelope processing: speech unmasking and auditory steady state responses. Ear Hear 30:568–575. 10.1097/AUD.0b013e3181ac128f [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hartwigsen G, Golombek T, Obleser J (2015) Repetitive transcranial magnetic stimulation over left angular gyrus modulates the predictability gain in degraded speech comprehension. Cortex 68:100–110. 10.1016/j.cortex.2014.08.027 [DOI] [PubMed] [Google Scholar]
- Henry KS, Kale S, Heinz MG (2014) Noise-induced hearing loss increases the temporal precision of complex envelope coding by auditory-nerve fibers. Front Syst Neurosci 8:20. 10.3389/fnsys.2014.00020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hopkins K. (2009) The role of temporal fine structure information in the perception of complex sounds for normal-hearing and hearing-impaired subjects. Ph.D. Thesis, University of Cambridge, Cambridge, United Kingdom. [Google Scholar]
- Huang MX, Mosher JC, Leahy RM (1999) A sensor-weighted overlapping-sphere head model and exhaustive head model comparison for MEG. Phys Med Biol 44:423–440. 10.1088/0031-9155/44/2/010 [DOI] [PubMed] [Google Scholar]
- Huang MX, Shih JJ, Lee RR, Harrington DL, Thoma RJ, Weisend MP, Hanlon F, Paulson KM, Li T, Martin K, Millers GA, Canive JM (2004) Commonalities and differences among vectorised beamformers in electromagnetic source imaging. Brain Topogr 16:139–158. [DOI] [PubMed] [Google Scholar]
- Johnson S, Prendergast G, Hymers M, Green G (2011) Examining the effects of one- and three-dimensional spatial filtering analyses in magnetoencephalography. PLoS One 6:e22251. 10.1371/journal.pone.0022251 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kale S, Heinz MG (2010) Envelope coding in auditory nerve fibers following noise-induced hearing loss. J Assoc Res Otolaryngol 11:657–673. 10.1007/s10162-010-0223-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kale S, Heinz MG (2012) Temporal modulation transfer functions measured from auditory-nerve responses following sensorineural hearing loss. Hear Res 286:64–75. 10.1016/j.heares.2012.02.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kozinska D, Carducci F, Nowinski K (2001) Automatic alignment of EEG/MEG and MRI data sets. Clin Neurophysiol 112:1553–1561. 10.1016/S1388-2457(01)00556-9 [DOI] [PubMed] [Google Scholar]
- Luo H, Poeppel D (2007) Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex. Neuron 54:1001–1010. 10.1016/j.neuron.2007.06.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Millman RE, Prendergast G, Hymers M, Green GG (2013) Representations of the temporal envelope of sounds in human auditory cortex: Can the results from invasive intracortical “depth” electrode recordings be replicated using non-invasive MEG “virtual electrodes”? Neuroimage 64:185–196. 10.1016/j.neuroimage.2012.09.017 [DOI] [PubMed] [Google Scholar]
- Millman RE, Johnson SR, Prendergast G (2015) The role of phase-locking to the temporal envelope of speech in auditory perception and speech intelligibility. J Cogn Neurosci 27:533–545. 10.1162/jocn_a_00719 [DOI] [PubMed] [Google Scholar]
- Moore BC. (2008) The role of temporal fine structure processing in pitch perception, masking, and speech perception for normal-hearing and hearing-impaired people. JARO 9:399–406. 10.1007/s10162-008-0143-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Moore BC, Glasberg BR (1993) Simulation of the effects of loudness recruitment and threshold elevation on the intelligibility of speech in quiet and in a background of speech. J Acoust Soc Am 94:2050–2062. 10.1121/1.407478 [DOI] [PubMed] [Google Scholar]
- Moore BC, Glasberg BR (1998) Use of a loudness model for hearing-aid fitting. I. Linear hearing aids. Br J Audiol 32:317–335. 10.3109/03005364000000083 [DOI] [PubMed] [Google Scholar]
- Moore BC, Shailer MJ (1992) Modulation discrimination interference and auditory grouping. Philos Trans R Soc Lond B Biol Sci 336:339–346. 10.1098/rstb.1992.0067 [DOI] [PubMed] [Google Scholar]
- Moore BC, Wojtczak M, Vickers DA (1996) Effect of loudness recruitment on the perception of amplitude modulation. J Acoust Soc Am 100:481–489. 10.1121/1.415861 [DOI] [Google Scholar]
- Moore BC, Peters RW, Stone MA (1999) Benefits of linear amplification and multichannel compression for speech comprehension in backgrounds with spectral and temporal dips. J Acoust Soc Am 105:400–411. 10.1121/1.424571 [DOI] [PubMed] [Google Scholar]
- Morita T, Naito Y, Nagamine T, Fujiki N, Shibasaki H, Ito J (2003) Enhanced activation of the auditory cortex in patients with inner-ear hearing impairment: a magnetoencephalographic study. Clin Neurophysiol 114:851–859. 10.1016/S1388-2457(03)00033-6 [DOI] [PubMed] [Google Scholar]
- Nourski KV, Reale RA, Oya H, Kawasaki H, Kovach CK, Chen H, Howard MA 3rd, Brugge JF (2009) Temporal envelope of time-compressed speech represented in the human auditory cortex. J Neurosci 29:15564–15574. 10.1523/JNEUROSCI.3065-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nourski KV, Brugge JF, Reale RA, Kovach CK, Oya H, Kawasaki H, Jenison RL, Howard MA 3rd (2013) Coding of repetitive transients by auditory cortex on posterolateral superior temporal gyrus in humans: an intrancranial electrophysiology study. J Neurophysiol 109: 1283–1295. 10.1152/jn.00718.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oxenham AJ, Kreft HA (2014) Speech perception in tone and noise via cochlear implants reveals influence of spectral resolution on temporal processing. Trends Hear 18. 10.1177/2331216514553783 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Poeppel D. (2003) The analysis of speech in different temporal integration windows: cerebral lateralization as “asymmetric sampling in time”. Speech Commun 41:245–255. 10.1016/S0167-6393(02)00107-3 [DOI] [Google Scholar]
- Prendergast G, Johnson JR, Green GGR (2011) Extracting amplitude modulations from speech in the time domain. Speech Commun 53:903–913. 10.1016/j.specom.2011.03.002 [DOI] [Google Scholar]
- Presacco A, Simon JZ, Anderson S (2016) Evidence of degraded representation of speech in noise, in the aging midbrain and cortex. J Neurophysiol 116:2346–2355. 10.1152/jn.00372.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rademacher J, Morosan P, Schormann T, Schleicher A, Werner C, Freund HJ, Zilles K (2001) Probabilistic mapping and volume measurement of human primary auditory cortex. Neuroimage 13:669–683. [DOI] [PubMed] [Google Scholar]
- Rothauser E, Chapman W, Guttman N, Silbiger H, Hecker M, Urbanek G, Nordby K, Weinstock M (1969) IEEE recommended practice for speech quality measurements. IEEE Transactions in Acoustics 17:225–246. [Google Scholar]
- Scott SK, Rosen S, Beaman CP, Davis JP, Wise RJ (2009) The neural processing of masked speech: Evidence for different mechanisms in the left and right temporal lobes. J Acoust Soc Am 125:1737–1743. 10.1121/1.3050255 [DOI] [PubMed] [Google Scholar]
- Sek A, Baer T, Crinnion W, Springgay A, Moore BC (2015) Modulation masking within and across carriers for subjects with normal and impaired hearing. J Acoust Soc Am 138:1143–1153. 10.1121/1.4928135 [DOI] [PubMed] [Google Scholar]
- Simon JZ. (2015) The encoding of auditory objects in auditory cortex: Insights from magnetoencephalography. Int J Psychophysiol 95:184–190. 10.1016/j.ijpsycho.2014.05.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stone MA, Moore BC (2007) Quantifying the effects of fast-acting compression on the envelope of speech. J Acoust Soc Am 121:1654–1664. 10.1121/1.2434754 [DOI] [PubMed] [Google Scholar]
- Stone MA, Moore BC (2014) Amplitude-modulation detection by recreational-noise-exposed humans with near-normal hearing thresholds and its medium-term progression. Hear Res 317:50–62. 10.1016/j.heares.2014.09.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stone MA, Füllgrabe C, Moore BC (2012) Notionally steady background noise acts primarily as a modulation masker of speech. J Acoust Soc Am 132:317–326. 10.1121/1.4725766 [DOI] [PubMed] [Google Scholar]
- Studebaker GA. (1985) A “rationalized” arcsine transform. J Speech Hear Res 28:455–462. 10.1044/jshr.2803.455 [DOI] [PubMed] [Google Scholar]
- Van Veen BD, van Drongelen W, Yuchtman M, Suzuki A (1997) Localization of brain electrical activity via linearly constrained minimum variance spatial filtering. IEEE Trans Biomed Eng 44:867–880. 10.1109/10.623056 [DOI] [PubMed] [Google Scholar]
- Vickers DA, Moore BC, Baer T (2001) Effects of low-pass filtering on the intelligibility of speech in quiet for people with and without dead regions at high frequencies. J Acoust Soc Am 110:1164–1175. 10.1121/1.1381534 [DOI] [PubMed] [Google Scholar]
- Wienbruch C, Paul I, Weisz N, Elbert T, Roberts LE (2006) Frequency organisation of the 40 Hz auditory steady-state response in normal hearing and in tinnitus. Neuroimage 33:180–194. 10.1016/j.neuroimage.2006.06.023 [DOI] [PubMed] [Google Scholar]
- Wilding TS, McKay CM, Baker RJ, Kluk K (2012) Auditory steady-state responses in normal-hearing and hearing-impaired adults: An analysis of between-session amplitude and latency repeatability, test time, and F ratio detection paradigms. Ear Hear 33:267–278. 10.1097/AUD.0b013e318230bba0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wong PC, Jin JX, Gunasekera GM, Abel R, Lee ER, Dhar S (2009) Aging and cortical mechanisms of speech perception in noise. Neuropsychologia 47:693–703. 10.1016/j.neuropsychologia.2008.11.032 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Yost WA, Sheft S (1989) Across-critical-band processing of amplitude-modulated tones. J Acoust Soc Am 85:848–857. 10.1121/1.397556 [DOI] [PubMed] [Google Scholar]
- Yost WA, Sheft S, Opie J (1989) Modulation interference in detection and discrimination of amplitude modulation. J Acoust Soc Am 86:2138–2147. 10.1121/1.398474 [DOI] [PubMed] [Google Scholar]
- Zhong Z, Henry KS, Heinz MG (2014) Sensorineural hearing loss amplifies neural coding of envelope information in the central auditory system of chinchillas. Hear Res 309:55–62. 10.1016/j.heares.2013.11.006 [DOI] [PMC free article] [PubMed] [Google Scholar]