Abstract
This electrophysiological study investigated the role of the medial olivocochlear (MOC) efferents in listening in noise. Both ears of eleven normal-hearing adult participants were tested. The physiological tests consisted of transient-evoked otoacoustic emission (TEOAE) inhibition and the measurement of cortical event-related potentials (ERPs). The mismatch negativity (MMN) and P300 responses were obtained in passive and active listening tasks, respectively. Behavioral responses for the word recognition in noise test were also analyzed. Consistent with previous findings, the TEOAE data showed significant inhibition in the presence of contralateral acoustic stimulation. However, performance in the word recognition in noise test was comparable for the two conditions (i.e., without contralateral stimulation and with contralateral stimulation). Peak latencies and peak amplitudes of MMN and P300 did not show changes with contralateral stimulation. Behavioral performance was also maintained in the P300 task. Together, the results show that the peripheral auditory efferent effects captured via otoacoustic emission (OAE) inhibition might not necessarily be reflected in measures of central cortical processing and behavioral performance. As the MOC effects may not play a role in all listening situations in adults, the functional significance of the cochlear effects of the medial olivocochlear efferents and the optimal conditions conducive to corresponding effects in behavioral and cortical responses remain to be elucidated.
Keywords: event-related potential (ERP), medial olivocochlear (MOC) efferents, otoacoustic emissions inhibition, contralateral acoustic stimulation (CAS), MMN, P300
1. Introduction
Human auditory perception depends on elaborate neural coding of acoustic properties of the target sounds and involves bidirectional interactions of the afferent and efferent systems along the auditory pathway. The neural coding of auditory stimulation begins in the afferent system of the cochlear inner hair cells on the basilar membrane, which is regulated by the active mechanical behavior of outer hair cells under the control of a complex efferent innervation system originating from the medial olivary complex in the brainstem. There has been considerable discussion regarding the role of the medial olivocochlear (MOC) efferents in auditory perception [1,2]. A major function of the MOC efferents is thought to be the enhancement of transient signals in noise. However, it remains unclear how the activation of the MOC system by ipsilateral and/or contralateral stimulation contributes to the enhanced central processing of the auditory target either autonomously or via interactions with attentional control. The present electrophysiological study was designed to address this question by selecting listening conditions known to be conducive to MOC activation and investigating how it affects cortical-level responses in passive and attentive listening tasks.
Descending efferent neurons from the MOC system synapse on cochlear outer hair cells (OHCs), the source of the cochlear amplifier, and influence their function. The MOC efferents may be activated by the corticofugal system, which extends from the cortex toward subcortical nuclei [3], or by acoustic stimulation of the ear [4]. The MOC efferents exert control in the cochlea by inhibiting OHC activity, thereby reducing basilar membrane vibration. This inhibition occurs in quiet [5,6,7] and in the presence of background noise, leading to reductions in neural fiber responses to noise and enhancing neural responses to transient signals [8,9,10]. This is referred to as the “antimasking effect” or “unmasking effect” [6,11]. The efferent “antimasking” effect is thought to have important implications for the detection of tones and tonal complexes in noise [12,13,14]. Other important impacts of the MOC system arise from its role in the process of learning [15,16,17,18] and its interactions with higher cognitive functions such as attention [17,19]. MOC inhibition may also have a protective effect in the cochlea by reducing the impact of acoustic trauma [20,21].
The beneficial link between MOC inhibition and speech perception remains controversial [15,22,23,24,25,26,27,28]. One complication here is that behavioral results and higher cognitive processing such as speech perception are influenced by several factors [15,22]. For instance, Mishra and Lutman [25] suggest that although the “unmasking” effect may be documented for speech stimuli, it may not be related to MOC effects measured under artificial conditions. In particular, the MOC system may not be used in natural listening conditions. Age may also play a role, with young children being more dependent on the efferent system for listening in noise compared with older children (>9 years) and adults [29]. The MOC-mediated mechanisms in central auditory processing may be recruited only in some specific conditions that remain to be investigated [25].
Auditory event-related potentials (ERPs) provide an avenue for investigating the neurophysiological underpinnings of behavioral effects and attentional processing, which influence the perception of complex signals in noise. ERPs are non-invasive recordings of electrical voltages created due to synchronous post-synaptic cortical activity time-locked to an auditory stimulus. ERPs reflect specific sensory and/or cognitive processes [30]. Specific ERPs that may be used to study the perception of signals in noise are the N1, P2, mismatch negativity (MMN) and P300 responses [31,32,33,34,35,36,37,38,39,40]. Of particular interest to the current study are the MMN and P300 components. MMN occurs as a negative wave in the frontal region of the brain and is elicited when there is a change in a repeated auditory signal. Typically, an oddball paradigm is used with “standard” stimuli occurring repeatedly and “deviant” or target stimuli randomly interspersed. The MMN occurs even when subjects are not attending to stimuli and thus represents an automatic change-detection process. The MMN is sensitive to small stimulus changes and seems to correlate with behavioral discrimination thresholds in normal subjects as well as in clinical populations [41]. By contrast, the P300 is a positive waveform, which peaks around 300 ms and is enhanced when subjects attend to and detect deviants. The P300 is associated with the cognitive processes of stimulus evaluation and categorization. The amplitude of P300 is proportional to the amount of attentional resources required for stimulus processing [42], with latency reflecting stimulus evaluation [43].
In this exploratory study, we used cortical potentials to study the consequences of MOC activation. MOC efferent effects in humans can be measured noninvasively using spontaneous otoacoustic emissions (OAEs) [44] or evoked OAEs [19,25,45,46,47,48], which are by-products of the cochlear amplifier mechanism. The MOC activity measured in one ear may be influenced by incoming sound at either ear, due to ipsilateral, contralateral and bilateral brainstem circuits. The most widely used procedure is to compare the amplitudes of click-evoked OAEs with and without contralateral broadband noise, also called “contralateral acoustic stimulation” (CAS), to document the reduction in the OAE amplitude when CAS is introduced. This contralateral inhibition of OAEs or the “MOC reflex” [49] has been used as a measure of MOC inhibition strength. For the present study, the transient-evoked otoacoustic emissions (TEOAEs) were recorded with and without CAS to document the MOC reflex in participants. This is also the most frequently used protocol for measuring the effects of the MOC reflex [11,12,13,14,15,16,17]. The ERP tasks for our study were designed with and without CAS to parallel the OAE paradigm. ERPs were obtained in an active P300 experimental task and a passive MMN task. Signals were embedded in ipsilateral noise; contralateral noise was used to elicit a greater MOC effect. All the stimulus levels and characteristics were chosen to optimize the recruitment of MOC efferent effects.
Our study was built upon two recent speculations: (1) individuals may not be using the MOC unmasking effect for listening in all speech-in-noise situations, and (2) the cochlear effects of MOC function may not be reflected in tasks that require higher-level processes such as attention. We explored the links between MOC function and listening in noise by obtaining combined electrophysiological and behavioral measurements while stimulating the MOC system. If target recognition is facilitated by the efferent unmasking effect, the modulation of the MMN would be expected in the non-attentive listening condition. In the active listening condition, however, the unmasking effect would not necessarily be reflected in the P300 for attention-related processing. Traditional word recognition in noise testing was also completed for the same participants for comparison.
2. Materials and Methods
2.1. Participants and Qualification Testing
Data obtained from eleven individuals (10 females, mean age of all participants = 22.3 years, SD = 1.8 years, range = 19–26 years) were analyzed. The data obtained from two participants in addition to the eleven could not be included for analysis due to the noisy measurements obtained on one or more physiological tests (OAEs and ERPs). A negative history of speech, language or cognitive difficulties was established through case history interview. The Human Subjects’ Protection Program at the University of Minnesota approved the study (ethical research approval code: 1308M41281), and all the participants provided informed consent. The participants were scheduled for two 2-hr sessions and were paid for their time. The first session consisted of a comprehensive audiologic examination, including otoscopy, pure-tone air and bone conduction testing, speech reception threshold testing, and word recognition testing using the Central Institute for the Deaf (CID) W-22 word lists. Participants were tested behaviorally on nonsense word recognition. Otoacoustic emissions were also recorded during the first session. The second session consisted of ERP recordings in both passive and active conditions to record the MMN and P300 ERPs, respectively.
All the participants were right-handed as determined by the Edinburgh Handedness Inventory [50]. The participants had hearing within normal limits as revealed by audiometric thresholds at octave frequencies ranging from 250 to 8000 Hz in both ears. The average thresholds and standard deviations at the different octave frequencies in the two ears are presented in Table 1.
Table 1.
Ear | 250 Hz | 500 Hz | 1000 Hz | 2000 Hz | 4000 Hz | 8000 Hz |
---|---|---|---|---|---|---|
Right | 5 (4.9) | 4.2 (3.9) | 2.3 (4.7) | 2.3 (7.5) | 3.8 (7.2) | 2.6 (7.8) |
Left | 5 (8.3) | 4.6 (5.7) | 3.5 (4.3) | 2.3 (7.9) | 1.9 (6.1) | 2.3 (8.5) |
The speech reception thresholds (SRTs) measured using spondees and presented live-voice [51] were consistent with pure-tone averages bilaterally. The word recognition scores (WRSs) obtained with the recorded W-22 lists [52] were excellent (average WRS for right ear = 95.27%, SD = 3.44%; average WRS for left ear = 95.27%, SD = 2.93%) at 30 dB Sensation Level (SL) with reference to the SRT (average right ear = 1.82 dB Hearing Level (HL), SD = 2.52 dB; average left ear = 2.72 dB HL, SD = 2.61 dB). The SRT and WRS results were used as qualification criteria to ensure good test reliability and adequate word recognition abilities in quiet, respectively.
The CAS used to elicit the MOC reflex was broadband and was presented from a portable screening audiometer (Beltone 120, Beltone Electronics Company, Glenview, IL, USA) through an Etymotic Research ER-3A (Etymotic Research Inc, Elk Grove Village, IL, USA) insert headphone. The portable audiometer was calibrated to ANSI-ASA standard S3.6-2010 [53]. The spectrum from 250 to 5000 Hz was within 5 dB of the level at 1000 Hz. The thresholds for the broadband noise (BBN) stimulus were found using the modified Hughson-Westlake procedure. The broadband noise was presented at 30 dB SLwith reference to the behavioral threshold for all conditions using CAS. For all participants, the BBN was between 50 and 55 dBA SPL when measured using a sound level meter. Middle-ear acoustic reflexes were not obtained to the different stimuli used in this experiment due to time and equipment constraints. While this level is below the average middle-ear reflex threshold for young adults [54], it has been used in similar studies [29].
2.2. OAEs
Click-evoked OAEs were recorded using the Biologic Scout OAE system (2009) (Natus Medical Incorporated, Mundelein, IL, USA). Clicks were presented at 70 dB peSPL at a presentation rate of 11.72 per second. The responses to 2000 clicks were averaged. The clicks were presented in two conditions in each ear: (1) in quiet and (2) in the presence of CAS (broadband noise) presented at 30 dB SL with reference to the BBN behavioral threshold. The recordings for each listening condition were repeated three times per ear, and the results were averaged. The participants were asked to relax in order to obtain good measurements. The filter range was 800 to 6000 Hz. The responses between 3.5 ms and 16.6 ms were analyzed. The total amplitudes of the click-evoked OAEs were calculated by summing the energy in the fast Fourier transform spectrum between 800 and 4500 Hz. The amplitudes of the click-evoked OAEs were compared between ears and conditions.
2.3. Listening Conditions for Word Recognition Testing and Event-Related Potentials
Four listening conditions were used during word recognition testing and the ERP experiments: (1) the stimuli were presented to the right ear with ipsilateral noise at a +10 dB signal-to-noise ratio (SNR) without CAS; (2) the same stimuli were presented to the right ear with the addition of CAS; (3) the stimuli were presented to the left ear with ipsilateral noise at a +10 dB SNR, without CAS; and (4) the stimuli were presented to the left ear with CAS. Therefore, Conditions 2 and 4 were similar to Conditions 1 and 3, respectively, but with CAS. Stimuli were presented to the right ear in Conditions 1 and 2 and to the left ear in Conditions 3 and 4. The conditions were randomized throughout the experiment for each participant. The ipsilateral and contralateral noises used in this study were uncorrelated, to prevent the effects of central masking and to reduce masking level difference. A quiet condition was not used, as the premise of the testing was the perception of signals in noise with and without CAS.
2.4. Word Recognition in Noise Test
The Nonsense Syllable Test (NST) [55] was used to evaluate monosyllabic nonsense word recognition. This test was chosen to eliminate the effects of semantic content on word recognition. Participants listened to words using TDH 39P headphones and were instructed to repeat back what they heard from the compact-disc recording. Two research assistants knowledgeable in phonetics scored the phoneme responses and averaged the results for a final score. A different list of 25 words was used for each listening condition. Words were presented to the test ear at 30 dB SL with reference to the speech reception threshold (SRT) a +10 dB SNR through an Aurical diagnostic audiometer (GN Otometrics, Schaumburg, IL, USA). Speech-shaped noise was presented continuously through the second channel of the audiometer to the same ear at a +10 dB SNR. As with the other tests in this protocol, CAS was presented continuously at 30 dB SL with reference to the BBN behavioral threshold.
2.5. ERPs
Participants were seated in a comfortable chair in an acoustically and electrically treated booth (IAC systems, Bronx, NY, USA). An oddball paradigm included a pure-tone contrast with frequently repeating 2000 Hz standard stimuli and less frequently occurring 2016 Hz deviant stimuli. The signals were mixed with a white-noise masker presented ipsilaterally at a +10 dB SNR. The signals were presented at 50 dB SL relative to the behavioral detection threshold at 2000 Hz. The signal intensity as measured using a sound level meter was 58 to 62 dB SPL. The deviant of 2016 Hz was chosen based on behavioral pilot experiments in which participants scored 80% accuracy. The pure tone stimuli and masker were created using Sony Sound Forge 10 (version 10, Sony Creative Software, Middleton, WI, USA). The pure tone stimuli were 100 ms in duration and included a 10 ms rise time and 10 ms fall time. The presentation probability was 70% for standards and 30% for deviants. Five blocks of 48 trials were presented with 170 standards and 70 deviants per listening condition. The interstimulus interval between two stimuli was randomized within the range of 900~1100 ms. The stimulus order was randomized within the following two constraints: the deviants must have always been separated by at least two standards, and the first stimulus presented in each block must have been a standard. To measure efferent effects, contralateral broadband noise was presented at 30 dB SL with reference to the BBN behavioral threshold (50 to 55 dB SPL A) using the same portable audiometer used for the OAE and word recognition test protocol.
The same oddball paradigm was used in passive MMN and active P300 conditions. The session always began with the passive MMN blocks to avoid drawing participants’ attention to deviants. In this portion of the experiment, participants were asked to watch a movie of their choice that was played silently with subtitles. In the active P300 task, the participants were asked to press a button to identify the deviant stimulus. Participants were given a short practice block in quiet before being presented with the active listening conditions. Behavioral accuracy and response times were recorded from the participant’s button press responses to calculate the sensitivity index d’ and Criterion [56].
EEG data were collected using the Advanced Neuro Technology (ANT) electroencephalography system and a 64-channel Waveguard cap (ANT Neuro, Enschede, Netherlands) in the Zhang Lab at the University of Minnesota. The impedance of the electrodes was below 15 kΩ. The data were band-pass filtered between 0.016 Hz and 200 Hz and digitized using a sampling frequency of 512 Hz. ERP responses were analyzed off-line using ANT’s Advanced Source Analysis software (version 4.6, ANT Neuro, Enschede, Netherlands). The responses were band-pass filtered off-line between 0.2 Hz and 40 Hz. Trials with potentials exceeding ±50 µV were then rejected. The analysis window was from −100 ms to +600 ms relative to the stimulus onset.
Waveforms were referenced to linked mastoids and analyzed using the peak measures from individual electrodes for amplitude and latency values. The latency ranges used for analysis were 120–350 ms and 200–600 ms for MMN and P300 ERPs, respectively. MMN analysis was conducted for data obtained from frontal, mid-frontal, central and mid-central electrodes. The electrode sites for analysis were chosen based on scalp maps that showed intense activation in these regions. Similar methods for electrode grouping were used in previous ERP studies [57,58,59,60]. The frontal electrodes included F3, F5, F7, FC3, FC5, FT7 and the corresponding electrodes on the right hemisphere. The central electrodes included T7, TP7, C3, C5, CP3, CP5 and the corresponding electrodes on the right hemisphere. The midline frontal electrodes included F1, Fz, F2, FC1, FCz and FC2. The midline central electrodes included C1, Cz, C2, CP1, CPz and CP2. P300 analysis was conducted for central, mid-central, parietal and mid-parietal sites. The parietal electrodes included P3, P5, P7, PO3, PO5, PO7 and the corresponding right-hemisphere electrodes. The midline parietal electrodes included P1, Pz, P2 and POz. All the ERP measures from the 11 subjects were subjected to repeated-measures ANOVA with the ear and condition as independent variables.
3. Results
3.1. OAE Data
Click-evoked OAEs were present at acceptable SNRs in both ears for all the conditions, and the probe stability was always close to 100%. The average amplitudes of the click-evoked OAEs were compared for conditions with and without contralateral masking noise. The mean overall amplitudes for the right and left ears are shown in Figure 1. The amplitude data from the two ears were analyzed using repeated-measures ANOVA with the ear and condition as independent variables. There was a significant ear effect (F(1,10) = 6.68; p = 0.03, ηp2 = 0.40). A main effect of condition was also seen (F(1,10) = 6.87; p = 0.03, ηp2 = 0.41), with lower amplitudes in the presence of CAS. The interaction between the ear and noise was not significant (F(1,10) = 0.02; p = 0.91, ηp2 = 0.002).
3.2. WRS Data
The word recognition scores for the NST calculated for the conditions with and without CAS are presented in Table 2. The main effects (ear: F(1,10) = 0.17; p = 0.67, ηp2 = 0.02; condition: F(1,10) = 1.79; p = 0.21, ηp2 = 0.15) and interaction (ear × interaction: F(1,10) = 0.25; p = 0.63, ηp2 = 0.02) were non-significant.
Table 2.
Ear | Condition | Word Recognition Score in Percentage (SD) |
---|---|---|
Right | Without CAS | 78.3 (3.6) |
With CAS | 76.8 (3.7) | |
Left | Without CAS | 77.4 (5.2) |
With CAS | 77 (3.3) |
3.3. MMN Data for Passive Listening Condition
Grand average waveforms of the MMNs obtained for the two conditions are presented in Figure 2. A repeated-measures ANOVA was performed after selecting electrodes from the frontal, mid-frontal, central and mid-central locations of the scalp, where the MMN was robust. The latencies and amplitudes were subjected to separate repeated-measures ANOVAs with condition and electrode location as factors. Descriptive statistics of the MMN are presented in Table 3, and the repeated-measures ANOVA results, in Table 4. A significant main effect of ear was seen for the MMN latencies, with peak latencies longer for the left ear than for the right ear. The other main effects and interactions were non-significant.
Table 3.
Ear | Condition | MMN | |
---|---|---|---|
Amplitude (µV) (SD) |
Latency (ms) (SD) |
||
Right | Without CAS | −1.22 (0.47) | 227.3 (23.5) |
With CAS | −1.15 (0.34) | 219.65 (18.63) | |
Left | Without CAS | −1.21 (0.33) | 247.72 (20.93) |
With CAS | −1.1 (0.52) | 228.95 (18.84) |
Table 4.
Main Effects and Interactions | MMN Peak Amplitude | MMN Peak Latency |
---|---|---|
Ear |
F(1,10) = 0.145 p = 0.71 ηp2 = 0.014 |
F(1,10) = 6.71 p = 0.03 * ηp2 = 0.40 |
Condition |
F(1,10) = 1.25 p = 0.29 ηp2 = 0.11 |
F(1,10) = 2.11 p = 0.17 ηp2 = 0.175 |
Ear × Condition |
F(1,10) = 0.058 p = 0.81 ηp2 = 0.006 |
F(1,10) = 1.4 p = 0.26 ηp2 = 0.263 |
3.4. P300 Data for Active Listening Condition
Grand average waveforms of the P300s obtained for the two conditions are presented in Figure 3. Descriptive statistics are presented in Table 5. A repeated-measures ANOVA was performed after selecting the central, mid-central, parietal and mid-parietal scalp locations where the P300 was most prominent, and the results are shown in Table 6. Once again, the condition without contralateral noise was compared with the condition with contralateral noise. The Peak P300 latencies and peak amplitudes did not show significant differences between the two conditions.
Table 5.
Ear | Condition | MMN | P300 | ||
---|---|---|---|---|---|
Amplitude (µV) (SD) |
Latency (ms) (SD) |
Amplitude (µV) (SD) |
Latency (ms) (SD) |
||
Right | Without CAS | −1.22 (0.47) | 227.3 (23.5) | 2.09 (0.99) | 420.5 (54.97) |
With CAS | −1.15 (0.34) | 219.65 (18.63) | 1.8 (0.95) | 417.03 (64.5) | |
Left | Without CAS | −1.21 (0.33) | 247.72 (20.93) | 1.99 (0.98) | 409.52 (49.83) |
With CAS | −1.1 (0.52) | 228.95 (18.84) | 1.97 (0.98) | 444.12 (50.45) |
Table 6.
Main Effects and Interactions | P300 Peak Amplitude | P300 Peak Latency |
---|---|---|
Ear |
F(1,10) = 0.022 p = 0.88 ηp2 = 0.002 |
F(1,10) = 0.72 p = 0.416 ηp2 = 0.067 |
Condition |
F(1,10) = 1.20 p = 0.298 ηp2 = 0.107 |
F(1,10) = 0.85 p = 0.378 ηp2 = 0.078 |
Ear × Condition |
F(1,10) = 0.533 p = 0.482 ηp2 = 0.051 |
F(1,10) = 1.65 p = 0.228 ηp2 = 0.142 |
The criterion and d’ were measured from the hits and false alarms obtained in the active discrimination (P300) task. The reaction times from the two conditions (with and without CAS) were also subjected to statistical analysis. Descriptive statistics for the d’, criterion and reaction times are shown in Table 7. None of the measures showed significant differences (Table 8).
Table 7.
Ear | Condition | d’ (SD) | Criterion (SD) | Reaction Times (SD) |
---|---|---|---|---|
Right | Without CAS | 2.07 (1.12) | −0.69 (0.38) | 556.01 (115.66) |
With CAS | 2.09 (1.18) | −0.76 (0.34) | 567.29 (134.59) | |
Left | Without CAS | 2.59 (1.34) | −0.66 (0.5) | 543.18 (121.04) |
With CAS | 2.64 (1.42) | −0.67 (0.43) | 551.19 (110.43) |
Table 8.
Main Effects and Interactions | d’ | Criterion | Reaction Time |
---|---|---|---|
Ear |
F(1,10) = 7.133 p = 0.02 * ηp2 = 0.41 |
F(1,10) = 0.702 p = 0.42 ηp2 = 0.06 |
F(1,10) = 1.24 p = 0.292 ηp2 = 0.11 |
Condition |
F(1,10) = 0.62 p = 0.80 ηp2 = 0.006 |
F(1,10) = 1.81 p = 0.208 ηp2 = 0.153 |
F(1,10) = 0.420 p = 0.53 ηp2 = 0.04 |
Ear × Condition |
F(1,10) = 0.121 p = 0.735 ηp2 = 0.012 |
F(1,10) = 0.708 p = 0.420 ηp2 = 0.066 |
F(1,10) = 0.015 p = 0.91 ηp2 = 0.001 |
4. Discussion
This exploratory study attempted to address an important gap in our understanding of the function of the auditory MOC efferent system in specific listening conditions by assessing how contralateral noise affected behavioral and physiological measures. The effects of efferent suppression on OAE amplitude and word recognition in noise were measured along with ERPs in passive and active tasks. For one physiological measure, namely the otoacoustic emissions elicited by clicks presented in quiet, contralateral noise decreased the level. For the other physiological and behavioral measures, stimuli were presented in ipsilateral noise with and without contralateral noise, and no statistically significant effect of contralateral noise was seen in any condition.
4.1. OAE Inhibition
In this study, click-evoked OAEs were analyzed in the 1000 Hz to 4000 Hz frequency range, where maximum suppression is seen in humans [61,62,63,64]. Animal studies also reveal that mid-to-high frequencies are strongly affected by the activation of the MOC system [11]. The overall amplitudes of the click-evoked OAEs were reduced in the presence of CAS, which is consistent with previous findings, exhibiting an inhibitory effect at the cochlear level [4,45]. The differences in OAE levels were seen as an ear effect, with the right ear showing higher levels compared with the left. The ear effect may be attributed to the asymmetry between the two ears [65], which could be related to handedness [66]. In our study, all the participants were right-handed.
4.2. Word Recognition Testing
In the present study, the nonsense-syllable recognition scores in noise from the nonsense syllable test were unchanged in the presence of CAS. This contrasts with previous findings obtained using similar methodology but with words that carried semantic content. Studies of speech perception have used monosyllables in noise in the past [23,24]. When nonsense syllables are used, listeners have to rely solely on acoustic features due to the lack of linguistic context. These results suggest that the antimasking effect alone is insufficient to lead to improvement in word recognition scores in adults.
In a 2012 study by de Boer et al. [22], phoneme discrimination was uncorrelated with OAE inhibition. Similarly, in another study [15], once subjects had been sufficiently trained in phoneme discrimination, the correlation between OAE inhibition and phoneme discrimination was lost, suggesting a stronger role for attention and/or other central mechanisms. The fact that several other studies have failed to show a relationship between word recognition scores and OAE inhibition may be attributed to differences in the methods used to test MOC effects on word recognition [27]. Modeling of efferent effects [67] shows that optimal speech perception is achieved when the amount of efferent activity is proportional to the level of noise, with the amount of unmasking dependent on both the signal level and noise level. It has also been shown that individuals with stronger MOC efferent responses are more responsive to changes in the SNR [68]. Therefore, it seems that a specific combination of signal level and noise level is required to achieve maximal improvement. Additionally, improvements may be age related [24], with young participants showing greater reliance on efferent function in noise. Those who have received a vestibular neurectomy show nearly normal performance, which points to the role of possible compensatory mechanisms in word recognition [26,66]. In summary, it is improbable that the MOC effect on the cochlea induced using CAS alone at a single SNR would appreciably affect speech intelligibility in ipsilateral noise at the group level.
4.3. ERPs
Although significant OAE inhibition was seen due to contralateral acoustic stimulation, the MMN (passive task) showed no statistically significant difference between the conditions with and without CAS. Obligatory changes that facilitate signal processing in noise can be measured in the cochlea and the cochlear nerve (in terms of changes to cochlear gain and the enhanced coding of signals in noise, respectively), but these may not be reflected at the cortical level in the MMN response. Alternatively, perhaps the changes are reflected at the cortical level but are sufficiently subtle that the MMN was not sensitive enough to capture them—at least when using these particular sound levels. Interestingly, an ear effect was noted for the MMN peak latency, with the latencies for the stimuli presented to the left ear longer than the latencies for the stimuli presented to the right ear. This latency difference suggests longer processing times for stimuli in the left ear. Ear differences have been reported in MMN amplitudes in response to monaural stimulation, especially in patients with various cortical lesions [69]. Interestingly, the left ear was also where the average d’ wsa highest in the P300 behavioral task.
The P300, which is reflective of conscious perception, was not shown by the ANOVA results to be significantly affected by the presence or absence of CAS. Behavioral performance was also essentially identical at the group level regardless of whether CAS was present. Once again, although a peripheral facilitatory mechanism is likely to be operating, its effects were not observed at the cortical level in the presence of attentional processing. As with MMN, the possibility remains that the P300 was simply not sensitive enough to capture the subtle changes transmitted from the cochlear mechanism. Interestingly, the ANOVA of the behavioral results showed a significant ear effect, with d’ being higher when stimuli were presented to the left ear rather than the right. It is possible that the participants were expending greater effort when stimuli were being presented to the left ear, with consequent gains in behavioral outcomes.
Evidence from others [22,29] points to a dynamic relationship between MOC function and central mechanisms during tasks in which individuals are attending to signals in noise. Behavioral output is modulated by interactions between the MOC system and central mechanisms that are attention- and experience-dependent. Indeed, long-term training was found to preserve performance in a signal-in-noise sound localization task in cats with MOC lesions. The authors attributed this to the development of alternate listening strategies, which were able to minimize the functional consequences of the auditory lesions [70]. Therefore, central mechanisms may compensate for significant changes in peripheral function in the processing of complex signals involving redundancy (e.g., speech) or during highly routine tasks (e.g., the psychophysical testing of intensity or frequency discrimination). Alternately, the unmasking provided by the MOC system may not be used at all [16,25]. It has been shown that the efferent system may have a greater role to play during auditory development in early childhood, when the central mechanisms for listening in noise are still immature [29]. The ERP data from the present adult study also support the hypothesis that passive and active central mechanisms may not reflect the facilitatory changes recorded at the periphery in laboratory conditions.
4.4. Methodological Challenges and Lessons
All the physiological measures for our study were obtained with stimuli ideal for eliciting a particular type of response while maintaining the general characteristics of the protocol. For example, transient tone bursts were embedded in white noise to elicit ERPs, and clicks were used to elicit OAEs to record OAE inhibition. Attentional focus was dependent on the task. In the OAE task, the participants were asked to be quiet and relax as fully as possible. In the passive MMN task, the participants were watching a movie; during the active P300 task, the participants’ focus was on the ear being tested. This variation in methodology was inevitable due to the range of physiological indices used. A frequency discrimination task was used to elicit ERPs, as it is an important skill for speech perception in noise [71]. Additionally, the frequency discrimination thresholds of the second formant frequency are reportedly affected by the lesioning of the MOC system in cats [65]. However, we need to acknowledge the possibility that the stimuli and presentation levels chosen, including for the contralateral noise, may be unnatural and work against the original hypotheses. ERP measurements require a considerable amount of the participant’s time, and we had four conditions even without multiple signal levels. Given the time constraint, we chose the most frequently reported SNR in the literature that showed enhancement with syllables or speech. For each condition in our study, only one level for the stimulus and the ipsilateral noise was chosen, which may not have been the optimal levels for demonstrating the associations between MOC effects and central auditory processing [72]. This is particularly important given the fact that the ipsilateral noise would also be evoking the MOC reflex in the presence of considerable within- and across-subject variabilities upon contralateral acoustic stimulation [73].
Another possible limitation is the methodology used to measure OAE inhibition [74]. Stimulus-frequency OAEs (SFOAEs) have been proposed to be superior to click-evoked OAEs in eliciting OAE efferent inhibition, given that the signal levels required to elicit click-evoked or tone-evoked OAEs may themselves cause MOC activation. SFOAEs may be elicited with continuous tones at lower stimulus intensities than click-evoked OAEs. However, commercial systems that measure SFOAEs are not currently available. The subclinical activation of the middle-ear muscle reflex could contaminate MOC reflex effects. We present the caveat that due to the signal levels used in this study, although unlikely, it is possible that the middle-ear reflex may have contaminated the results obtained from couple of participants with low middle-ear reflex activation thresholds [75,76].
As the measures of OAE, word recognition in noise, MMN and P300 were obtained in different test sessions with different sets of stimuli and listening conditions, involving considerable amounts of test time for each of our participants, it is methodologically challenging to test many different signal levels to find out what may be sensitive to the interactions between MOC activity and higher-level cortical processing when listening in noise. Given that the MOC efferent system works both ipsilaterally and contralaterally with much of the innervation being ipsilateral, we cannot rule out the possibility that the ipsilateral MOC could have been fully activated in our behavioral experiment, providing enhanced listening-in-noise recognition. Presumably, this should apply whether or not contralateral noise stimulation was applied. Likewise, if the ipsilateral MOC has achieved full activation in the EEG experiments for delivering the hypothesized anti-masking benefits, there is no reason to expect any additional benefit from adding contralateral noise. Therefore, our conditions may represent a degree of MOC activation (less activation with ipsilateral noise only vs. more activation with added contralateral noise stimulation) as opposed to conditions with vs. without efferent activation. In experiments with human subjects, it is methodologically challenging to implement protocols for introducing a condition with no MOC activation for comparison to verify the contributions of MOC to listening in noise [2].
Changes have been reported in ERPs in the presence of ipsilateral noise compared with a quiet condition [33,34,38,39,77,78,79,80]. Generally, amplitudes are reduced and latencies are prolonged for the MMN and P300 when stimuli are presented in ipsilateral noise, which may be activating the efferent system. As we did not include a quiet condition for comparison, we were unable to capture this effect.
Future studies may need to consider simultaneous experimental protocols that can measure attentional modulation effects in the periphery as well as at the cortical level in the same test session [81]. Nevertheless, the exploratory results reported here could provide a cautionary note to avoid simplistic expectations or interpretation with regard to cortical and medial efferent systems in auditory perception [25]. For instance, one should not erroneously assume that due to the MOC “antimasking” effect, presenting contralateral noise at a single predetermined “optimal” SNR would automatically lead to enhanced ipsilateral speech-in-noise performance. The relationship of MOC efferent activity with higher-level auditory and speech processing is highly dependent on the task characteristics, including the SNRs [72].
4.5. Clinical Relevance and Future Directions
In many clinical populations exhibiting difficulties with speech perception in noise (for example, individuals with learning disabilities and children with auditory processing disorders), inadequate functioning of the MOC system has been documented [65,82,83]. Based on the results of this study, it can be hypothesized that inefficient MOC function is just one piece of the puzzle in these individuals who have not been able to develop alternative listening strategies, as normal individuals do, to compensate for deficiencies in MOC function.
Interestingly, increased MOC activity has been documented in musicians [84]. Greater OAE inhibition has been found in musicians compared with age- and gender-matched normal subjects. Evidence suggests that musicians have superior behavioral speech-in-noise processing skills and brainstem coding [85,86]. An interesting hypothesis is that the enhanced auditory experience in musicians leads to a cumulative strengthening of these mechanisms.
Future studies should be directed toward understanding the role of the MOC system and its interactions with attention, learning and speech perception. Furthermore, information is required about its role during the developmental period [29] and in clinical populations [87,88]. This will inform the optimal procedures to use for the testing and rehabilitation of those who exhibit speech-in-noise difficulties.
5. Conclusions
In summary, this auditory perception study implemented different behavioral and neurophysiological protocols that involved varying degrees of MOC activation. OAE inhibition was seen in the presence of contralateral acoustic stimulation. The recognition of nonsense syllables did not appear to change with the degree of MOC stimulation. Similarly, cortical ERPs as assessed by the MMN and P300 responses in passive and active listening conditions did not reflect the facilitatory effects seen in the contralateral inhibition of OAE at the cochlear level. Our findings are consistent with the view that individuals do not necessarily make use of the available MOC induced unmasking mechanisms for higher-level auditory processing and speech perception in noise, which are highly subject to influence from age-dependent attentional, cognitive and experiential factors [25]. The dissociation patterns demonstrate the limitations of the materials and methods implemented in the present study, which underlines the need for further studies to address the complexity and challenge with suitable protocols to reveal the possible associations and interactions between cortical and MOC efferent mechanisms in auditory perception.
Acknowledgments
We would like to thank Robert D. Melara, Glenis R. Long, Sharon Miller and Luodi Yu for their valuable comments for improving this manuscript.
Abbreviations
The following abbreviations are used in this manuscript:
ANOVA | Analysis of variance |
BBN | Broadband noise |
CAS | Contralateral acoustic stimulation |
CID | Central Institute for the Deaf |
ER | Etymotic research |
ERP | Event-related potentials |
MOC | Medial olivocochlear |
MMN | Mismatch negativity |
NST | Nonsense Syllable Test |
P300 | Positive peak at 300 ms |
OHC | Outer hair cell |
OAE | Otoacoustic emission |
SL | Sensation level |
SPL | Sound pressure level |
SNR | Signal-to-noise ratio |
SRT | Speech reception threshold |
TEOAE | Transient-evoked otoacoustic emissions |
WRS | Word recognition score |
Author Contributions
A.R. and Y.Z. conceived and designed the experiments; A.R., T.K.K. and B.M. performed the experiments and collected the data; A.R., T.K.K. and Y.Z. analyzed the data; A.R. and Y.Z. wrote the paper; all the authors revised the manuscript. All authors have read and agreed to the published version of the manuscript.
Funding
This study was made possible through a grant-in-aid grant from the University of Minnesota to the first author (A.R.). Zhang was additionally supported by the University of Minnesota’s Grand Challenges Exploratory Research Project award.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Smith D.W., Keil A. The biological role of the medial olivocochlear efferents in hearing: Separating evolved function from exaptation. Front. Syst. Neurosci. 2015;9:12. doi: 10.3389/fnsys.2015.00012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Lopez-Poveda E.A. Olivocochlear efferents in animals and humans: From anatomy to clinical relevance. Front. Neurol. 2018;9:197. doi: 10.3389/fneur.2018.00197. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Xiao Z., Suga N. Modulation of cochlear hair cells by the auditory cortex in the mustached bat. Nat. Neurosci. 2002;5:57–63. doi: 10.1038/nn786. [DOI] [PubMed] [Google Scholar]
- 4.Guinan J.J.J. Olivocochlear efferents: Anatomy, physiology, function, and the measurement of efferent effects in humans. Ear Hear. 2006;27:589–607. doi: 10.1097/01.aud.0000240507.83072.e7. [DOI] [PubMed] [Google Scholar]
- 5.Cooper N.P., Guinan J.J.J. Separate mechanical processes underlie fast and slow effects of medial olivocochlear efferent activity. J. Physiol. 2003;548:307–312. doi: 10.1113/jphysiol.2003.039081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Cooper N.P., Guinan J.J.J. Efferent-mediated control of basilar membrane motion. J. Physiol. 2006;576:49–54. doi: 10.1113/jphysiol.2006.114991. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Murugasu E., Russell I.J. The effect of efferent stimulation on basilar membrane displacement in the basal turn of the guinea pig cochlea. J. Neurosci. 1996;16:325–332. doi: 10.1523/JNEUROSCI.16-01-00325.1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Dolan D.F., Nuttall A.L. Masked cochlear whole-nerve response intensity functions altered by electrical stimulation of the crossed olivocochlear bundle. J. Acoust. Soc. Am. 1988;83:1081–1086. doi: 10.1121/1.396052. [DOI] [PubMed] [Google Scholar]
- 9.Kawase T., Delgutte B., Liberman M.C. Antimasking effects of the olivocochlear reflex. II. Enhancement of auditory-nerve response to masked tones. J. Neurophysiol. 1993;70:2533–2549. doi: 10.1152/jn.1993.70.6.2533. [DOI] [PubMed] [Google Scholar]
- 10.Winslow R.L., Sachs M.B. Single-tone intensity discrimination based on auditory-nerve rate responses in backgrounds of quiet, noise, and with stimulation of the crossed olivocochlear bundle. Hear. Res. 1988;35:165–189. doi: 10.1016/0378-5955(88)90116-5. [DOI] [PubMed] [Google Scholar]
- 11.Liberman M.C., Guinan J.J.J. Feedback control of the auditory periphery: Anti-masking effects of middle ear muscles vs. olivocochlear efferents. J. Commun. Disord. 1998;31:471–482. doi: 10.1016/S0021-9924(98)00019-7. [DOI] [PubMed] [Google Scholar]
- 12.Micheyl C., Collet L. Involvement of the olivocochlear bundle in the detection of tones in noise. J. Acoust. Soc. Am. 1996;99:1604–1610. doi: 10.1121/1.414734. [DOI] [PubMed] [Google Scholar]
- 13.Micheyl C., Morlet T., Giraud A.L., Collet L., Morgon A. Contralateral suppression of evoked otoacoustic emissions and detection of a multi-tone complex in noise. Acta Oto-Laryngol. 1995;115:178–182. doi: 10.3109/00016489509139286. [DOI] [PubMed] [Google Scholar]
- 14.Micheyl C., Perrot X., Collet L. Relationship between auditory intensity discrimination in noise and olivocochlear efferent system activity in humans. Behav. Neurosci. 1997;111:801–807. doi: 10.1037/0735-7044.111.4.801. [DOI] [PubMed] [Google Scholar]
- 15.De Boer J., Thornton A.R. Neural correlates of perceptual learning in the auditory brainstem: Efferent activity predicts and reflects improvement at a speech-in-noise discrimination task. J. Neurosci. 2008;28:4929–4937. doi: 10.1523/JNEUROSCI.0902-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Irving S., Moore D.R., Liberman M.C., Sumner C.J. Olivocochlear efferent control in sound localization and experience-dependent learning. J. Neurosci. 2011;31:2493–2501. doi: 10.1523/JNEUROSCI.2679-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Maison S., Micheyl C., Collet L. Influence of focused auditory attention on cochlear activity in humans. Psychophysiology. 2001;38:35–40. doi: 10.1111/1469-8986.3810035. [DOI] [PubMed] [Google Scholar]
- 18.Veuillet E., Magnan A., Ecalle J., Thai-Van H., Collet L. Auditory processing disorder in children with reading disabilities: Effect of audiovisual training. Brain. 2007;130:2915–2928. doi: 10.1093/brain/awm235. [DOI] [PubMed] [Google Scholar]
- 19.Smith S.B., Cone B. The medial olivocochlear reflex in children during active listening. Int. J. Audiol. 2015;54:518–523. doi: 10.3109/14992027.2015.1008105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Maison S.F., Liberman M.C. Predicting vulnerability to acoustic injury with a noninvasive assay of olivocochlear reflex strength. J. Neurosci. 2000;20:4701–4707. doi: 10.1523/JNEUROSCI.20-12-04701.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Maison S.F., Usubuchi H., Liberman M.C. Efferent feedback minimizes cochlear neuropathy from moderate noise exposure. J. Neurosci. 2013;33:5542–5552. doi: 10.1523/JNEUROSCI.5027-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.De Boer J., Thornton A.R., Krumbholz K. What is the role of the medial olivocochlear system in speech-in-noise processing? J. Neurophysiol. 2012;107:1301–1312. doi: 10.1152/jn.00222.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Giraud A.L., Garnier S., Micheyl C., Lina G., Chays A., Chery-Croze S. Auditory efferents involved in speech-in-noise intelligibility. Neuroreport. 1997;8:1779–1783. doi: 10.1097/00001756-199705060-00042. [DOI] [PubMed] [Google Scholar]
- 24.Kumar U.A., Vanaja C.S. Functioning of olivocochlear bundle and speech perception in noise. Ear Hear. 2004;25:142–146. doi: 10.1097/01.AUD.0000120363.56591.E6. [DOI] [PubMed] [Google Scholar]
- 25.Mishra S.K., Lutman M.E. Top-down influences of the medial olivocochlear efferent system in speech perception in noise. PLoS ONE. 2014;9:e85756. doi: 10.1371/journal.pone.0085756. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Scharf B., Magnan J., Collet L., Ulmer E., Chays A. On the role of the olivocochlear bundle in hearing: A case study. Hear. Res. 1994;75:11–26. doi: 10.1016/0378-5955(94)90051-5. [DOI] [PubMed] [Google Scholar]
- 27.Stuart A., Butler A.K. Contralateral suppression of transient otoacoustic emissions and sentence recognition in noise in young adults. J. Am. Acad. Audiol. 2012;23:686–696. doi: 10.3766/jaaa.23.9.3. [DOI] [PubMed] [Google Scholar]
- 28.Zeng F.G., Martino K.M., Linthicum F.H., Soli S.D. Auditory perception in vestibular neurectomy subjects. Hear. Res. 2000;142:102–112. doi: 10.1016/S0378-5955(00)00011-3. [DOI] [PubMed] [Google Scholar]
- 29.Mishra S. The role of medial efferents in human auditory development: Efferent inhibition predicts frequency discrimination in noise for children. J. Neurophysiol. 2020 doi: 10.1152/jn.00136.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Luck S.J. An Introduction to the Event-Related Potential Technique. 2nd ed. MIT Press; Cambridge, MA, USA: 2014. [Google Scholar]
- 31.Koerner T.K., Zhang Y. Effects of background noise on inter-trial phase coherence and auditory N1-P2 responses to speech stimuli. Hear. Res. 2015;328:113–119. doi: 10.1016/j.heares.2015.08.002. [DOI] [PubMed] [Google Scholar]
- 32.Koerner T.K., Zhang Y. Differential effects of hearing impairment and age on electrophysiological and behavioral measures of speech in noise. Hear. Res. 2018;370:130–142. doi: 10.1016/j.heares.2018.10.009. [DOI] [PubMed] [Google Scholar]
- 33.Koerner T.K., Zhang Y., Nelson P.B., Wang B., Zou H. Neural indices of phonemic discrimination and sentence-level speech intelligibility in quiet and noise: A mismatch negativity study. Hear. Res. 2016;339:40–49. doi: 10.1016/j.heares.2016.06.001. [DOI] [PubMed] [Google Scholar]
- 34.Koerner T.K., Zhang Y., Nelson P.B., Wang B., Zou H. Neural indices of phonemic discrimination and sentence-level speech intelligibility in quiet and noise: A P3 study. Hear. Res. 2017;350:58–67. doi: 10.1016/j.heares.2017.04.009. [DOI] [PubMed] [Google Scholar]
- 35.Shepherd D., Lodhia V., Hautus M.J. Electrophysiological indices of amplitude modulated sounds and sensitivity to noise. Int. J. Psychophysiol. 2019;139:59–67. doi: 10.1016/j.ijpsycho.2019.03.005. [DOI] [PubMed] [Google Scholar]
- 36.Gustafson S.J., Billings C.J., Hornsby B.W.Y., Key A.P. Effect of competing noise on cortical auditory evoked potentials elicited by speech sounds in 7- to 25-year-old listeners. Hear. Res. 2019;373:103–112. doi: 10.1016/j.heares.2019.01.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Billings C.J., Tremblay K.L., Stecker G.C., Tolin W.M. Human evoked cortical activity to signal-to-noise ratio and absolute signal level. Hear. Res. 2009;254:15–24. doi: 10.1016/j.heares.2009.04.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Kozou H., Kujala T., Shtyrov Y., Toppila E., Starck J., Alku P., Naatanen R. The effect of different noise types on the speech and non-speech elicited mismatch negativity. Hear. Res. 2005;199:31–39. doi: 10.1016/j.heares.2004.07.010. [DOI] [PubMed] [Google Scholar]
- 39.Martin B.A., Kurtzberg D., Stapells D.R. The effects of decreased audibility produced by high-pass noise masking on N1 and the mismatch negativity to speech sounds /ba/and/da. J. Speech Lang. Hear. Res. 1999;42:271–286. doi: 10.1044/jslhr.4202.271. [DOI] [PubMed] [Google Scholar]
- 40.McCullagh J., Shinn J.B. Auditory P300 in noise in younger and older adults. J. Am. Acad. Audiol. 2018;29:909–916. doi: 10.3766/jaaa.17077. [DOI] [PubMed] [Google Scholar]
- 41.Naatanen R., Paavilainen P., Rinne T., Alho K. The mismatch negativity (MMN) in basic research of central auditory processing: A review. Clin. Neurophysiol. 2007;118:2544–2590. doi: 10.1016/j.clinph.2007.04.026. [DOI] [PubMed] [Google Scholar]
- 42.Polich J. Updating P300: An integrative theory of P3a and P3b. Clin. Neurophysiol. 2007;118:2128–2148. doi: 10.1016/j.clinph.2007.04.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Donchin E., Coles M.G.H. Is the P300 component a manifestation of context updating. Behav. Brain Sci. 1988;11:357–374. doi: 10.1017/S0140525X00058027. [DOI] [Google Scholar]
- 44.Mott J.B., Norton S.J., Neely S.T., Warr W.B. Changes in spontaneous otoacoustic emissions produced by acoustic stimulation of the contralateral ear. Hear. Res. 1989;38:229–242. doi: 10.1016/0378-5955(89)90068-3. [DOI] [PubMed] [Google Scholar]
- 45.Collet L., Kemp D.T., Veuillet E., Duclaux R., Moulin A., Morgon A. Effect of contralateral auditory stimuli on active cochlear micro-mechanical properties in human subjects. Hear. Res. 1990;43:251–261. doi: 10.1016/0378-5955(90)90232-E. [DOI] [PubMed] [Google Scholar]
- 46.Henin S., Thompson S., Abdelrazeq S., Long G.R. Changes in amplitude and phase of distortion-product otoacoustic emission fine-structure and separated components during efferent activation. J. Acoust. Soc. Am. 2011;129:2068–2079. doi: 10.1121/1.3543945. [DOI] [PubMed] [Google Scholar]
- 47.Lilaonitkul W., Guinan J.J.J. Human medial olivocochlear reflex: Effects as functions of contralateral, ipsilateral, and bilateral elicitor bandwidths. J. Assoc. Res. Otolaryngol. 2009;10:459–470. doi: 10.1007/s10162-009-0163-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Mishra S.K., Lutman M.E. Repeatability of click-evoked otoacoustic emission-based medial olivocochlear efferent assay. Ear Hear. 2013;34:789–798. doi: 10.1097/AUD.0b013e3182944c04. [DOI] [PubMed] [Google Scholar]
- 49.Guinan J.J.J. Cochlear efferent innervation and function. Curr. Opin. Otolaryngol. Head Neck Surg. 2010;18:447–453. doi: 10.1097/MOO.0b013e32833e05d6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Oldfield R.C. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia. 1971;9:97–113. doi: 10.1016/0028-3932(71)90067-4. [DOI] [PubMed] [Google Scholar]
- 51.American Speech-Language-Hearing Association Guidelines for determining the threshold levels for speech. ASHA. 1988;30:85–89. [PubMed] [Google Scholar]
- 52.Hirsh I.J., Davis H., Silverman S.R., Reynolds E.G., Eldert E., Benson R.W. Development of materials for speech audiometry. J. Speech Hear. Disord. 1952;17:321–337. doi: 10.1044/jshd.1703.321. [DOI] [PubMed] [Google Scholar]
- 53.American National Standards Institute . ANSI S3.6-2010. ANSI; Washington, DC, USA: 2010. Specification for Audiometers. [Google Scholar]
- 54.Johnsen N.J., Terkildsen K. The normal middle-ear reflex thresholds towards white noise and acoustic clicks in young-adults. Scand. Audiol. 1980;9:131–135. doi: 10.3109/01050398009076346. [DOI] [PubMed] [Google Scholar]
- 55.Edgerton B.J., Danhauer J.L. Clinical Implications of Speech Discrimination Testing Using Nonsense Stimuli. University Park Press; Baltimore, MD, USA: 1979. [Google Scholar]
- 56.Macmillan N.A., Creelman C.D. Detection Theory: A User’s Guide. Cambridge University Press; New York, NY, USA: 1991. [Google Scholar]
- 57.Rao A., Zhang Y., Miller S. Selective listening of concurrent auditory stimuli: An event-related potential study. Hear. Res. 2010;268:123–132. doi: 10.1016/j.heares.2010.05.013. [DOI] [PubMed] [Google Scholar]
- 58.Stevens J., Zhang Y. Relative distance and gaze in the use of entity-referring spatial demonstratives: An event-related potential study. J. Neurolinguist. 2013;26:31–45. doi: 10.1016/j.jneuroling.2012.02.005. [DOI] [Google Scholar]
- 59.Zhang Y., Koerner T., Miller S., Grice-Patil Z., Svec A., Akbari D., Tusler L., Carney E. Neural coding of formant-exaggerated speech in the infant brain. Dev. Sci. 2011;14:566–581. doi: 10.1111/j.1467-7687.2010.01004.x. [DOI] [PubMed] [Google Scholar]
- 60.Miller S., Zhang Y. Neural coding of phonemic fricative contrast with and without hearing aid. Ear Hear. 2014;35:e122–e133. doi: 10.1097/AUD.0000000000000025. [DOI] [PubMed] [Google Scholar]
- 61.Giraud A.L., Collet L., Chery-Croze S., Magnan J., Chays A. Evidence of a medial olivocochlear involvement in contralateral suppression of otoacoustic emissions in humans. Brain Res. 1995;705:15–23. doi: 10.1016/0006-8993(95)01091-2. [DOI] [PubMed] [Google Scholar]
- 62.Kemp D.T. Otoacoustic emissions, their origin in cochlear function, and use. Brit. Med. Bull. 2002;63:223–241. doi: 10.1093/bmb/63.1.223. [DOI] [PubMed] [Google Scholar]
- 63.Norton S.J., Neely S.T. Tone-burst-evoked otoacoustic emissions from normal-hearing subjects. J. Acoust. Soc. Am. 1987;81:1860–1872. doi: 10.1121/1.394750. [DOI] [PubMed] [Google Scholar]
- 64.Mertes I.B. Medial olivocochlear reflex effects on synchronized spontaneous otoacoustic emissions. J. Acoust. Soc. Am. 2020;147:EL235. doi: 10.1121/10.0000886. [DOI] [PubMed] [Google Scholar]
- 65.Garinis A.C., Glattke T., Cone B.K. The MOC reflex during active listening to speech. J. Speech Lang. Hear. Res. 2011;54:1464–1476. doi: 10.1044/1092-4388(2011/10-0223). [DOI] [PubMed] [Google Scholar]
- 66.Khalfa S., Veuillet E., Collet L. Influence of handedness on peripheral auditory asymmetry. Eur. J. Neurosci. 1998;10:2731–2737. doi: 10.1046/j.1460-9568.1998.00286.x. [DOI] [PubMed] [Google Scholar]
- 67.Brown G.J., Ferry R.T., Meddis R. A computer model of auditory efferent suppression: Implications for the recognition of speech in noise. J. Acoust. Soc. Am. 2010;127:943–954. doi: 10.1121/1.3273893. [DOI] [PubMed] [Google Scholar]
- 68.Mertes I.B., Wilbanks E.C., Leek M.R. Olivocochlear efferent activity is associated with the slope of the psychometric function of speech recognition in noise. Ear Hear. 2018;39:583–593. doi: 10.1097/AUD.0000000000000514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Alain C., Woods D.L., Knight R.T. A distributed cortical network for auditory sensory memory in humans. Brain Res. 1998;812:23–37. doi: 10.1016/S0006-8993(98)00851-8. [DOI] [PubMed] [Google Scholar]
- 70.May B.J., Budelis J., Niparko J.K. Behavioral studies of the olivocochlear efferent system: Learning to listen in noise. Arch. Otolaryngol. Head Neck Surg. 2004;130:660–664. doi: 10.1001/archotol.130.5.660. [DOI] [PubMed] [Google Scholar]
- 71.Baer T., Moore B.C. Effects of spectral smearing on the intelligibility of sentences in the presence of interfering speech. J. Acoust. Soc. Am. 1994;95:2277–2280. doi: 10.1121/1.408640. [DOI] [PubMed] [Google Scholar]
- 72.Mertes I.B., Johnson K.M., Dinger Z.A. Olivocochlear efferent contributions to speech-in-noise recognition across signal-to-noise ratios. J. Acoust. Soc. Am. 2019;145:1529. doi: 10.1121/1.5094766. [DOI] [PubMed] [Google Scholar]
- 73.Mertes I.B., Goodman S.S. Within- and across-subject variability of repeated measurements of medial olivocochlear-induced changes in transient-evoked otoacoustic emissions. Ear Hear. 2016;37:e72–e84. doi: 10.1097/AUD.0000000000000244. [DOI] [PubMed] [Google Scholar]
- 74.Guinan J.J.J., Backus B.C., Lilaonitkul W., Aharonson V. Medial olivocochlear efferent reflex in humans: Otoacoustic emission (OAE) measurement issues and the advantages of stimulus frequency OAEs. J. Assoc. Res. Otolaryngol. 2003;4:521–540. doi: 10.1007/s10162-002-3037-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Goodman S.S., Mertes I.B., Lewis J.D., Weissbeck D.K. Medial olivocochlear-induced transient-evoked otoacoustic emission amplitude shifts in individual subjects. J. Assoc. Res. Otolaryngol. 2013;14:829–842. doi: 10.1007/s10162-013-0409-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Mertes I.B. Establishing critical differences in ear-canal stimulus amplitude for detecting middle ear muscle reflex activation during olivocochlear efferent measurements. Int. J. Audiol. 2020;59:140–147. doi: 10.1080/14992027.2019.1673491. [DOI] [PubMed] [Google Scholar]
- 77.Martin B.A., Stapells D.R. Effects of low-pass noise masking on auditory event-related potentials to speech. Ear Hear. 2005;26:195–213. doi: 10.1097/00003446-200504000-00007. [DOI] [PubMed] [Google Scholar]
- 78.Muller-Gass A., Marcoux A., Logan J., Campbell K.B. The intensity of masking noise affects the mismatch negativity to speech sounds in human subjects. Neurosci. Lett. 2001;299:197–200. doi: 10.1016/S0304-3940(01)01508-7. [DOI] [PubMed] [Google Scholar]
- 79.Bennett K.O., Billings C.J., Molis M.R., Leek M.R. Neural encoding and perception of speech signals in informational masking. Ear Hear. 2012;33:231–238. doi: 10.1097/AUD.0b013e31823173fd. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Zhang X., Li X., Chen J., Gong Q. Background suppression and its relation to foreground processing of speech versus non-speech streams. Neuroscience. 2018;373:60–71. doi: 10.1016/j.neuroscience.2018.01.009. [DOI] [PubMed] [Google Scholar]
- 81.Wittekindt A., Kaiser J., Abel C. Attentional modulation of the inner ear: A combined otoacoustic emission and EEG study. J. Neurosci. 2014;34:9995–10002. doi: 10.1523/JNEUROSCI.4861-13.2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Lagace J., Jutras B., Gagne J.P. Auditory processing disorder and speech perception problems in noise: Finding the underlying origin. Am. J. Audiol. 2010;19:17–25. doi: 10.1044/1059-0889(2010/09-0022). [DOI] [PubMed] [Google Scholar]
- 83.Sanches S.G., Carvallo R.M. Contralateral suppression of transient evoked otoacoustic emissions in children with auditory processing disorder. Audiol. Neurotol. 2006;11:366–372. doi: 10.1159/000095898. [DOI] [PubMed] [Google Scholar]
- 84.Micheyl C., Carbonnel O., Collet L. Medial olivocochlear system and loudness adaptation: Differences between musicians and non-musicians. Brain Cognit. 1995;29:127–136. doi: 10.1006/brcg.1995.1272. [DOI] [PubMed] [Google Scholar]
- 85.Strait D.L., Parbery-Clark A., Hittner E., Kraus N. Musical training during early childhood enhances the neural encoding of speech in noise. Brain Lang. 2012;123:191–201. doi: 10.1016/j.bandl.2012.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Wong P.C., Skoe E., Russo N.M., Dees T., Kraus N. Musical experience shapes human brainstem encoding of linguistic pitch patterns. Nat. Neurosci. 2007;10:420–422. doi: 10.1038/nn1872. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Bar-Haim Y., Henkin Y., Ari-Even-Roth D., Tetin-Schneider S., Hildesheimer M., Muchnik C. Reduced auditory efferent activity in childhood selective mutism. Biol. Psychiatry. 2004;55:1061–1068. doi: 10.1016/j.biopsych.2004.02.021. [DOI] [PubMed] [Google Scholar]
- 88.Mishra J., Gazzaley A. Harnessing the neuroplastic potential of the human brain & the future of cognitive rehabilitation. Front. Hum. Neurosci. 2014;8 doi: 10.3389/fnhum.2014.00218. [DOI] [PMC free article] [PubMed] [Google Scholar]