Skip to main content
PLOS One logoLink to PLOS One
. 2020 Oct 15;15(10):e0240752. doi: 10.1371/journal.pone.0240752

Effects of noise on integration of acoustic and electric hearing within and across ears

Shelby Willis 1, Brian C J Moore 2, John J Galvin III 3, Qian-Jie Fu 1,*
Editor: Hussain Md Abu Nyeem4
PMCID: PMC7561114  PMID: 33057396

Abstract

In bimodal listening, cochlear implant (CI) users combine electric hearing (EH) in one ear and acoustic hearing (AH) in the other ear. In electric-acoustic stimulation (EAS), CI users combine EH and AH in the same ear. In quiet, integration of EH and AH has been shown to be better with EAS, but with greater sensitivity to tonotopic mismatch in EH. The goal of the present study was to evaluate how external noise might affect integration of AH and EH within or across ears. Recognition of monosyllabic words was measured for normal-hearing subjects listening to simulations of unimodal (AH or EH alone), EAS, and bimodal listening in quiet and in speech-shaped steady noise (10 dB, 0 dB signal-to-noise ratio). The input/output frequency range for AH was 0.1–0.6 kHz. EH was simulated using an 8-channel noise vocoder. The output frequency range was 1.2–8.0 kHz to simulate a shallow insertion depth. The input frequency range was either matched (1.2–8.0 kHz) or mismatched (0.6–8.0 kHz) to the output frequency range; the mismatched input range maximized the amount of speech information, while the matched input resulted in some speech information loss. In quiet, tonotopic mismatch differently affected EAS and bimodal performance. In noise, EAS and bimodal performance was similarly affected by tonotopic mismatch. The data suggest that tonotopic mismatch may differently affect integration of EH and AH in quiet and in noise.

Introduction

Despite considerable efforts over the last 30 years, advances in cochlear implant (CI) technology and signal processing have yet to show substantial gains in speech performance. While high stimulation rates, deeply inserted electrodes, and current focusing all offer theoretical advantages over previous technology, none have shown consistent advantages for speech perception [13]. The poor functional spectral resolution and limited temporal information provided by CIs continues to limit the perception of speech in noise, speech prosody, vocal emotion, tonal language, and music [410]. Arguably, one of the greatest improvements in CI outcomes has come from combining electric hearing (EH) from the CI with acoustic hearing (AH) in the same ear (electric-acoustic stimulation, or EAS) or in opposite ears (bimodal listening). Note that these definitions do not include BiEAS (residual AH in both ears, CI in one ear) or BiBiEAS (residual AH in both ears, CI in both ears). Residual AH provides detailed low-frequency information that can greatly benefit CI users under challenging listening conditions, including perception of speech in noise and music perception [1123]. Most of these previous studies of combined AH and EH (AEH) have assessed bimodal listening.

With relaxed candidacy criteria and the advance of electrode designs and hearing preservation surgery, increasing numbers of CI users have some residual AH in the implanted ear, allowing for EAS. Several CI manufacturers have developed hybrid speech processors that integrate hearing aid (HA) amplification and CI signal processing in the same ear. For EAS CI users, the current spread associated with electric stimulation may interfere with AH at the periphery. For bimodal users, there is no peripheral interaction between AH and EH, which may be advantageous. Similar to bimodal listening, Gantz et al. [24] found better speech performance with EAS than with the CI alone.

For AEH, clinical fitting of the CI is often performed without regard to the characteristics of AH, which may limit the integration of AH and EH due to two major factors–energetic/modulation interference and information interference. For EAS patients, the broad current spread with EH may interact with the spread of excitation (SOE) produced by AH [25]. This may lead to interference both in terms of energetic masking and of modulation masking, whereby the amplitude modulation (AM) patterns associated with acoustic and electric stimulation interfere with one another [26, 27]. For EAS, this interference occurs more at the periphery; for bimodal listening, such interference would necessarily be more central in origin. For bimodal and EAS patients, the same temporal envelope information may also be delivered to different cochlear locations within or across ears (“tonotopic mismatch”).

Sound quality differences between AH and EH may also affect integration of acoustic and electric hearing. To our knowledge, such sound quality differences have not been systematically studied in CI users, in quiet or noise. If the peripheral representations are sufficiently different, it is likely that sound quality may also be different across ears. While not the same as sound quality measurements across ears, inter-aural pitch matching experiments have shown difficulties matching the pitch of an acoustic signal in one ear [which contains temporal fine-structure (TFS) information that is important to pitch perception] to the pitch associated with stimulation of a single electrode [which does not contain TFS information and stimulates a broader cochlear region]. Work by Reiss and colleagues [2830] has shown that inter-aural frequency mismatch can limit binaural fusion of acoustic and electric stimulation across ears or bilateral electric stimulation. Such inter-aural frequency mismatch can also limit perception of inter-aural timing differences (ITDs) that are important for sound source localization and image fusion [3133].

Other studies have shown advantages for combining non-overlapping spectral information across ears. For example, Kulkarni et al. [34] showed a binaural advantage in spectral resolution for AH listeners when spectral bands were combined across ears. Similarly, Aronoff et al. [35] found that, for bilateral CI listeners, spectral resolution was significantly better when distributing 12 spectral channels across ears than within ears. Kong and Braida [36] found that the benefit of combining low-frequency AH with contralateral broadband EH was greater for vowels than for consonants, for both real and simulated CI patients. Luo and Fu [37], testing Chinese Mandarin-speaking NH participants listening to bimodal CI simulations, found that acoustic information < 500 Hz was more important for lexical tone perception (where voice pitch cues are important), and that acoustic information > 500 Hz was important for Mandarin phoneme recognition. The authors also found no difference for speech understanding in noise whether or not there was inter-aural spectral overlap between the simulated AH and EH. “Zipper” CI signal processing, where independent electrode information is provided to each ear, has shown binaural advantages in temporal processing and/or speech performance relative to monaural performance [38, 39].

Fu et al. [40] measured recognition of vowels in quiet for normal-hearing (NH) subjects listening to simulations of residual AH and EH. NH subjects and simulations were used to explicitly control the extent of stimulation within the cochlea and to directly compare the perception of EAS and bimodal stimulation. Such comparisons cannot easily be made using real EAS and bimodal CI patients, as the extent/quality of residual AH and the electrode-neural interface (the number and position of intra-cochlear electrodes relative to healthy neurons) are likely to vary markedly across ears and/or CI users. In the Fu et al. study [40], results showed that AH and EH were better combined for EAS than for bimodal listening. Relative to simulated EH, the benefits of simulated AEH were nearly twice as large for EAS as for bimodal listening. Also, acoustic-electric integration efficiency (IE, defined as the ratio between observed and predicted AEH performance; see later for details) was generally better for EAS than for bimodal listening. EAS IE was significantly affected by tonotopic mismatch, while bimodal IE was not. For CI-only and bimodal listening, there appeared to be a tradeoff between the amount of speech information and tonotopic mismatch. Compared to bimodal listening, EAS was less affected by small amounts of tonotopic mismatch, but more affected by larger mismatches.

In a subsequent study, Fu et al. [41] varied the carriers used in the CI simulations, including broad-band noise (limited only by the channel bandwidth), narrow-band noise, and sine waves. Reducing the bandwidth significantly improved simulated EH and bimodal performance, but not EAS performance. These results suggest that the inherent amplitude fluctuations in a broad-band noise carrier may negatively affect integration of AH and EH across ears. While reducing the bandwidth may improve both bimodal performance and integration efficiency for speech in quiet, it is possible that the deleterious effects of the inherent noise fluctuations on bimodal performance may be smaller or even absent for speech in noise. The amplitude fluctuations in external noise may mask the inherent noise fluctuations in broad-band noise carriers used to simulate EH. External noise may also reduce sound quality differences between AH and EH, thus improving the integration of AH and EH across ears; this might be true for both CI simulations and the real CI case, especially for bimodal listening. While previous studies showed that bimodal benefits are highly variable across individual CI users and different listening tasks and conditions, bimodal benefits are generally greater for speech in noise than for speech in quiet [21]. The bimodal benefit for speech in noise has often been attributed to a better ability to segregate the speech from the noise using residual AH. If external noise can mask the inherent noise fluctuations in simulated EH, it is possible that integration of AH and EH across ears may be better than previously observed in quiet.

The main goal of the present study was to evaluate integration of AH and simulated EH within and across ears in the presence of noise. Monosyllabic word recognition was measured in NH subjects using simulations of unimodal (AH or EH), EAS, and bimodal listening for speech in quiet or in steady speech-shaped noise. We hypothesized that in noise, AH and EH would be better integrated within (EAS) than across ears (bimodal), but with greater sensitivity to tonotopic mismatch than with bimodal listening, similar to our previous results in quiet [40, 41]. We also expected that, similar to the greater benefits observed in noise than in quiet in real bimodal and EAS listeners [21], we expected integration of AH and EH to be better in noise than in quiet in the present CI simulations.

Materials and methods

This study was approved by the Institutional Review Board of the University of California, Los Angeles (UCLA). Prior to participation, written informed consent was obtained from all subjects, in accordance with a protocol approved by the Institutional Review Board at UCLA.

Subjects

Ten NH subjects (5 males and 5 females) participated in this study. Participants were recruited from UCLA main campus. Recruitment postings were placed on the Digest News & Events website at the UCLA David Geffen School of Medicine (https://digest.dgsom.ucla.edu/). Inclusion criteria was that subjects have normal AH in both ears for all audiometric frequencies, that they were at least 18 years old, and were native speakers of American English. The mean age at testing was 25 years (range: 23–30 years). All subjects had thresholds <20 dB HL for audiometric frequencies 250, 500, 1000, 2000, 4000, and 8000 Hz in each ear. While the subject sample is small, the normal AH status in each ear presumably reflected uniform neural survival. As such, the various spectral degradation manipulations could be applied with the expectation that the effects would be similar across subjects.

Test stimuli and procedures

Monosyllable CNC words spoken by one male talker were used. Stimuli were delivered via circumaural headphones (Sennheiser HDA-200) connected to separate channels of a mixer (Mackie 402 VLZ3), which was connected to an audio interface (Edirol UA-EX25). Before signal processing, all stimuli were normalized to have the same long-term root-mean-square (RMS) energy. Word recognition was measured in quiet and in noise at two signal-to-noise ratios (SNRs): 10 dB and 0 dB. Steady noise was filtered to have same spectrum averaged across all words. For each condition, a CNC list was randomly selected and words were randomly selected from within the list (without replacement) and presented to the subject, who was asked to repeat what they heard as accurately as possible. The experimenter calculated the percent of words correctly identified. All words in the list were scored, resulting in a total of 50 words for each subject and condition. No training or trial-by-trial feedback was provided. The test order of different listening conditions and SNRs was randomized within and counter-balanced across subjects.

Simulations

Residual AH was simulated by bandpass filtering the speech signal between 0.1 and 0.6 kHz (20th order Butterworth filters; 240 dB/octave). This range was selected to represent the residual hearing available to some EAS and bimodal CI listeners, and to convey some information about the first formant frequency in speech. EH was simulated using 8-channel noise vocoders, similar to those described by Shannon et al. [42]. The input frequency range was divided into 8 bands (4th order Butterworth filters; 48 dB/octave), distributed according to Greenwood’s frequency-to-place formula [43]. The temporal envelope was extracted from each analysis band by half-wave rectification and low-pass filtering (4th order Butterworth filter with 160-Hz cutoff frequency). The temporal envelope from each channel was used to modulate corresponding noise bands; the filter slope for the noise band carriers was same as that of the analysis filters. The modulated noise-bands were summed and the output was adjusted to have the same long-term RMS energy as the input. The output frequency range of the noise vocoders was fixed at 1.2–8.0 kHz. The lowest output frequency (1.2 kHz) corresponds to the cochlear location for a 20-mm insertion of an electrode array according to Greenwood [43] and is slightly higher than the median upper edge frequency of the residual AH (approximately 1.1 kHz) for hybrid CI patients reported by Karsten et al. [44]. The highest output frequency (8.0 kHz) is similar to the highest input frequency commonly used in commercial CI speech processors. Note that the output range of the CI simulations was not intended to simulate specific commercial CI devices, which vary in terms of array length, the number of electrodes and electrode spacing. Rather, the output frequency range was fixed, and the input frequency range was varied to present different amounts of acoustic information while introducing different amounts of tonotopic mismatch. The upper limit of the simulated CI input frequency range was always 8.0 kHz. The lower limit was 0.6 or 1.2 kHz. When the limit was 0.6 kHz, all speech information was presented via the combination of AH and EH, but with a place mismatch of 4.3 mm at the apical end of the simulated electrode array according to Greenwood’s [43] function. This condition is denoted as the “EH-mismatched” frequency allocation. When the lower limit of the input frequency range was 1.2 kHz, less acoustic information was presented, but there was no place mismatch between AH and EH and no overlap between the AH and EH input frequency ranges. This condition is denoted as the “EH-matched” frequency allocation. Performance with the AH and EH simulations alone (unimodal) was measured with one ear only. For the EAS simulation, AH and EH simulations were delivered to one earpiece of the headphones. For the bimodal simulation, AH and EH simulations were delivered to opposite earpieces of the headphones. The overall presentation level for the stimuli (combined speech and noise) was 60 dBA.

It should be noted that the present simulations (and indeed, all CI simulations) do not capture potential deficits in speech processing at higher levels of the auditory system that may be impaired by long-term hearing loss. For example, previous studies have shown that for prolonged auditory deprivation, cortical deficits and/or slower cortical development that may persist even after cochlear implantation [45, 46]. Such central deficits were not captured by the present simulations, which were more intended to simulate different peripheral limitations, but perceived with normal hearing and (presumably) normal higher-level auditory processing.

Results

Fig 1 shows mean word recognition scores with the AH, EH, bimodal, and EAS listening conditions as a function of SNR; the left panels show data with the EH-mismatched allocation and the right panels show data with EH-matched allocation. Performance generally worsened as the SNR was reduced, and generally improved with AEH over AH or EH alone.

Fig 1. Mean CNC word recognition for the EH-mismatched (Panel A) and the EH-matched frequency allocations (Panel B).

Fig 1

Data are shown for the AH, EH, bimodal, and EAS listening conditions. The error bars show the standard error.

A repeated-measures analysis of variance (RM ANOVA) was performed on arc-sine transformed unimodal word recognition scores, with SNR (Quiet, 10 dB, 0 dB) and unimodal listening condition (AH, EH-mismatched, EH-matched) as factors; complete statistical results are shown in Table 1. Significant main effects were observed for SNR and unimodal listening condition (P < 0.001 in all cases); there was no significant interaction (P = 0.071). Another RM ANOVA was performed on the arcsine-transformed word recognition scores in Fig 1, this time with SNR (Quiet, 10 dB, 0 dB), CI mode (EH, bimodal, EAS), and CI allocation (mismatched, matched) as factors; complete statistical results are shown in Table 1. Significant main effects were observed for SNR, CI mode, and CI allocation (P < 0.001 in all cases). Significant interactions were observed between CI mode and CI allocation (P = 0.011), and among SNR, CI mode and CI allocation (P = 0.033).

Table 1. Statistical results for word recognition scores.

Unimodal dF, res F P ŋ2 Observed power Bonferroni post-hoc (P < 0.05)
SNR 2,18 93.7 *< 0.001 0.91 > 0.99 Quiet>10 dB>0 dB
Unimodal 2,18 32.8 *< 0.001 0.79 > 0.99 EH-matched>EH-mismatched>AH
SNR*Unimodal 4,36 2.4 0.071 0.21 0.62
CI mode dF, res F P ŋ2 Observed power Bonferroni post-hoc (P < 0.05)
SNR 2,18 110.7 *< 0.001 0.93 > 0.99 Quiet>10 dB>0 dB
CI mode 2,18 175.8 *< 0.001 0.95 > 0.99 Bimodal, EAS>EH
CI allocation 1,9 51.9 *< 0.001 0.85 > 0.99 Matched>Mismatched
SNR*CI mode 4,36 1.0 0.399 0.10 0.30
SNR*CI allocation 2,18 1.4 0.275 0.13 0.26
CI mode*CI allocation 2,18 5.9 *0.011 0.40 0.81 Mismatched: Bimodal>EAS>EH; Matched: Bimodal, EAS>EH
SNR*CI mode*CI allocation 4,36 3.0 * 0.033 0.25 0.73 Quiet, Mismatched: BI>EAS>EH; Quiet, Matched: BI, EAS>EH

For the unimodal analysis in the top part of the table, factors included SNR (Quiet, 10 dB SNR, 0 dB SNR) and Unimodal listening (AH, EH-mismatched, EH-matched). For the CI mode analysis in the bottom part of the table, factors included SNR (Quiet, 10 dB SNR, 0 dB SNR), CI mode (EH, EAS, Bimodal), and CI allocation (mismatched, matched).

The advantage for combined AH and EH (“AEH advantage”) was calculated relative to EH-only performance. Fig 2 shows the AEH advantage for word recognition with bimodal or EAS listening (relative to EH performance) as a function of SNR. In quiet (Panel A), the AEH advantage for the EH-mismatched condition (black bars) was much larger for bimodal than for EAS listening. In contrast, the AEH advantage for the EH-matched condition (gray bars) was larger for EAS than for bimodal listening. In noise, the AEH advantage was generally larger for the EH-matched than for the EH-mismatched condition, with no strong differences between bimodal and EAS listening. AEH advantage was generally reduced as the SNR was reduced. A RM ANOVA was performed on the AEH advantage data shown in Fig 2, with SNR (Quiet, 10 dB, 0 dB), AEH mode (bimodal, EAS), and CI allocation (EH-mismatched, EH-matched) as factors. A significant main effect was observed only for CI allocation [dF(res) = 1(9); F = 36.3; P < 0.001; ŋ2 = 0.80]. However, there was a significant interaction among SNR, AEH mode, and CI allocation [dF(res) = 2(18); F = 5.1; P = 0.017; ŋ2 = 0.36], most likely driven by the strong difference in AEH according to CI allocation for EAS in the quiet listening condition.

Fig 2. Mean AEH advantage for bimodal and EAS listening (relative to EH performance) for CNC words in quiet (Panel A), at 10 dB SNR (Panel B), and at 0 dB SNR (Panel C).

Fig 2

Data are shown for the EH-mismatched (black bars) and EH-matched frequency allocations (gray bars). The error bars show the standard error.

Integration efficiency (IE) for AEH was calculated for the present bimodal and EAS data as the ratio between observed and predicted AEH performance, as in Fu et al. [40].

IE=PAEH/P^AEH (1)

where PAEH is the observed (measured) AEH proportion correct and P^AEH is the predicted AEH proportion correct. The predicted AEH proportion correct was based on the probability-summation rule:

P^AEH=1(1PAH)(1PEH) (2)

where PAH and PEH represent the observed (measured) proportions correct and (1 –PAH) and (1 –PEH) represent the observed proportions incorrect for AH alone and EH alone, respectively. According to this rule, the product (1 –PAH)(1 –PEH) corresponds to the probability of being incorrect for both AH and EH. It should be noted that other methods for predicting the way that AH and EH are combined could give different outcomes [47]. However, the simple method defined by Eq 1 was considered sufficient for comparing IE across conditions. IE values > 1 imply “super-additive” or synergistic integration of acoustic and electric hearing.

Fig 3 shows IE as a function of SNR with bimodal (left panel) and EAS listening (right panel) with the EH-mismatched and EH-matched allocations. In all cases, IE increased as the SNR was reduced, largely due to the poorer AH and EH performance with increasing noise. In general, there was very little effect of listening condition or EH allocation on IE, except for the higher IE for the EH-matched allocation for EAS in the quiet listening condition. A RM ANOVA was performed on the IE data shown in Fig 3, with SNR (Quiet, 10 dB, 0 dB), AEH mode (binaural, EAS) and CI allocation (EH-mismatched, EH-matched) as factors. Note that data for 1 subject was excluded because of 0% correct scores for AH and EH at 0 dB SNR, which prohibited accurate estimation of IE. A significant main effect was observed only for SNR [dF(res) = 2(16); F = 19.1; P < 0.001; ŋ2 = 0.70], and there were no significant interactions. Post-hoc Bonferroni pairwise comparisons showed that IE was significantly larger at 0 dB SNR than in quiet or at 10 dB SNR (P < 0.05 in both cases), with no significant difference between quiet and 10 dB SNR.

Fig 3. Mean IE for CNC words as a function of SNR with the EH-matched (filled circles) and EH-mismatched frequency allocations (open circles) for bimodal (Panel A) and EAS listening (Panel B).

Fig 3

Values > 1 indicate that observed AEH performance was better than predicted by AH and EH performance. The error bars show the standard error.

Discussion

Consistent with our previous studies [40, 41], CNC word recognition in quiet was poorer with EAS than with bimodal listening when there was a tonotopic mismatch in the simulated EH. Different from our previous studies, performance in quiet was similar with bimodal and EAS listening when there was no tonotopic mismatch. Contrary to our hypothesis, there was no difference between bimodal and EAS performance in noise, regardless of the degree of tonotopic mismatch in EH. The present data suggest that integration of AH and EH may be affected by interactions among the degree of tonotopic mismatch, the distribution of the AH (same or opposite ear of EH), and the presence of noise.

Effect of outcome measures on integration of AH and EH

In Fu et al. [40, 41], integration of AH and EH was measured in quiet using a closed-set vowel recognition task, which was expected to be more sensitive to tonotopic mismatch than would be sentence recognition, where top-down processes might better accommodate the mismatch. For example, Fu et al. [48] showed stronger and more rapid adaptation to shifted frequency allocations in real CI users for recognition of sentences than words in quiet. In the present study, open-set CNC word recognition was measured at some “midway” point between closed-set vowel recognition (more bottom-up) and open-set sentence recognition (more top-down).

In quiet, the EAS advantage over bimodal listening observed for vowel recognition in Fu et al. [40, 41] was not observed for word recognition in the present study. Interestingly, EH performance (with or without mismatch) was comparable between word recognition in the present study and vowel recognition in Fu et al. [40, 41]. However, adding residual AH appeared to provide a greater benefit for word recognition than for vowel recognition. It is possible that adding the residual AH allowed for better perception of initial and final consonants in words than with EH alone. The vowel recognition task in Fu et al. [40, 41] required only perception of medial vowels, where low-frequency acoustic cues may have largely facilitated perception of some first formant cues. For real bimodal CI listeners, Yoon et al. [20] showed a greater AEH benefit for sentence recognition in quiet than for vowel or consonants, and a greater benefit for consonants than for vowels. Taken together, the present and previous data suggest that some outcome measures may be more sensitive than others to tonotopic mismatch in EH, as well as the benefits of adding residual AH, depending on the listening demands.

Effect of tonotopic mismatch on integration of AH and EH

The present data suggest that the effects of tonotopic mismatch between AH and EH may be mitigated by the presence of noise. In quiet, word recognition was significantly better when the EH allocation was tonotopically matched (Fig 1; Table 1). However, there were some interactions between EH tonotopic mismatch and other test conditions. For example, a significant AEH advantage for EAS was observed only when the EH allocation was tonotopically matched (Fig 2). For bimodal listening, a significant AEH advantage was observed regardless of the degree of mismatch in the EH frequency allocation; with either allocation, the AEH benefit was significantly greater for bimodal than for EAS listening.

In noise, word recognition was significantly better when the EH allocation was tonotopically matched (Fig 1; Table 1). The effect of tonotopic mismatch on AEH advantage was similar for bimodal and EAS listening. The differential effects of mismatch on the distribution of AH (same or opposite ear as EH) observed in quiet were not observed in noise. As such, adding residual AH similarly benefitted bimodal and EAS listening in noise whether or not there was a tonotopic mismatch in EH. The present results in noise were different from those in quiet, and different from previous results in quiet from Fu et al. [40], suggesting that bimodal and EAS listening may be similarly sensitive to tonotopic mismatch in noise.

The present data show that the tradeoff between tonotopic matching and information loss observed in Fu et al. [40, 41] for vowel recognition in quiet persisted in noise. This is not consistent with Gifford et al. [49], who found that for real bimodal CI users, the best performance was observed when a wide input acoustic frequency range was used (more speech information, but with more mismatch), rather than a more tonotopically matched range (less information, but with less mismatch). For the present data, performance was always poorer with a tonotopic mismatch. Note that only one electrode insertion depth was used to create the EH tonotopic mismatch in the present study, compared to previous studies where a range of insertion depths/tonotopic mismatches were tested [40]. It is possible that the patterns of results may differ for words in noise (used in this study) and sentences in noise (used in Gifford et al. [49]), where top-down auditory processing might play a greater role. Also, real bimodal CI listeners have much greater experience with tonotopic mismatch in their device. Such adaptation to the mismatch in the EH simulation was not tested in the present study, as only acute performance was measured. It is possible that some short-term training may have reduced the effects of the mismatch [5053]. It is also possible that the present CI simulations did not capture real CI users’ performance.

Effect of distribution of AH

While overall AEH was better with EAS than with bimodal listening in quiet, there was no significant difference between EAS and bimodal listening in noise, even when the SNR was relatively high (10 dB). This suggests that limited AH in either ear is beneficial in noise. It could be that the presence of AH in either ear improved perception of voicing information [16], rather than integration with the EH spectral pattern. The distribution of AH did not significantly affect AEH advantage with the EH-mismatched and EH-matched allocations. This suggests that AH did not necessarily facilitate spectral integration in noise in the same way as observed in quiet, where there was an advantage for EAS, but also less sensitivity to mismatch with bimodal listening.

Effect of noise on integration of AH and EH

Not surprisingly, overall performance significantly worsened for all listening conditions as the amount of noise increased from 10 dB to 0 dB SNR (Fig 1; Table 1). The effects of tonotopic mismatch observed in quiet also persisted in noise, as performance was better with the EH-matched allocation despite the associated information loss. Significant interactions were observed between CI listening mode (EH, bimodal, EAS), and EH allocation only in quiet; here, bimodal performance was better than EAS when there was a tonotopic mismatch, but similar when there was no mismatch. In noise, there were no such interactions. This suggests that noise was a great equivocator on the differential effects of AH distribution (same or opposite ear from EH) and tonotopic mismatch observed in quiet. As such, bimodal and EAS may be similarly advantageous and similarly sensitive to tonotopic mismatch when listening to speech in noise.

It is possible that noise may have masked sound quality differences between residual AH and the noise-band vocoders used for the EH simulations, resulting in better fusion between AH and EH. However, noise caused performance to worsen for all unimodal (AH, EH) and AEH (bimodal, EAS) listening modes. As such, low-amplitude consonant and vowel information may have been masked by noise, which would obviate putative advantages due to consistent sound quality between AH and EH. This is somewhat borne out by the similar decrements in performance across listening modes as the SNR worsened. It is unclear whether a similar pattern of results would have been observed if sine-wave carriers (which have no inherent noise fluctuations) had been used instead of noise-band carriers.

Effects of tonotopic mismatch, distribution of AH, and noise on IE

It is important to note that IE does necessarily not reflect absolute performance or even AEH performance gains per se. Instead, IE reflects the potential super-additivity of the unimodal components (AH, EH) for AEH listening under the various SNR, AH distribution, and tonotopic mismatch conditions. In general, the IE data show that as unimodal performance worsened due to increased noise and/or tonotopic mismatch, combining AH and EH contributed more strongly to performance gains. The present bimodal IE data are comparable to those from Yang and Zeng [54], who showed a mean value of 1.31 for concurrent vowel recognition for simulations of bimodal listening. IR was generally similar for the EH-matched and EH-mismatched allocations (except for in quiet, where there was a bimodal advantage for the EH-mismatched allocation). This reflects the greater sensitivity to tonotopic mismatch in quiet for EAS, and the greater sensitivity to mismatch at poor SNRs for bimodal listening.

The present IE data suggest that the relative contribution of AH to AEH listening increases as the listening conditions become more adverse (noise, tonotopic mismatch), regardless of how the AH is distributed across ears. Note that BiEAS listening were not simulated in this study. It is possible that IE might have been even greater for BiEAS, as AH and EH might have only been marginally affected by bilateral presentation, but AEH performance would likely be even better than presently observed for bimodal or EAS listening. Gifford et al. [49] found a consistent advantage for BiEAS over bimodal listening in real CI users.

Clinical implications and limitations

As indications for cochlear implantation continue to expand, and as surgical techniques and electrode designs continue to improve, combining AH and EH in the same ear, opposite ears, and even both ears will become more commonplace. Depending on the stability of residual AH and/or the electrode insertion depth, complete or partial residual AH preservation may be possible in the implanted ear. The present results suggest that minimizing tonotopic mismatch for EH may increase the benefit of AEH, especially for speech in noise, where both EAS and bimodal hearing were highly sensitive to tonotopic mismatch.

In the present study, the effects of tonotopic mismatch were acutely measured in NH participants listening to simulations of EH, bimodal and EAS, with no time for adaptation or training. Real CI patients can at least partly adapt to tonotopic mismatch with extended experience [48, 5556]. CI simulation studies have shown that while listeners can automatically adapt to small amounts of tonotopic mismatch [57]; even greater adaptation is possible with explicit training [5052]. It is possible that the effects of tonotopic mismatch observed in this CI simulation study might be mitigated over time in real CI users via passive adaptation and/or explicit auditory training.

To the extent that reducing tonotopic mismatch is advantageous for AEH, using radiological imaging to estimate the intra-cochlear electrode position, which then can be used to optimize frequency allocations [58, 59]. Intracochlear electrocochleography (ECochG), in which intra-cochlear electrodes are used to record responses to acoustic stimulation [60, 61], may also be used to estimate electrode position and guide tonotopic matching for CI patients with residual AH, especially for EAS listening.

One limitation of the current present study is that only a single frequency gap (0.6–1.2 kHz) between AH and EH was simulated; AH was limited to frequencies between 0.1 and 0.6 Hz, and the output frequency range for EH was 1.2–8.0 kHz. Increasing numbers of CI patients are being implanted with long electrode arrays and relatively deep insertions. Some CI patients also have residual AH in the implanted ear for frequencies above 0.6 kHz. For such CI patients, there may be no frequency gap or even an overlap between the frequency ranges covered by AH and EH. It is unclear how noise may affect integration of AH and EH in such cases. Also, there are increasing numbers of BiEAS CI users, as well as single-sided deaf CI users (normal AH in one ear, CI in the other ear). CI users’ speech and localization performance has been shown to greatly benefit when there is limited AH in both ears [62, 63]. It is unclear how the availability of residual AH in both ears might have affected the present pattern of results with the present CI simulations. In future studies, it would be worthwhile to evaluate the effects of tonotopic mismatch in EH in the context of residual AH in both ears and CI in one or both ears.

Finally, it is unclear the extent to which the present CI simulation data extend to real bimodal and EAS CI users. CI simulations allow for better control of stimulation parameters, and NH listeners are likely to be a homogenous group in terms of performance, and have uniform auditory neural survival. This is rarely the case across CI users, who may use different devices, different parameter settings, different etiologies of deafness, different amounts of experience with their device, etc. Still, data from the present and previous CI simulation studies allow researchers to examine the effects of parameter manipulations in the “best-case scenario” of normal AH. Aspects of CI patient variability (e.g., patchy neural survival, electrode insertion depth, the number of implanted electrodes) can also be simulated to better understand how CI signal processing may interact with patient-specific factors. The simulations also can be used to identify areas of interest that might be fruitful in optimizing stimulation parameters for real CI patients. Future studies with real bimodal and EAS CI users may further elucidate the importance of tonotopic matching for combined acoustic and electric hearing.

Summary and conclusions

The present study examined monosyllabic word recognition in quiet and in noise for NH participants listening to simulations of residual AH, EH, bimodal (residual AH and EH in opposite ears), and EAS (residual AH and EH in the same ear). The EH simulations were either tonotopically matched or mismatched. Key findings include:

  1. Overall performance was poorer when there was a tonotopic mismatch in EH In quiet, bimodal and EAS performance was similar when there was no tonotopic mismatch, but poorer for EAS than for bimodal when there was a mismatch. In noise, performance was similar between bimodal and EAS regardless of the degree of tonotopic mismatch.

  2. In quiet, the AEH advantage relative to EH differed between bimodal and EAS, especially when there was a tonotopic mismatch, where performance was much poorer with EAS than with bimodal listening. In noise, the AEH advantage was similar for bimodal and EAS, regardless of the degree of tonotopic mismatch. However, the AEH advantage reduced with increasing noise level.

  3. IE was generally similar between EAS and bimodal listening, and was largely unaffected by tonotopic mismatch.

Supporting information

S1 File. S1 File contains the raw word percent correct data for all subjects and listening conditions.

(XLSX)

Acknowledgments

We thank all participants for their time and effort. We also thank the editor and three anonymous reviewers for helpful comments.

Data Availability

All relevant data are within its supporting information files.

Funding Statement

The research was supported by The National Institute on Deafness and Other Communication Disorders (NIDCD) R01-DC-016883. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Boyd PJ. Potential benefits from deeply inserted cochlear implant electrodes. Ear Hear. 2011; 32(4):411–427. 10.1097/AUD.0b013e3182064bda [DOI] [PubMed] [Google Scholar]
  • 2.Shannon RV, Cruz RJ, Galvin JJ 3rd. Effect of stimulation rate on cochlear implant users' phoneme, word and sentence recognition in quiet and in noise. Audiol Neurootol. 2011; 16(2):113–23. 10.1159/000315115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Bierer JA, Litvak L. Reducing channel interaction through cochlear implant programming may improve speech perception: current focusing and channel deactivation. Trends Hear. 2016; 17(20): 1–12. 10.1177/2331216516653389 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Fu QJ, Zeng FG, Shannon RV, Soli SD. Importance of tonal envelope cues in Chinese speech recognition. J Acoust Soc Am. 1998; 104(1): 505–510. 10.1121/1.423251 [DOI] [PubMed] [Google Scholar]
  • 5.Friesen LM, Shannon RV, Baskent D, Wang X. Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. J Acoust Soc Am. 2001; 110(2): 1150–1163. 10.1121/1.1381538 [DOI] [PubMed] [Google Scholar]
  • 6.Gfeller K, Turner C, Mehr M, Woodworth G, Fearn R, Knutson JF, et al. Recognition of familiar melodies by adult cochlear implant recipients and normal-hearing adults. Cochlear Implants Int. 2002; 3(1):29–53. 10.1179/cim.2002.3.1.29 [DOI] [PubMed] [Google Scholar]
  • 7.Shannon RV, Fu QJ, Galvin J 3rd. The number of spectral channels required for speech recognition depends on the difficulty of the listening situation. Acta Otolaryngol Suppl. 2004; 552:50–54. 10.1080/03655230410017562 [DOI] [PubMed] [Google Scholar]
  • 8.Galvin JJ III, Fu Q-J, Nogaki G. Melodic contour identification by cochlear implant listeners. Ear Hear. 2007; 28:302–319. 10.1097/01.aud.0000261689.35445.20 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Luo X, Fu QJ, Galvin JJ 3rd. Vocal emotion recognition by normal-hearing listeners and cochlear implant users. Trends Amplif. 2007; 11(4): 301–315. 10.1177/1084713807305301 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Chatterjee M, Peng SC. Processing F0 with cochlear implants: Modulation frequency discrimination and speech intonation recognition. Hear Res. 2008; 235(1–2): 143–156. 10.1016/j.heares.2007.11.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Armstrong M, Pegg P, James C, Blamey P. Speech perception in noise with implant and hearing aid. Am J Otol. 1997; 18(6 Suppl): S140–141. [PubMed] [Google Scholar]
  • 12.Tyler RS, Parkinson AJ, Wilson BS, Witt S, Preece JP, Noble W. Patients utilizing a hearing aid and a cochlear implant: speech perception and localization. Ear Hear. 2002; 23(2): 98–105. 10.1097/00003446-200204000-00003 [DOI] [PubMed] [Google Scholar]
  • 13.Kong YY, Stickney GS, Zeng FG. Speech and melody recognition in binaurally combined acoustic and electric hearing. J Acoust Soc Am. 2005; 117(3 Pt 1): 1351–1361. 10.1121/1.1857526 [DOI] [PubMed] [Google Scholar]
  • 14.Looi V, McDermott H, McKay C, Hickson L. The effect of cochlear implantation on music perception by adults with usable pre-operative acoustic hearing. Int J Audiol. 2008; 47(5): 257–268. 10.1080/14992020801955237 [DOI] [PubMed] [Google Scholar]
  • 15.Dorman MF, Gifford RH, Spahr AJ, McKarns SA. The benefits of combining acoustic and electric stimulation for the recognition of speech, voice and melodies. Audiol Neurootol. 2008; 13(2):105–112. 10.1159/000111782 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Brown CA, Bacon SP. Achieving electric-acoustic benefit with a modulated tone. Ear Hear. 2009; 30(5): 489–493. 10.1097/AUD.0b013e3181ab2b87 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Dorman MF, Gifford RH. Combining acoustic and electric stimulation in the service of speech recognition. Int J Audiol. 2010; 49(12):912–919. 10.3109/14992027.2010.509113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Zhang T, Dorman MF, Spahr AJ. Information from the voice fundamental frequency (F0) region accounts for the majority of the benefit when acoustic stimulation is added to electric stimulation. Ear Hear. 2010a; 31(1):63–69. 10.1097/aud.0b013e3181b7190c [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Zhang T, Spahr AJ, Dorman MF. Frequency overlap between electric and acoustic stimulation and speech-perception benefit in patients with combined electric and acoustic stimulation. Ear Hear. 2010b; 31(2):195–201. 10.1097/AUD.0b013e3181c4758d. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Yoon YS, Li Y, Fu QJ. Speech recognition and acoustic features in combined electric and acoustic stimulation. J Speech Lang Hear Res. 2012; 55(1):105–124. 10.1044/1092-4388(2011/10-0325) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Yoon YS, Shin YR, Gho JS, Fu QJ. Bimodal benefit depends on the performance difference between a cochlear implant and a hearing aid. Cochlear Implants Int. 2015; 16(3):159–167. 10.1179/1754762814Y.0000000101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Crew JD, Galvin JJ 3rd, Landsberger DM, Fu QJ. Contributions of electric and acoustic hearing to bimodal speech and music perception. PLoS One. 2015; 10(3). 10.1371/journal.pone.0120279. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Crew JD, Galvin JJ 3rd, Fu QJ. Perception of sung speech in bimodal cochlear implant users. Trends Hear. 2016; 11:20 pii: 2331216516669329. 10.1177/2331216516669329 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Gantz BJ, Dunn C, Oleson J, Hansen M, Parkinson A, Turner C. Multicenter clinical trial of the Nucleus Hybrid S8 cochlear implant: Final outcomes. Laryngoscope. 2016; 126(4):962–973. 10.1002/lary.25572 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Krüger B, Büchner A, Nogueira W. Simultaneous masking between electric and acoustic stimulation in cochlear implant users with residual low-frequency hearing. Hear Res. 2017; 353:185–196. 10.1016/j.heares.2017.06.014 [DOI] [PubMed] [Google Scholar]
  • 26.Stone MA, Moore BCJ. On the near non-existence of "pure" energetic masking release for speech. J Acoust Soc Am. 2014; 135:1967–1977. 10.1121/1.4868392 [DOI] [PubMed] [Google Scholar]
  • 27.Stone MA, Canavan S. The near non-existence of "pure" energetic masking release for speech: Extension to spectro-temporal modulation and glimpsing. J Acoust Soc Am. 201; 140:832–842. 10.1121/1.4960483 [DOI] [PubMed] [Google Scholar]
  • 28.Reiss LAJ, Fowler JR, Hartling CL, Oh Y. Binaural pitch fusion in bilateral cochlear implant users. Ear Hear. 2018;39(2):390–397. 10.1097/AUD.0000000000000497 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Reiss LA, Ito RA, Eggleston JL, Wozny DR. Abnormal binaural spectral integration in cochlear implant users. J Assoc Res Otolaryngol. 2014;15(2):235–248. 10.1007/s10162-013-0434-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Oh Y, Reiss LAJ. Binaural pitch fusion: Pitch averaging and dominance in hearing-impaired listeners with broad fusion. J Acoust Soc Am. 2017;142(2):780 10.1121/1.4997190 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Goupell MJ, Stoelb C, Kan A, Litovsky RY. Effect of mismatched place-of-stimulation on the salience of binaural cues in conditions that simulate bilateral cochlear-implant listening. J Acoust Soc Am. 2013;133(4):2272–2287. 10.1121/1.4792936 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kan A, Litovsky RY, Goupell MJ. Effects of interaural pitch matching and auditory image centering on binaural sensitivity in cochlear implant users. Ear Hear. 2015;36(3):e62–e68. 10.1097/AUD.0000000000000135 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Laback B, Egger K, Majdak P. Perception and coding of interaural time differences with bilateral cochlear implants. Hear Res. 2015;322:138–150. 10.1016/j.heares.2014.10.004 [DOI] [PubMed] [Google Scholar]
  • 34.Kulkarni PN, Pandey PC, Jangamashetti DS. Binaural dichotic presentation to reduce the effects of spectral masking in moderate bilateral sensorineural hearing loss. Int J Audiol. 2012;51:334–344. 10.3109/14992027.2011.642012 [DOI] [PubMed] [Google Scholar]
  • 35.Aronoff JM, Stelmach J, Padilla M, Landsberger DM. Interleaved processors improve cochlear implant patients' spectral resolution. Ear Hear. 2016;37(2):e85–e90. 10.1097/AUD.0000000000000249 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Kong YY, Braida LD. Cross-frequency integration for consonant and vowel identification in bimodal hearing. J Speech Lang Hear Res. 2011;54(3):959–980. 10.1044/1092-4388(2010/10-0197) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Luo X, Fu QJ. Contribution of low-frequency acoustic information to Chinese speech recognition in cochlear implant simulations. J Acoust Soc Am. 2006;120(4):2260–2266. 10.1121/1.2336990 [DOI] [PubMed] [Google Scholar]
  • 38.Zhou N, Pfingst BE. Psychophysically based site selection coupled with dichotic stimulation improves speech recognition in noise with bilateral cochlear implants. J Acoust Soc Am. 2012;132:994–1008. 10.1121/1.4730907 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Aronoff JM, Amano-Kusumoto A, Itoh M, Soli SD. The effect of interleaved filters on normal hearing listeners' perception of binaural cues. Ear Hear. 2014;35(6):708–710. 10.1097/AUD.0000000000000060 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Fu QJ, Galvin JJ, Wang X. Integration of acoustic and electric hearing is better in the same ear than across ears. Sci Rep. 2017a; 7(1):12500 10.1038/s41598-017-12298-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Fu QJ, Galvin JJ 3rd, Wang X. Effect of carrier bandwidth on integration of simulations of acoustic and electric hearing within or across ears. J Acoust Soc Am. 2017b; 142(6): EL561 10.1121/1.5017530 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Shannon RV, Zeng FG, Kamath V, Wygonski J, Ekelid M. Speech recognition with primarily temporal cues. Science. 1995. October 13;270(5234):303–304. 10.1126/science.270.5234.303 [DOI] [PubMed] [Google Scholar]
  • 43.Greenwood DD. A cochlear frequency-position function for several species– 29 years later. J Acoust Soc Am. 1990; 87:2592–2605. 10.1121/1.399052 [DOI] [PubMed] [Google Scholar]
  • 44.Karsten SA, Turner CW, Brown CJ, Jeon EK, Abbas PJ, Gantz BJ. Optimizing the combination of acoustic and electric hearing in the implanted ear. Ear Hear. 2013; 34:142–150. 10.1097/AUD.0b013e318269ce87 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Kral A, Tillein J, Heid S, Klinke R, Hartmann R. Cochlear implants: cortical plasticity in congenital deprivation. Prog Brain Res. 2006;157:283–313. 10.1016/s0079-6123(06)57018-9 [DOI] [PubMed] [Google Scholar]
  • 46.Gordon KA, Wong DD, Valero J, Jewell SF, Yoo P, Papsin BC. Use it or lose it? Lessons learned from the developing brains of children who are deaf and use cochlear implants to hear. Brain Topogr. 2011;24(3–4):204–219. 10.1007/s10548-011-0181-2 [DOI] [PubMed] [Google Scholar]
  • 47.Micheyl C, Oxenham AJ. Comparing models of the combined-stimulation advantage for speech recognition. J Acoust Soc Am. 2012; 131:3970–3980. 10.1121/1.3699231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Fu QJ, Shannon RV, Galvin JJ 3rd. Perceptual learning following changes in the frequency-to-electrode assignment with the Nucleus-22 cochlear implant. J Acoust Soc Am. 2002; 112: 1664–1674. 10.1121/1.1502901 [DOI] [PubMed] [Google Scholar]
  • 49.Gifford RH, Davis TJ, Sunderhaus LW, Menapace C, Buck B, Crosson J, et al. Combined electric and acoustic stimulation with hearing preservation: effect of cochlear implant low-frequency cutoff on speech understanding and perceived listening difficulty. Ear Hear. 2017; 38(5):539–553. 10.1097/AUD.0000000000000418 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Li T, Galvin JJ 3r d, Fu QJ. Interactions between unsupervised learning and the degree of spectral mismatch on short-term perceptual adaptation to spectrally shifted speech. Ear Hear. 2009; 30: 238–249. 10.1097/AUD.0b013e31819769ac [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Fu Q-J, Nogaki G, Galvin JJ III. Auditory training with spectrally shifted speech: an implication for cochlear implant users' auditory rehabilitation. J Assoc Res Otolaryngol. 2005; 6(2): 180–189. 10.1007/s10162-005-5061-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Faulkner A. Adaptation to distorted frequency-to-place maps: implications of simulations in normal listeners for cochlear implants and electroacoustic stimulation. Audiol Neurootol. 2006; 11(Suppl 1): 21–26. 10.1159/000095610 [DOI] [PubMed] [Google Scholar]
  • 53.Rosen S, Faulkner A, Wilkinson L. Adaptation by normal listeners to upward spectral shifts of speech: implications for cochlear implants. J Acoust Soc Am. 1999;106(6):3629–3636. 10.1121/1.428215 [DOI] [PubMed] [Google Scholar]
  • 54.Yang HI, Zeng FG. Reduced acoustic and electric integration in concurrent-vowel recognition. Sci Rep. 2013; 3:1419 10.1038/srep01419 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Svirsky MA, Silveira A, Neuburger H, Teoh SW, Suárez H. Long-term auditory adaptation to a modified peripheral frequency map. Acta Otolaryngol. 2004; 124:381–386. [PubMed] [Google Scholar]
  • 56.Sagi E, Fu QJ, Galvin JJ 3rd, Svirsky MA. A model of incomplete adaptation to a severely shifted frequency-to-electrode mapping by cochlear implant users. J Assoc Res Otolaryngol. 2010; 11(1):69–78. 10.1007/s10162-009-0187-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Svirsky MA, Talavage TM, Sinha S, Neuburger H, Azadpour M. Gradual adaptation to auditory frequency mismatch. Hear Res. 2015; 322, 163–170. 10.1016/j.heares.2014.10.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Labadie RF, Noble JH, Hedley-Williams AJ, Sunderhaus LW, Dawant BM, Gifford RH. Results of postoperative, CT-based, electrode deactivation on hearing in prelingually deafened adult cochlear implant recipients. Otol Neurotol. 2016; 37(2):137–145. 10.1097/MAO.0000000000000926 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Noble JH, Hedley-Williams AJ, Sunderhaus L, Dawant BM, Labadie RF, Camarata SM, et al. Initial results with image-guided cochlear implant programming in children. Otol Neurotol. 2016; 37(2):e63–9. 10.1097/MAO.0000000000000909 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Dalbert A, Pfiffner F, Röösli C, Thoele K, Sim JH, Gerig R, et al. Extra- and intra-cochlear electrocochleography in cochlear implant recipients. Audiol Neurootol. 2015; 20(5):339–348. 10.1159/000438742 [DOI] [PubMed] [Google Scholar]
  • 61.Tejani VD, Abbas PJ, Brown CJ, Woo J. An improved method of obtaining electrocochleography recordings from Nucleus Hybrid cochlear implant users. Hear Res. 2019; 373:113–120. 10.1016/j.heares.2019.01.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Gifford RH, Dorman MF, Sheffield SW, Teece K, Olund AP. Availability of binaural cues for bilateral implant recipients and bimodal listeners with and without preserved hearing in the implanted ear. Audiol Neurootol. 2014;19(1):57–71. 10.1159/000355700 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Gifford RH, Driscoll CL, Davis TJ, Fiebig P, Micco A, Dorman MF. A within-subject comparison of bimodal hearing, bilateral cochlear implantation, and bilateral cochlear implantation with bilateral hearing preservation: high-performing patients. Otol Neurotol. 2015;36(8):1331–1337. 10.1097/MAO.0000000000000804 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Hussain Md Abu Nyeem

18 Aug 2020

PONE-D-20-21206

Effects of noise on the integration of acoustic and electric hearing within and across ears

PLOS ONE

Dear Dr. Fu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

This manuscript is structured well with useful findings on the simulated electric hearing (EH) and acoustic hearing (AH) in the presence of noise in both the cases, hearing preservation implantation, and bimodal hearing. Please see the Editor's and reviewers' comments below this email.

Please submit your revised manuscript by Oct 02 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Hussain Md Abu Nyeem, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments:

While the presented findings on the effect of tonotopic mismatch hold promises for both the acoustic and electric hearing (AEH), its effect on the electric-acoustic stimulation (EAS) and bimodal hearing in the presence of noise appear to be slightly obvious. In other words, the presence of noise is more likely to contribute to the tonotopic mismatch, and thus, specific noises can be distinguished by their effect to make a more definite sense here. For example, we may want to know- how are the specific noise (with its properties or model) responsible for both EAS and bimodal hearing to be ‘highly’ sensitive to tonotopic mismatch? With the analysis of a few typical noise types in bimodal listening, the sensitivity variation of EAS and bimodal hearing thus may be contrasted to conclude their trend of sensitivity variation in noisy conditions.

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Methods section, please provide additional information about the demographic details of your participants. Please ensure you have provided sufficient details to replicate the analyses such as: a)  a description of any inclusion/exclusion criteria that were applied to participant inclusion in the analysis, b) a table of relevant demographic details, c) a statement as to whether your sample can be considered representative of a larger population, d) how participants were recruited to the study. In addition, please ensure you have described the statistical analyses within your Methods section.

3. Thank you for stating the following financial disclosure:

"NO"

At this time, please address the following queries:

  1. Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution.

  2. State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

  3. If any authors received a salary from any of your funders, please state which authors and which funders.

  4. If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. Thank you for stating the following in your Competing Interests section: 

"NO"

Please complete your Competing Interests on the online submission form to state any Competing Interests. If you have no competing interests, please state "The authors have declared that no competing interests exist.", as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now

 This information should be included in your cover letter; we will change the online submission form on your behalf.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The simulation of CI signal processing tested on listeners with normal hearing serves as one method for evaluating speech understanding when the speech signal has been distorted using the noise vocoder. This assumes that the auditory system beyond the cochlea is "normal" and not affected by whatever oto-toxic agent led to the hearing loss. Thus, these simulations may, or may not, be good benchmarks for evaluation of such processing schemes. This caveat is rarely discussed in papers such as this one that are designed to evaluate speech processors.

Another small "flaw" is the lack of a reference for speech processing by listeners with normal hearing when the speech has been corrupted or distorted in other ways. For example, what level of performance would be expected if the noise vocoder were eliminated and only band pass filtering and additive noise were used to degrade the speech signals? Would similar results be obtained at SNRs below zero? Are there other models for combining information across channels that could be evaluated and used for comparison with these results?

This manuscript presents useful and timely results that were obtained from carefully designed and conducted experiments. A broader discussion of the implications of the work would be helpful.

Reviewer #2: This manuscript provides a detailed analysis of speech perception results for normal hearing subjects listening to simulations of 'electrical hearing' (EH) as applied in cochlear implant devices, and combinations of low frequency acoustic hearing combined with electrical hearing in the same ear (AEH) and opposite ear (bimodal). Monosyllabic word and phoneme scores are tabulated for quiet and for two competing noise conditions for ten young subjects. Two versions of the acoustic input were assessed in an attempt to elucidate the effect of 'tonotopic mismatch' between the electrical and acoustic hearing. The introduction provides appropriate background to the study and the decisions regarding the technical parameters are well justified. The procedures are clearly described and statistical analyses appear appropriate and comprehensive. I had some concerns about the applicability of simulation studies to the real world clinical environment but this is not overplayed by the authors and the limitations are clearly documented. There are a few typos along the way but the paper is generally well-written. Some parts of the discussion are a little difficult to follow and could be reworked somewhat for clarity. Most of the discussion provides reasonable conjectures about the main results that can be summarized as follows: 1. Low frequency acoustic hearing may provide improved speech perception when added to simulated cochlear implant hearing in either the ipsilateral (AEH) or contralateral ear (bimodal). 2. Tonotopic mismatch was found to have a more detrimental effect for speech perception in quiet for AEH than bimodal simulated listening. This effect appeared to decrease with increasing noise. The main conclusion that 'overall performance was poorer when there was a tonotopic mismatch' seems to be overstating the actual results which showed a loss of advantage for AEH (identified as EAS in the final summary - could be better to stick with the same terminology), but not for bimodal simulations or for AEH in noise. Despite my mainly positive comments relating to this manuscript, I have two issues that I believe need to be reworked. These both relate to the analysis of phoneme and word scores in this study and some conclusions that I believe are not justified. Firstly, by definition, word scores must be less than phoneme scores in this type of testing and the relationship is quite deterministic within the experimental error of these measures. Boothroyd and Nittrouer (1988 JASA 84:101-114) used a probability theory model of speech perception to look at contextual effects and showed that the relationship between phoneme and word scores in monosyllables is of the form - word score/100 = phoneme score/100^E, where E, the exponent, is approximately 3 for nonsense words and approximately 2.4 for real words. The exponent reflects the effective number of individual parts (phonemes) that need to identified to identify the word, and the difference between 3 for nonsense, and 2.4 for real words reflects the lexical context effects - ie., the increased probability of guessing a real word correctly compared to nonsense. A recent study has duplicated this result using real clinical data from cochlear implant users (Au, et al, 2018, Hearing Across the Lifespan (HEAL) conference, Lake Como Italy. see attachment). This relationship between phoneme and word scores in monosyllabic testing means that not only will phoneme scores always be lower than word scores (by differing amounts for different scores as the relationship is non-linear) but changes in word scores from one condition to another will predict changes in phonemes scores for the same conditions fairly accurately. The relevance to this study is that phoneme scores being significantly greater than word scores is claimed at numerous points as being an outcome, but it is not an outcome of the study, it is just a consequence of the more or less fixed relationship between such scores. In addition, another conclusion, that the integration efficiency (IE) factor was larger for phonemes than for words in some conditions does not really make sense as the scores (words and phonemes) are inextricably linked by the above relationship. I believe this outcome is a quirk of the mathematics and that, particularly for low scores, phoneme scores grow more rapidly than word scores (eg. phoneme scores have to improve from 0 to around 25% before word scores move from 0%). Note that the AEH advantages (fig.2) and the IE factors are derived from generally low word scores, particularly for SNR=0dB. I feel that the paper should be reworked removing the discussion and conclusions relating to different effects for word and phoneme scores.

Reviewer #3: This is a review of PONE-D-20-21206 entitled “Effects of noise on the integration of acoustic and electric hearing within and across ears.” The purpose of this study was to describe the integration of simulated electric hearing (EH) and acoustic hearing (AH) in the presence of noise within an ear, as is the case with hearing preservation implantation, or across ears (bimodal hearing). They investigated cases of simulated EH+AH with and without tonotopic mismatch. This manuscript was relatively straightforward, well written, and the results hold high clinical application. There are a few items requiring attention which would significantly improve the manuscript and its overall impact.

General comments:

A major concern with the study design and interpretation of results is that EAS conditions (EH+AH in the same ear) do not include integration of EH+AH across ears. However, most EAS patients also have acoustic hearing in the contralateral ear. As such, this would result in potential interference both peripherally (within an ear) and centrally (across ears). Though this isn’t a critical flaw, it is something that must be addressed as it impacts the clinical impact of the study. It would also be interesting to mention this as a point for additional study in future simulation studies.

Point-by-point comments:

Lines 75-77: An explanation here would be valuable regarding how qualitative differences impact integration (include references). Also, there is no mention here regarding binaural integration (particularly for binaural cues) and its influence on source segregation. As mentioned in “General comments,” both would be present with EAS listeners who were combining AH+EH within an ear and across ears simultaneously.

Line 97: “inherent” is included twice in this sentence

Lines 111-112: Could this also be pertinent for EAS (CI + binaural acoustic hearing)?

Figure 1: Red and green bars next to each other are difficult to distinguish for individuals with color blindness.

Lines 290-291: This should not be unexpected.

Lines 337-346: Real bimodal listeners typically have chronic listening experience in a “mismatch” condition. Also, a singular simulated insertion depth was used in the current study whereas most studies with real bimodal (or EAS) listeners include various different insertion depths. Both issues should be mentioned here as potential reasons for the discrepancy in findings.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

<gdiv></gdiv>

Attachment

Submitted filename: HEAL Poster - linguistic context effects final.ppt

PLoS One. 2020 Oct 15;15(10):e0240752. doi: 10.1371/journal.pone.0240752.r002

Author response to Decision Letter 0


9 Sep 2020

To the editor:

Please find enclosed our revised submission “Effects of noise on integration of acoustic and electric hearing within and across ears”.

We thank the editor and the reviewers and have greatly revised the MS with their comments in mind. Major changes include:

1. Removal of phoneme data. We had originally included the phoneme data because it offered better comparison to the vowel recognition data in our previous related studies. However, there seemed to be some confusion regarding percent correct, AEH advantage (the difference in performance when AH was added to EH, relative to EH-only), and IE [the ratio between observed AEH and predicted AEH (according to the AH and EH only performance]. It was always true that phoneme recognition was higher than word recognition. However, it seemed confusing that AEH advantage and IE might be greater for words than for phonemes. This had nothing to do with the relationship between phoneme and word recognition, but rather how adding AH to EH benefitted performance. As Rev. 2 points out, the lower overall AH and EH word scores allowed more “room for improvement.” But it appears that the phoneme data distracted from the main story, which was about how noise, the degree of tonotopic mismatch, and the distribution of AH across ears might affect integration of AH and EH. Accordingly, we have deleted all phoneme data, have revised the three figures, and have re-analyzed the data after excluding the phoneme data. We feel that this does not change the story at all, but may help reduce confusion.

2. We have expanded the Introduction and Discussion sections according to reviewers’ suggestions.

3. All figures are now in gray scale

We hope you find this revision acceptable, and let me know if you need further information.

Sincerely,

Qian-jie Fu

Additional Editor Comments:

While the presented findings on the effect of tonotopic mismatch hold promises for both the acoustic and electric hearing (AEH), its effect on the electric-acoustic stimulation (EAS) and bimodal hearing in the presence of noise appear to be slightly obvious. In other words, the presence of noise is more likely to contribute to the tonotopic mismatch, and thus, specific noises can be distinguished by their effect to make a more definite sense here. For example, we may want to know- how are the specific noise (with its properties or model) responsible for both EAS and bimodal hearing to be ‘highly’ sensitive to tonotopic mismatch? With the analysis of a few typical noise types in bimodal listening, the sensitivity variation of EAS and bimodal hearing thus may be contrasted to conclude their trend of sensitivity variation in noisy conditions.

>>Actually, we expected noise to interact with tonotopic mismatch with bimodal or EAS hearing similarly as in quiet (perhaps an “obvious” expectation). But this was not the case. There was a very strong effect of tonotopic mismatch for EAS in quiet, but not for bimodal listening. When noise was added, the effects of tonotopic mismatch were similar for EAS and bimodal listening. Noise does not contribute to tonotopic mismatch per se, but it appears to similarly mask the effects of tonotopic mismatch with EAS and bimodal listening. Of course, overall performance also declines with noise, whether or not there is a tonotopic mismatch. In the present study, speech and noise were combined before subsequent processing; as such the effects of noise within AH, EH-mismatched, and EH matched were spectrally limited within the bandwidth of each of these conditions. Different types of noise (multi-talker babble, gated noise, etc.) would be similarly processed by these frequency input/output function. It is possible that AEH benefit might have differed between steady noise (as used in the present study), competing speech or multi-talker babble. Perhaps a good comparison for future work….

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

>>We have proofread the paper to ensure that the style conforms to PLOS One guidelines.

2. In your Methods section, please provide additional information about the demographic details of your participants. Please ensure you have provided sufficient details to replicate the analyses such as: a) a description of any inclusion/exclusion criteria that were applied to participant inclusion in the analysis, b) a table of relevant demographic details, c) a statement as to whether your sample can be considered representative of a larger population, d) how participants were recruited to the study. In addition, please ensure you have described the statistical analyses within your Methods section.

>> We have added: Ten NH subjects (5 males and 5 females) participated in this study. Participants were recruited from UCLA main campus. Recruitment postings were placed on the Digest News & Events website at the UCLA David Geffen School of Medicine (https://digest.dgsom.ucla.edu/). Inclusion criteria was that subjects have normal AH in both ears for all audiometric frequencies, that they were at least 18 years old, and were native speakers of American English. The mean age at testing was 25 years (range: 23-30 years). All subjects had thresholds <20 dB HL for audiometric frequencies 250, 500, 1000, 2000, 4000, and 8000 Hz in each ear. While the subject sample is small, the normal AH status in each ear presumably reflected uniform neural survival. As such, the various spectral degradation manipulations could be applied with the expectation that the effects would be similar across subjects.”

3. Thank you for stating the following financial disclosure:

"NO"

At this time, please address the following queries:

a. Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution.

b. State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

c. If any authors received a salary from any of your funders, please state which authors and which funders.

d. If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

>> We have added this information in the revised cover letter.

4. Thank you for stating the following in your Competing Interests section:

"NO"

Please complete your Competing Interests on the online submission form to state any Competing Interests. If you have no competing interests, please state "The authors have declared that no competing interests exist.", as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now

This information should be included in your cover letter; we will change the online submission form on your behalf.

>> We have included this information in the revised cover letter.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

>>We have added the Supporting information file at the end of the MS.

5. Review Comments to the Author

Reviewer #1: The simulation of CI signal processing tested on listeners with normal hearing serves as one method for evaluating speech understanding when the speech signal has been distorted using the noise vocoder. This assumes that the auditory system beyond the cochlea is "normal" and not affected by whatever oto-toxic agent led to the hearing loss. Thus, these simulations may, or may not, be good benchmarks for evaluation of such processing schemes. This caveat is rarely discussed in papers such as this one that are designed to evaluate speech processors.

>>You make an important point…We have added to the end of the Simulations section of the Methods: It should be noted that the present simulations (and indeed, all CI simulations) do not capture potential deficits in speech processing at higher levels of the auditory system that may be impaired by long-term hearing loss. For example, previous studies have shown that for prolonged auditory deprivation, cortical deficits and/or slower cortical development that may persist even after cochlear implantation [45-46]. Such central deficits were not captured by the present simulations, which were more intended to simulate different peripheral limitations, but perceived with normal hearing and (presumably) normal higher-level auditory processing.”

Another small "flaw" is the lack of a reference for speech processing by listeners with normal hearing when the speech has been corrupted or distorted in other ways. For example, what level of performance would be expected if the noise vocoder were eliminated and only band pass filtering and additive noise were used to degrade the speech signals? Would similar results be obtained at SNRs below zero? Are there other models for combining information across channels that could be evaluated and used for comparison with these results?

>>We have added: “Other studies have shown advantages for combining non-overlapping spectral information across ears. For example, Kulkarni et al. [34] showed a binaural advantage in spectral resolution for AH listeners when spectral bands were combined across ears. Similarly, Aronoff et al. [35] found that, for bilateral CI listeners, spectral resolution was significantly better when distributing 12 spectral channels across ears than within ears. Kong and Braida [36] found that the benefit of combining low-frequency AH with contralateral broadband EH was greater for vowels than for consonants, for both real and simulated CI patients. Luo and Fu [37], testing Chinese Mandarin-speaking NH participants listening to bimodal CI simulations, found that acoustic information < 500 Hz was more important for lexical tone perception (where voice pitch cues are important), and that acoustic information > 500 Hz was important for Mandarin phoneme recognition. The authors also found no difference for speech understanding in noise whether or not there was inter-aural spectral overlap between the simulated AH and EH. “Zipper” CI signal processing, where independent electrode information is provided to each ear, has shown binaural advantages in temporal processing and/or speech performance relative to monaural performance [38-39].”

This manuscript presents useful and timely results that were obtained from carefully designed and conducted experiments. A broader discussion of the implications of the work would be helpful.

>>We have somewhat expanded the “Clinical implications and limitations” section of the Discussion

Reviewer #2:

There are a few typos along the way but the paper is generally well-written.

>>We have carefully proof-read the revision and hopefully have removed typos and grammatical errors.

Some parts of the discussion are a little difficult to follow and could be reworked somewhat for clarity.

>>Given that we have removed Discussion of the phoneme data (see below), we have greatly revised the Results and Discussion.

Despite my mainly positive comments relating to this manuscript, I have two issues that I believe need to be reworked. These both relate to the analysis of phoneme and word scores in this study and some conclusions that I believe are not justified. Firstly, by definition, word scores must be less than phoneme scores in this type of testing and the relationship is quite deterministic within the experimental error of these measures. Boothroyd and Nittrouer (1988 JASA 84:101-114) used a probability theory model of speech perception to look at contextual effects and showed that the relationship between phoneme and word scores in monosyllables is of the form - word score/100 = phoneme score/100^E, where E, the exponent, is approximately 3 for nonsense words and approximately 2.4 for real words. The exponent reflects the effective number of individual parts (phonemes) that need to identified to identify the word, and the difference between 3 for nonsense, and 2.4 for real words reflects the lexical context effects - ie., the increased probability of guessing a real word correctly compared to nonsense. A recent study has duplicated this result using real clinical data from cochlear implant users (Au, et al, 2018, Hearing Across the Lifespan (HEAL) conference, Lake Como Italy. see attachment). This relationship between phoneme and word scores in monosyllabic testing means that not only will phoneme scores always be lower than word scores (by differing amounts for different scores as the relationship is non-linear) but changes in word scores from one condition to another will predict changes in phonemes scores for the same conditions fairly accurately. The relevance to this study is that phoneme scores being significantly greater than word scores is claimed at numerous points as being an outcome, but it is not an outcome of the study, it is just a consequence of the more or less fixed relationship between such scores.

>>” This relationship between phoneme and word scores in monosyllabic testing means that not only will phoneme scores always be lower than word scores…” Maybe a typo, but word scores should always be lower than phoneme scores, as you point out above, as is shown in your poster, and is shown in our data. We believe there may be some confusion regarding the relationships among word recognition, phoneme recognition, AEH advantage (the difference in performance when AH was added to EH), and IE (the ratio between the observed and predicted performance). AEH and IE are both derivative measures that look at different aspects of how AH and EH are combined within or across ears. It is entirely possible that adding AH may have benefitted word recognition (which had lower EH-only scores) more than phoneme recognition (which had higher EH-only scores). IE was also much better for words than phonemes, due to the poorer AH and EH performance. Thus, while word recognition was always poorer than phoneme recognition, word recognition may have benefitted more greatly from the addition of AH due to the generally poorer performance.

We realize that there may be unintended confusion when comparing recognition scores, AEH, and IE between words and phonemes at the various SNR. We have decided to remove all mention of phoneme scores in the revision. We had originally included the phoneme scores as they were more comparable to the vowel recognition measures in our previous study in quiet. As the phoneme scores do not add much to the general pattern of results, we now only present word recognition scores, as well as AEH and IE for word recognition.

In addition, another conclusion, that the integration efficiency (IE) factor was larger for phonemes than for words in some conditions does not really make sense as the scores (words and phonemes) are inextricably linked by the above relationship. I believe this outcome is a quirk of the mathematics and that, particularly for low scores, phoneme scores grow more rapidly than word scores (eg. phoneme scores have to improve from 0 to around 25% before word scores move from 0%). Note that the AEH advantages (fig.2) and the IE factors are derived from generally low word scores, particularly for SNR=0dB. I feel that the paper should be reworked removing the discussion and conclusions relating to different effects for word and phoneme scores.

>>There must be some misunderstanding, because we clearly state on line 390 of the original MS: “As shown in Fig 3, IE was much stronger for words than for phonemes.” But as noted above, we have removed all mention of phoneme scores in the revision.

Reviewer #3:

A major concern with the study design and interpretation of results is that EAS conditions (EH+AH in the same ear) do not include integration of EH+AH across ears. However, most EAS patients also have acoustic hearing in the contralateral ear. As such, this would result in potential interference both peripherally (within an ear) and centrally (across ears). Though this isn’t a critical flaw, it is something that must be addressed as it impacts the clinical impact of the study. It would also be interesting to mention this as a point for additional study in future simulation studies.

>>We have added to the “Clinical implications and limitations” section of the Discussion: Also, there are increasing numbers of BiEAS CI users, as well as single-sided deaf CI users (normal AH in one ear, CI in the other ear). CI users’ speech and localization performance has been shown to greatly benefit when there is limited AH in both ears [62-63]. It is unclear how the availability of residual AH in both ears might have affected the present pattern of results with the present CI simulations. In future studies, it would be worthwhile to evaluate the effects of tonotopic mismatch in EH in the context of residual AH in both ears and CI in one or both ears.”

Lines 75-77: An explanation here would be valuable regarding how qualitative differences impact integration (include references). Also, there is no mention here regarding binaural integration (particularly for binaural cues) and its influence on source segregation. As mentioned in “General comments,” both would be present with EAS listeners who were combining AH+EH within an ear and across ears simultaneously.

>>Note that “BiEAS” was not simulated in this study. To clarify, we have revised a section of the first paragraph as: “Arguably, one of the greatest improvements in CI outcomes has come from combining electric hearing (EH) from the CI with acoustic hearing (AH) in the same ear (electric-acoustic stimulation, or EAS) or in opposite ears (bimodal listening). Note that these definitions do not include BiEAS (residual AH in both ears, CI in one or both ears)].”

In response to the other comments, we have added: “Sound quality differences between AH and EH may also affect integration of acoustic and electric hearing. To our knowledge, such sound quality differences have not been systematically studied in CI users, in quiet or noise. If the peripheral representations are sufficiently different, it is likely that sound quality may also be different across ears. While not the same as sound quality measurements across ears, inter-aural pitch matching experiments have shown difficulties matching the pitch of an acoustic signal in one ear [which contains temporal fine-structure (TFS) information that is important to pitch perception] to the pitch associated with stimulation of a single electrode [which does not contain TFS information and stimulates a broader cochlear region]. Work by Reiss and colleagues [28-30] has shown that inter-aural frequency mismatch can limit binaural fusion of acoustic and electric stimulation across ears or bilateral electric stimulation. Such inter-aural frequency mismatch can also limit perception of inter-aural timing differences (ITDs) that are important for sound source localization and image fusion [31-33].”

Line 97: “inherent” is included twice in this sentence

>>Corrected

Lines 111-112: Could this also be pertinent for EAS (CI + binaural acoustic hearing)?

>>In this paper and many others, EAS refers to combined electric and acoustic hearing in the same ear. We have added to the Discussion: “Note that BiEAS listening were not simulated in this study. It is possible that IE might have been even greater for BiEAS, as AH and EH might have only been marginally affected by bilateral presentation, but AEH performance would likely be even better than presently observed for bimodal or EAS listening. Gifford et al. [49] found a consistent advantage for BiEAS over bimodal listening in real CI users.” And later: “Also, there are increasing numbers of BiEAS CI users, as well as single-sided deaf CI users (normal AH in one ear, CI in the other ear). CI users’ speech and localization performance has been shown to greatly benefit when there is limited AH in both ears [62-63]. It is unclear how the availability of residual AH in both ears might have affected the present pattern of results with the present CI simulations.”

Figure 1: Red and green bars next to each other are difficult to distinguish for individuals with color blindness.

>>Figure 1 now in gray scale

Lines 290-291: This should not be unexpected.

>>Revised “As expected, performance was significantly better for phoneme than for word recognition. However, in terms of AEH advantage, there was a significant interaction between outcome measure and the degree of tonotopic mismatch in EH.”

Lines 337-346: Real bimodal listeners typically have chronic listening experience in a “mismatch” condition. Also, a singular simulated insertion depth was used in the current study whereas most studies with real bimodal (or EAS) listeners include various different insertion depths. Both issues should be mentioned here as potential reasons for the discrepancy in findings.

>>We had somewhat discussed these issues in the “Clinical implications and limitations” section of the Discussion of the original MS, but we have also added here in response to the reviewer’s comments. Paragraph revised as: The present data show that the tradeoff between tonotopic matching and information loss observed in Fu et al. [40-41] for vowel recognition in quiet persisted in noise. This is not consistent with Gifford et al. [49], who found that for real bimodal CI users, the best performance was observed when a wide input acoustic frequency range was used (more speech information, but with more mismatch), rather than a more tonotopically matched range (less information, but with less mismatch). For the present data, performance was always poorer with a tonotopic mismatch. Note that only one electrode insertion depth was used to create the EH tonotopic mismatch in the present study, compared to previous studies where a range of insertion depths/tonotopic mismatches were tested [40]. It is possible that the patterns of results may differ for words in noise (used in this study) and sentences in noise (used in Gifford et al. [49]), where top-down auditory processing might play a greater role. Also, real bimodal CI listeners have much greater experience with tonotopic mismatch in their device. Such adaptation to the mismatch in the EH simulation was not tested in the present study, as only acute performance was measured. It is possible that some short-term training may have reduced the effects of the mismatch [50-53]. It is also possible that the present CI simulations did not capture real CI users’ performance.”

Attachment

Submitted filename: PLOS-R1-response.docx

Decision Letter 1

Hussain Md Abu Nyeem

2 Oct 2020

Effects of noise on integration of acoustic and electric hearing within and across ears

PONE-D-20-21206R1

Dear Dr. Fu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hussain Md Abu Nyeem, Ph.D.

Academic Editor

PLOS ONE

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: My apologies to the authors for typographical errors in my review. Of course, word scores are always lower than phoneme scores - I should have checked my review more closely! Removing all mention of the phoneme scores is an interesting way of dealing with some of the confusing findings. Maybe I would have gone the other way and removed the word scores, but I will not argue about the approach. Thank you for your careful revision.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Hussain Md Abu Nyeem

6 Oct 2020

PONE-D-20-21206R1

Effects of noise on integration of acoustic and electric hearing within and across ears

Dear Dr. Fu:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Hussain Md Abu Nyeem

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. S1 File contains the raw word percent correct data for all subjects and listening conditions.

    (XLSX)

    Attachment

    Submitted filename: HEAL Poster - linguistic context effects final.ppt

    Attachment

    Submitted filename: PLOS-R1-response.docx

    Data Availability Statement

    All relevant data are within its supporting information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES