Abstract
In the case of hearing loss, cochlear implants (CI) allow for the restoration of hearing. Despite the advantages of CIs for speech perception, CI users still complain about their poor perception of their auditory environment. Aiming to assess non-verbal auditory perception in CI users, we developed five listening tests. These tests measure pitch change detection, pitch direction identification, pitch short-term memory, auditory stream segregation, and emotional prosody recognition, along with perceived intensity ratings. In order to test the potential benefit of visual cues for pitch processing, the three pitch tests included half of the trials with visual indications to perform the task. We tested 10 normal-hearing (NH) participants with material being presented as original and vocoded sounds, and 10 post-lingually deaf CI users. With the vocoded sounds, the NH participants had reduced scores for the detection of small pitch differences, and reduced emotion recognition and streaming abilities compared to the original sounds. Similarly, the CI users had deficits for small differences in the pitch change detection task and emotion recognition, as well as a decreased streaming capacity. Overall, this assessment allows for the rapid detection of specific patterns of non-verbal auditory perception deficits. The current findings also open new perspectives about how to enhance pitch perception capacities using visual cues.
Keywords: hearing loss, pitch perception, auditory scene analysis, prosody, audiovisual integration
1. Introduction
Auditory cognition allows us to perceive our environment and react properly to external stimuli. Communication via language is essential, but non-verbal auditory cognition also plays a primary role in the understanding of perceived stimuli, including prosody. Indeed, in speech communication, prosody perception is essential to understand the intentions and emotions of the speaker [1]. Moreover, non-verbal auditory abilities allow us to detect auditory signals in noise, and to better perceive and analyze the auditory scene [2]. Music perception also relies on our abilities to process and determine melody and harmony in non-verbal auditory signals [3].
Non-verbal auditory perception can be altered in the case of peripheral hearing loss [4,5,6]. When the hearing loss is profound, one or two cochlear implants can partially restore hearing capacities. However, auditory perception in cochlear implant (CI) users can remain impaired due to the limited frequency resolution of the implant [7,8]. Indeed, the technical constraints of an implant do not allow for a fine-grained decomposition of the auditory signal compared to the decomposition of a healthy cochlea [9]. As a result, CI users still have difficulties hearing speech in noise, understanding prosody and perceiving music, even though post-lingually deafened CI users have a strong desire to enjoy music again. These non-verbal auditory perception difficulties are assumed to be related to a pitch perception deficit resulting from the degraded auditory input [10]. In particular, as CIs do not always stimulate the apical part of the cochlea and only a frequency range around 150 to 8500 Hz is transmitted (depending on the CI device and settings), the poor perception of low frequencies could negatively impact the sound quality for CI users [10]. For instance, music and emotional prosody contain spectral energy outside this range [11,12]. In order to simulate the pitch perception deficit of CI users in normal-hearing (NH) listeners, a signal-processing tool referred to as a “vocoder” can be used [13]. Similarly to the CI, vocoders filter the sounds in frequency bands and extract the temporal envelope of the sound [14]. Consequently, the pitch discrimination thresholds measured in NH participants with vocoded sounds are higher than those in NH participants with original sounds (meaning non-vocoded sounds), showing similar pitch deficits to CI users [9,13,14]. Here, we aim to assess, using a short series of listening tests, the pattern of non-verbal auditory perception deficits in CI users and NH participants with vocoded sounds, compared to NH listeners with original sounds. We thus review below the documented deficits of non-verbal auditory perception in CI users.
1.1. Music Perception
Pitch perception has a key role in music processing: pitch discrimination is essential to perceive melody and harmony in music [14,15]. In CI users, deficits of music processing affect multiple dimensions. Musical features relating to pitch, such as timbre and harmonicity, are less well perceived by CI users than NH listeners [16,17,18,19]. Musical time processing was first thought to be relatively preserved in CI users [20,21,22]. However, when subjected to complex musical tasks (with pitch variations, not only sequences of tones), CI users do not perform as well on beat recognition and entrainment [22].
These deficits contribute to a general depreciation of music and a decreased quality of music listening in CI users [23], with a correlation between pitch perception abilities and music appreciation [24]. Regarding musical emotions, CI users showed a deficit for recognition compared to NH listeners, especially for sadness [16,25,26,27] or fear stimuli [28]. Some studies showed different arousal scores, but there were similar valence scores in CI users compared to NH listeners [8,25], suggesting that the musical emotion processing deficit is not a general deficit of music perception. Despite these limitations, some CI users still enjoy music, and have a strong desire to enhance this pleasure of music listening [29,30]. Overall, these results suggest that musical emotion recognition is possible in CI, but depends on features of the implantation as well as of the patient him/herself [31]. Music processing and appreciation might depend on the time of deafness and implantation: post-lingually deafened and implanted patients complain more about music depreciation compared to early-deafened late-implanted patients or prelingually implanted children [29]. Furthermore, some studies suggest that CI users might have implicit processing of pitch that could help them to enjoy music [32].
Overall, studies of music perception report that music appreciation remains poor in CI users [8,16,23,24,25,26,27,28,29]. Even if it might depend on previous musical knowledge [31], music perception—especially relating to pitch [24,33]—could possibly still be improved with CIs despite the technical limitations thanks to implicit knowledge related to pitch [29,32].
1.2. Prosody Perception
Even though speech in silent environments is quite well perceived by CI users, the pitch deficit still limits their non-verbal auditory perception in speech signals. For intentional prosody, CI users have demonstrated poor perception abilities [34,35] as well as a deficit for production [35]. This deficit was associated with difficulties in perceiving word and sentence stress [36]. For emotional prosody, CI users have deficits when visual cues are unavailable [28,37]. These difficulties are reflected in the electrophysiological correlates of emotional prosody processing, with a decreased N1-P2 in response to emotional bursts, reflecting an altered encoding of speech emotions in CI users [38]. However, it seems that this deficit could be partially compensated with enhanced musical exposure and training [39].
1.3. Auditory Scene Perception
Even when speech comprehension in a quiet environment might be quite effective, CI users still experience difficulties in hearing and understanding speech in noise [40,41,42]. Nevertheless, research investigating speech-in-noise perception has shown that a CI is helpful for enhancing speech perception in noisy situations [43].
As poor verbal working memory and lexical ability can limit speech perception in noise [44], tasks without speech were designed in order to better assess specific difficulties for hearing-in-noise, such as the Music-in-noise task developed for NH participants and musicians by Coffey et al. [45]. Hearing-in-noise perception mostly relies on the ability of the listener to separate auditory sources and to focus on the relevant source. Moreover, this ability seems not only to be associated with age-related hearing loss but also to depend on the cognitive abilities of participants, such as cognitive decline and dementia [46]. Streaming segregation tasks, initially developed for NH participants, allow us to determine how well listeners can discriminate between two auditory sources [47]. This segregation of auditory sources relies on the ability of the participant to spectrally separate these two sources [48]. Testing NH participants with vocoded sounds revealed that when the fundamental frequency (F0) discrimination becomes poorer, the segregation between two simultaneous sounds becomes very difficult [14,49], suggesting that this task could be a good index of CI users’ difficulties. Several studies investigating stream segregation in CI users revealed that they have increased perception of one stream, meaning that their abilities to distinguish several auditory sources might be reduced. Moreover, by using stream segregation tasks requiring the detection of rhythmic abnormalities, instead of the detection of frequency differences, as in classical stream segregation tasks [47,48,49,50], some studies measured indirectly the segregation abilities of CI users. These results showed that CI users seem not to experience automatic stream segregation [42,51,52,53,54]. CI users’ stream segregation performance is even worse when the background information is fluctuating, for example in a noisy environment [14]. Overall, the hearing-in-noise capacities of CI users remain a great challenge that needs to be addressed. Stream segregation tasks appear to be an efficient means of determining hearing-in-noise capacities in CI users, especially regarding the relevant pitch discrimination capacities [15].
1.4. Enhancing Pitch Perception with Visual Information
As reviewed above, non-verbal auditory perception is still a great challenge for CI users, and several studies have been conducted to improve this perception. It was suggested that cerebral plasticity could be induced in CI users (pre- and post-lingually) after implantation [7,55,56]. In particular, cross-modal plasticity in CI users could help to enhance their auditory capacities if correctly guided [7,57].
Indeed, in the perception of environmental stimuli, multisensory interactions are ubiquitous. For instance, the McGurk effect shows that the integration of visual and auditory information is essential for speech perception [58]. Several studies have shown the benefit of multisensory integration for perception rehabilitation [59,60,61]. Some findings suggest that multisensory integration is stronger when one of the sensory modalities is deficient. For example, in participants with reduced visual acuity, audiovisual interactions led to the improvement of visual detection thresholds beyond visual-only performance, which was not observed in control participants [62]. Similarly, in participants with a pitch processing deficit, such as in individuals with congenital amusia, visual stimulations helped them to improve their performance in an auditory pitch task [63]. Numerous studies have investigated audiovisual integration in CI users: the results using speech stimuli and faces showed enhanced auditory capacities in CI users thanks to audiovisual integration [64,65,66,67]. This audiovisual enhancement is more efficient in CI users compared to NH participants for all types of speech, even foreign-accented speech [68]. One study on non-linguistic speech processing, using voice gender identification, suggested that CI users are influenced more strongly by visual information than are NH participants, even for tasks not directly related to speech comprehension [69].
1.5. Rationale for the Present Study
Despite numerous studies investigating non-verbal abilities in CI users, few testing tools are available to assess these deficits. Testing batteries have been designed to detect auditory deficits in CI users, such as the Minimal Auditory Capabilities (MAC) [70] and the Minimum Speech Test Battery (MSTB) [71,72]. However, these tests mainly focus on speech intelligibility in quiet and noise, with only one subtest of the MAC testing for environmental sounds [70]. In comparison, the Basic Auditory Skills Evaluation Battery (BASE) [73] allows the detection of auditory deficits in CI users in a rather complete manner, including music processing. This battery is divided into twelve subtests of 3 to 5 min each that CI users can undertake to test their performance. However, the online implementation might create a difficult testing situation for some CI users. It is also noteworthy that all of these batteries require more than one hour of testing.
In the present study, we designed a new short series of listening tests to assess specifically certain aspects of non-verbal auditory perception in NH and CI users. This battery aimed to provide a rapid overview of non-verbal auditory capacities in an individual, and to be used for further auditory perception research on CI users and populations with hearing deficits, utilizing a quick assessment on a touchscreen tablet. This battery measures non-verbal auditory perception capacities with five tests: (1) pitch change detection [63], (2) pitch direction identification [74], (3) short-term memory for pitch [75], (4) stream segregation [50], and (5) emotional prosody [76]. Most of these tasks were selected on the basis of previous research investigating pitch processing in congenital amusia. Congenital amusia is a deficit in music perception and production that is associated with pitch perception [77] and pitch memory deficits [75,78,79,80,81]. These deficits have been traced down to an impaired fronto-temporal network which is involved in particular in pitch encoding [78,82,83]. Deficits in pitch change detection [84], pitch direction identification [85,86], and pitch short-term memory tasks [78,80,81] are hallmarks of the condition of congenital amusia, and we thus selected these three tasks as candidates to uncover the functioning of the combination of the peripheral auditory system and the cortical fronto-temporal network involved in pitch-related auditory cognition. Two further tests allow for the assessment of prosody perception and auditory scene analysis, as these two abilities are central in auditory non-verbal cognition and are altered in CI users, as is prosody perception in amusic participants [1,76,87,88,89]. Based on prior research with CI users on pitch and music processing, as well as on prosody and auditory scene analysis, we expected to observe deficits in CI users for all five tasks of the battery.
Finally, we also tested whether visual cues could improve the pitch processing of CI users, as well as the processing of full and vocoded signals in NH participants. We used informative and non-informative visual cues in non-verbal pitch perception and memory tasks (pitch change detection, pitch direction identification, and short-term memory) to test the potential usefulness of multisensory integration in these two populations. We hypothesized that CI users would have more difficulties than NH in general, but might benefit more strongly from the visual cues than do NH, even than NH in vocoded conditions, based on previous research on multisensory integration in CI users [64,66].
Overall, the present study presents a new set of five listening tests to assess important aspects of non-verbal auditory perception in NH participants and CI users. Its aim is to provide a rapid overview of the non-verbal auditory abilities of a given participant and, over the longer run, to provide a useful tool for further auditory research investigating various populations with hearing deficits.
2. Material and Methods
2.1. Participants
Ten cochlear implant users and ten normal-hearing participants were recruited (see Table 1 for demographical information). All of the control participants had normal hearing (hearing loss inferior to 20 dB at octave frequencies from 500 to 4000 Hz in both ears, measured with audiometry in silence, in accordance with the BIAP standards (Recommendation 02/1: Audiometric Classification of Hearing Impairments). There were six unilateral CI users with hearing-aids on the contralateral ear, and four bilateral CI users, all implanted after post-lingual deafness (see Table S1 for details about the CIs). We chose CI users with one year or more of implantation in order to have only CI users in the post-implantation chronic phase. Psychiatric and neuropsychological disorders were assessed by the Hospital where the CI users are medically taken care of. The participants were not diagnosed with psychiatric or neurological disorders, as assessed by a short questionnaire at inclusion in the study. The control participants were selected for their hearing-in-quiet capacities rather than their age, leading to age differences between the groups [20,69,90]. As age-difference could have played a role in the CI results, we performed Pearson correlations between each score obtained in the Results section and the age of the CI users, with Bonferroni correction for multiple testing. There was not any significant correlation between the age of the participants and their corresponding scores in all of the subtests (all pcorr > 0.4). We also ran a supplementary Bayesian Linear Regressions on the five tasks with age as a covariate, and no effect of age was significantly better than the null model for each task (all BF10 < 1.9). The study procedures were approved by an ethics committee (randomly selected at the national level, CPP Ile de France VI, ID RCD 2018-A02670-55), and all of the participants were paid for their participation and gave their written informed consent.
Table 1.
Group | CI (6 Unilateral and 4 Bilateral) |
Controls (N = 10) |
p-Value (Group Comparison) |
Effect Size (Lower and Upper Confidence Interval at 95%) |
---|---|---|---|---|
Sex | 8M 2F | 4M 6F | 0.07 | |
Age (years) | 51 (±14) Min: 24 Max: 73 |
22.1(±1.7) Min: 20 Max: 25 |
<0.001 | 2.9 (1.6–4.2) |
Education (years) | 16.1 (±2.8) Min: 10 Max: 20 |
15.5 (±1.2) Min: 14 Max: 17 |
0.5 | 0.28 (−0.6–1.2) |
Musical education (years) | 1.5 (±4.7) Min: 0 Max: 15 |
0.6 (±1.6) Min: 0 Max: 5 |
0.6 | 0.26 (−0.6–1.1) |
Laterality | 9R, 1L | 9R, 1L | 1 | |
Right Ear | 8 implants, 2 hearing-aids | NA | ||
Left Ear | 6 implants, 4 hearing-aids | NA | ||
Unilateral implant (n = 6):
Duration (years) |
2.33 (±1.5) Min: 1 Max: 5 |
NA | ||
Bilateral implants (n = 4):
First implant Duration (years) |
6.75 (±6.4) Min: 2 Max: 16 |
NA | ||
Bilateral implants (n = 4):
Second implant Duration (years) |
5 (±4.5) Min: 1 Max: 11 |
NA |
2.2. Listening Tests: Material and Procedure
Auditory non-verbal perception was assessed with five tests: pitch change detection (PCD), pitch direction change identification (DCI), pitch short-term memory (STM), auditory stream segregation (AS), and emotion recognition (EMO). All of the tests were implemented to run on an iPad touch tablet, allowing the participants to answer by touching large buttons presented on the screen (see Figure 1 for schematic representations of the five tests).
For the PCD, DCI and STM tasks, the same stimuli were used, with a roving of frequency across the trials. Note that we did not use intensity roving, either within or between the trials. We intended to measure how CI users could process frequency changes in stimuli as they can occur in everyday situations, and did not intend to assess which cues they used to do so: they might thus use some loudness cues [91,92] and not only pitch cues, in particular in the PCD and STM tasks. Furthermore, large intra-trial intensity roving creates a situation of interference which would be particularly deleterious in more cognitively demanding tasks such as the STM task. The stimuli for the PCD, DCI and STM tasks were synthetic harmonic tones (twelve harmonics), equalized in Root Mean Square (RMS) amplitude, each lasting 500 ms and presented with a within-sequence Inter-Stimulus-Interval (ISI) of 100 ms. The stimuli in the AS task were based on the same harmonic tones but with a duration of 100 ms. On half of the trials in the PCD, DCI and STM tasks, visual cues were presented (see Figure 1). They were composed of white disks with a diameter of 2 cm (100 px) on a black screen. For these three tests, auditory-only and audio-visual trials were presented in pseudo-random order, with no more than two repetitions of the same type of trial in a row. The visual stimuli were always informative regarding the timing of the audio stimulus (i.e., its onset). However, they provided task-relevant information about the tones’ pitch only for the STM and DCI tasks. In the DCI task, the visual stimuli indicated the correct answer for the task. The purpose of including audio-visual trials in the DCI task was not to gain any specific information, but rather for the sake of completeness, (1) to obtain visual information on half of the trials for the three pitch-based tasks, and (2) in anticipation of a training study, which is now ongoing, in which we are using informative, task-relevant visual information with the DCI task [74].
For the experiment with the NH participants, we also used vocoded sounds with 16, 8, and 4 channels (see below for details). The experiment took place in a quiet room. The participant was seated in front of the tablet with two loudspeakers (Logitech Z200) at a 70 cm distance from the participant’s head, with 40 cm between each speaker. We set the volume at 55 dB SPL for the NH participants and adjusted it to a comfortable listening level for the CI users (+/−5 dB SPL). The bilateral users performed the tests using both CIs. The unilateral users performed the tests using their CI and their hearing-aids in order to reproduce natural listening conditions in each CI user.
Each participant performed the five subtests in a random order, with each test presenting the stimuli in a pseudo-random order, with no more than two repetitions of the same type of stimulus in a row. For the NH participants, the order of the sound type (normal, vocoded four channels, vocoded eight channels, vocoded 16 channels) was also randomized, and all five tests for one sound type were presented in a row, in the same order for each sound type. Before each subtest, the participants received an oral explanation of the corresponding task, and the CI users also received written instructions to accompany the oral instructions. The participants performed a short training session in order to ensure that they understood the task, with the same range of frequencies as in the experimental trials. We checked the performance of this short training session for each participant before starting (at least 75% of correct responses). The entire session lasted about 30 min for CI users and two hours for NH participants. No response feedback was provided during the tests.
2.2.1. Pitch Change Detection (PCD) Test
In one trial, the participants were presented with a sequence of five isochronous harmonic tones, all identical (standard harmonic tone) except for the fourth harmonic tone, which could differ in frequency (adapted from [63,84]). The standard frequencies were 165, 196, 262 or 392 Hz. The deviant frequencies were between 131 and 494 Hz, with the changes relative to the standard harmonic tone being between 1/16, 1/8, ¼, ½, 1, or 2 tones, either up or down compared to the standard. In total, 64 sequences were constructed with five notes. There were 16 identical trials (four trials for each standard) and 48 different trials (twelve trials per each standard, that is, one trial per deviant size, up and down). Non-informative visual cues were presented on half of the trials in addition to the harmonic tones (as in Albouy et al. [63]). Five circles appeared sequentially from left to right, synchronously with the harmonic tones. They were always positioned at the center of the vertical axis on the screen (see Figure 1), and hence were not informative as far as the pitch of the harmonic tone was concerned, meaning that they did not give sufficient information to perform the task, but gave information regarding the onset of the sound. The participants had to determine whether the fourth harmonic tone of the sequence was the same as or different from the other tones. After the end of the sequence, the participants had unlimited time to give their answer by tapping on either the “Same” or the “Different” button. After they had given their answer, the next trial was played automatically after an average delay of 1000 ms (700–1300 ms).
2.2.2. Pitch Direction Change Identification (DCI) Test
The participants were presented with two harmonic tones at two different frequencies in each trial. The fundamental frequencies of the tones were between 123 and 523 Hz. The steps between the two harmonic tones could be of 0.5, 1, 1.5, 2, 2.5, 3, or 3.5 tones. In total, 56 two-tone sequences were constructed: 28 “up” sequences, with the second harmonic tone being higher in pitch than the first one, and 28 “down” sequences, with the second harmonic tone being lower in pitch than the first one. Informative visual cues were presented on half of the trials in addition to the harmonic tones. Two circles connected by a white bar appeared sequentially from left to right, and simultaneously with the onset of each harmonic tone. The vertical positions of the circles were centered on average on the two harmonic tones (to be at the center of the screen). Moreover, the vertical position of each circle was calculated according to the frequency of the corresponding harmonic tone: the higher the frequency, the higher the circle on the screen. In contrast to PCD, the visual cues were thus fully informative for pitch height (see Figure 1), but as they were only present in half of the trials, the participants were asked to base their judgements on their auditory perception. These visual cues aimed to reinforce the association between visual height and pitch [74], an effect we plan to exploit in future training experiments. The participants had to determine if the second harmonic tone was higher in pitch (Up) or lower (Down) than the first harmonic tone. After the end of the second harmonic tone, the participants had unlimited time to give their answer by tapping on either the “Up” or the “Down” button. After having given their answer, the next trial was played automatically after an average delay of 1000 ms (700–1300 ms).
2.2.3. Pitch Short-Term Memory (STM) Test
The participants were presented with two melodies of four harmonic tones (S1 and S2), with S2 being either identical or different from S1 [75,93]. The fundamental frequencies of the harmonic tones were between 262 and 440 Hz (corresponding to notes between C4 and A4). In total, 32 melodies were constructed with four harmonic tones; each melody thus lasted 2300 ms. In total, there were 16 identical and 16 different trials. For the different trials, changes of one harmonic tone could occur on the second or third harmonic tone. The changes could be of 1.5, 2, 2.5, 3.5, or 4.5 tones, all entailing a change of contour. The delay between the two melodies of a trial was 1000 ms. Informative visual cues were presented on half of the trials in addition to the harmonic tones; they were presented during S1, and in the delay between S1 and S2. This allowed us to draw conclusions about the potential benefit of visual information during the encoding and maintenance phase of a memory task. The circles were connected by white bars, appearing consecutively and simultaneously with the onset of each harmonic tone of the first sequence. The circles’ vertical positions were centered on average over the four harmonic tones, and each vertical circle position was calculated according to the frequency of the corresponding harmonic tone: the higher the frequency, the higher the circle would be on the screen (see Figure 1). The participants had to determine if the second melody was the same or different from the first melody. After the end of the second melody, the participants had unlimited time to give their answer by tapping on “Same” or “Different” button. After they gave their answer, the next trial was played automatically after an average delay of 1000 ms (700–1300 ms).
2.2.4. Auditory Stream Segregation (AS) Test
A sequence was constructed on the model ABA (with A being the standard harmonic tone and B being a harmonic tone with a varying frequency, both lasting 100 ms). The ISI between A and B was 20 ms, and the interval between two ABA triplets was 140 ms [47,50]. Five triplets were repeated for each frequency of B. The fundamental frequency of A was 196 Hz, and B was 196, 247, 294, 440, 659 or 988 Hz (i.e., ranging from 0 to 28 semitones with respect to 196 Hz). The sequence started with a fundamental frequency of B at 440, going down to 196 Hz, then up to 988 Hz, and down again to 196 Hz. This up-and-down pattern was repeated five times, terminating with B having a frequency of 294 Hz. In total, the AS sequence lasted approximately 2.5 min. During the sequence, the participants had to tell if they heard one stream (meaning they perceived the sequence as “integrated”) or two streams (meaning they perceived the sequence as “segregated”). They gave their answer by tapping on either the “1 stream” or the “2 streams” button. Once one button was selected, it remained selected until the participant changed their answer (the selected button remained highlighted). The participants could respond as many times as they wanted during the sequence.
2.2.5. Emotion Recognition (EMO) Test
Twenty sentences were taken from Pralus and Fornoni et al. [76]. These sentences were semantically neutral in French: “J’espère qu’il va m’appeler bientôt” (“I hope he will call me soon”) and “L’avion est presque plein” (“The plane is almost full”). These sentences were uttered with different emotions by male and female actors. For each emotion (joy, neutral, sadness, anger, fear), four sentences were used, half pronounced by a male voice and half by a female voice. The stimuli lasted on average 1470 ms (+/−278 ms) and were equalized in RMS amplitude. In each trial, the participants listened to a stimulus and were asked to select the recognized emotion from five options (joy, neutral, sadness, anger, fear). After having given their response, they were asked to rate the intensity of the selected emotion from 1 (not intense) to 5 (very intense), except for the stimuli judged as neutral (as in Pralus and Fornoni et al., [76]). They had unlimited time to give their answer. After the intensity rating response, the following stimulus was played automatically after an average delay of 1000 ms (700–1300 ms).
2.2.6. Vocoded Sounds
Three vocoded conditions simulating cochlear implants with different numbers of channels were created using MatLab R2016a (Mathworks, Inc, Natick, MA, USA). For the complete vocoding procedure, see Massida et al. [94] and Rouger et al. [64]. All of the sounds presented in the battery were analyzed through 4, 8, or 16 frequency bands using sixth-order IIR elliptical analysis filters. We extracted the temporal envelope by half-wave rectification for each of these frequency bands. The envelope was smoothed with a 500 Hz low-pass third order IIR elliptical filter. We used this extracted envelope to modulate a white noise given by a generator. The obtained signal was filtered with the same filters used previously for the frequency bands. We additively recombined the signals from each frequency band and adjusted the acoustic level obtained to match the original sound level based on RMS.
2.3. Data Analysis
A Bayesian approach was used to analyze the data [95,96]. We analyzed the data with Bayesian mixed repeated-measures analyses of variance (ANOVA), as implemented in the software JASP [97]. Bayesian analyses allow us to perform a model comparison and to select the best model (with the best evidence) given the data. In the first set of analyses, we investigated the effect of vocoding in NH participants’ data with Sound Type as a within-subject factor, with four levels (Non-vocoded, Vocoded with four, eight, or 16 channels). Other relevant factors depending on the task are detailed below. In the second set of analyses, we compared the groups (NH control data for non-vocoded sounds vs. CI), hence the between-participants factor Group was included for all of the tasks. We report the Bayes Factor (BF) as a relative measure of evidence. In order to interpret the strength of evidence (according to [95]), we considered a BF under three to be weak evidence, a BF between three and 10 to be positive evidence, a BF between 10 and 100 to be strong evidence, and a BF higher than 100 to be decisive evidence. BF10 indicates the evidence of H1 (a given model) compared to H0 (the null model), and BFinclusion indicates the evidence of one effect over all of the models. As no post-hoc tests with correction for multiple comparison have as yet been developed for Bayesian statistics [96,97], we used t-tests with Holm–Bonferroni correction for multiple comparisons.
For the PCD, DCI, and STM tests, we analyzed the percentage of correct responses with Modality (auditory or audiovisual) as a within-participant factor, and the factors Sound Type or Group (as described above). For PCD (different trials only) and DCI, we ran an additional analysis including the factor of Difficulty (different change sizes). For PCD, we also ran the analysis for identical trials only. For STM, we also analyzed the percentage of Hits (correct responses for different trials) minus the percentage of False Alarms (incorrect responses for the same trials) to correct for potential response bias.
For the AS test, we analyzed first the total time spent in the perception of one stream or two streams (thus excluding the time needed to give the first answer), and second, the mean frequency corresponding to changes in the number of streams perceived, with Sound Type or Group as a factor.
For the EMO test, we analyzed the percentages of correct responses and intensity ratings with Emotion (joy, sadness, anger, fear or neutral) as a within-participant factor, and the factors Sound Type or Group. Note that for the intensity ratings, the emotion factor had only four levels, as the neutral stimuli were not rated for intensity. We analyzed only the intensity ratings for trials with correctly recognized emotions (as in Pralus, Fornoni et al., 2019). The confusion matrices were calculated based on the percentage of responses given for each type of emotion compared to the expected emotion.
In addition, in order to better understand the relationships between the pitch tasks (PCD, DCI, and STM) with similar auditory stimuli, we performed a Bayesian ANOVA on accuracy for the auditory trials only, with Task as a within-participant factor, and with the between-participants factor Group (CI users, NH participants).
3. Results
3.1. PCD Test
3.1.1. Normal-Hearing Participants and Vocoded Sounds
After comparison to the null model, the best model showing positive evidence was the one with the main effect of Sound Type (BF10 = 4.09) (Figure 2A). The other models showed no noticeable evidence (BF10 < 1.2) (Table 2). This was confirmed by a decisive positive effect of Sound Type (BFinclusion = 3.15) only, while other specific effects showed no evidence (BFinclusion < 0.48). According to the t-tests with Holm–Bonferroni correction, the original sounds were significantly better recognized than the vocoded sounds with eight and four channels (all pcorr < 0.03).
Table 2.
Models | P(M) | P(M|Data) | BFM | BF10 | Error % | |
---|---|---|---|---|---|---|
PCD | Null model (incl. subject) | 0.2 | 0.136 | 0.629 | 1.000 | |
Sound Type | 0.2 | 0.555 | 4.986 | 4.085 | 0.865 | |
Sound Type + Modality | 0.2 | 0.164 | 0.782 | 1.204 | 3.341 | |
Sound Type + Modality + Sound Type × Modality | 0.2 | 0.107 | 0.479 | 0.787 | 3.020 | |
Modality | 0.2 | 0.039 | 0.161 | 0.285 | 0.956 | |
DCI | Null model (incl. subject) | 0.2 | 3.164 × 10−7 | 1.266 × 10−6 | 1.000 | |
Modality | 0.2 | 0.881 | 29.605 | 2.784 × 106 | 5.315 | |
Sound Type + Modality | 0.2 | 0.099 | 0.438 | 311,979.244 | 1.818 | |
Sound Type + Modality + Sound Type × Modality | 0.2 | 0.020 | 0.083 | 64,215.941 | 1.075 | |
Sound Type | 0.2 | 3.109 × 10−8 | 1.244 × 10−7 | 0.098 | 1.218 | |
STM | Null model (incl. subject) | 0.2 | 0.522 | 4.370 | 1.000 | |
Sound Type | 0.2 | 0.239 | 1.255 | 0.458 | 0.612 | |
Modality | 0.2 | 0.147 | 0.687 | 0.281 | 1.416 | |
Sound Type + Modality | 0.2 | 0.071 | 0.305 | 0.136 | 3.380 | |
Sound Type + Modality + Sound Type × Modality | 0.2 | 0.021 | 0.088 | 0.041 | 1.351 | |
AS-total time | Null model (incl. subject) | 0.2 | 2.668 × 10−49 | 1.067 × 10−48 | 1.000 | |
Sound Type + Percept + Sound Type × Percept | 0.2 | 0.875 | 28.028 | 3.280 × 1048 | 3.159 | |
Percept | 0.2 | 0.116 | 0.523 | 4.332 × 1047 | 1.046 | |
Sound Type + Percept | 0.2 | 0.009 | 0.038 | 3.484 × 1046 | 1.184 | |
Sound Type | 0.2 | 1.883 × 10−50 | 7.532 × 10−50 | 0.071 | 2.129 | |
AS-frequency | Null model (incl. subject) | 0.5 | 0.073 | 0.079 | 1.000 | |
Sound Type | 0.5 | 0.927 | 12.720 | 12.720 | 0.307 | |
EMO-recognition | Null model (incl. subject) | 0.2 | 1.560 × 10−20 | 6.240 × 10−20 | 1.000 | |
Sound Type + Emotion + Sound Type × Emotion | 0.2 | 0.886 | 31.106 | 5.680 × 1019 | 0.682 | |
Sound Type + Emotion | 0.2 | 0.114 | 0.514 | 7.301 × 1018 | 0.934 | |
Sound Type | 0.2 | 6.305 × 10−5 | 2.522 × 10−4 | 4.042 × 1015 | 0.600 | |
Emotion | 0.2 | 3.105 × 10−19 | 1.242 × 10−18 | 19.904 | 0.658 | |
EMO-intensity | Null model (incl. subject) | 0.2 | 3.605 × 10−6 | 1.442 × 10−5 | 1.000 | |
Sound Type + Emotion + Sound Type × Emotion | 0.2 | 0.999 | 4038.321 | 277,105.29 | 0.859 | |
Sound Type + Emotion | 0.2 | 8.423 × 10−4 | 0.003 | 233.632 | 0.864 | |
Sound Type | 0.2 | 1.304 × 10−4 | 5.218 × 10−4 | 36.182 | 0.789 | |
Emotion | 0.2 | 1.320 × 10−5 | 5.280 × 10−5 | 3.661 | 0.550 |
The best model for each task is in bold font. P(M): prior probability assigned to the model; P(M|data): probability of the model knowing the data; BFM: Bayesian Factor of the model; BF10: Bayesian Factor of the model compared to the null model.
In addition, we analyzed the percentage of correct responses for the different trials with the additional factor of Difficulty (six change sizes, see Methods) (Figure 3A). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Sound Type, Difficulty, and the interaction between the two (BF10 = 6.52 × 1049). It was 8.8 times better than the model with the main effects of Sound Type, Difficulty, Modality, and the interaction between Sound Type and Difficulty (BF10 = 7.37 × 1048), and 162 times better than the model with the main effects of Sound Type, Modality, Difficulty, the interaction between Sound Type and Modality, and the interaction between Sound Type and Difficulty (BF10 = 4.02 × 1047). All of the other models were at least 250 times less likely (BF10 < 2.6 × 1047). This was confirmed by a decisive specific effect of Difficulty (BFinclusion = 1.3 × 1014), Sound Type (BFinclusion = 1.3 × 1014), and the interaction between Sound Type and Difficulty (BFinclusion = 9.8 × 104). Other specific effects showed no significant evidence (BFinclusion < 0.04). According to the t-tests with Holm–Bonferroni correction, the original sounds were significantly better recognized than the vocoded sounds with 16, eight, and four channels (all pcorr < 0.001). Over all of the types of sounds (original or vocoded), trials of difficulty of 1/16 and 1/8 tones led to poorer performance than all of the other trials (all pcorr < 0.017), trials of difficulty of 1/4 tone were less well performed than trials of difficulties of one and two tones (both pcorr < 0.009), and trials of difficulties of ½ and one tone were less well performed than trials of difficulty of two tones (both pcorr < 0.019). For the original sounds, trials of difficulty of 1/16 were less well performed than that with difficulties of ¼ and two (both pcorr < 0.038). For 16-channels vocoded sounds, trials of difficulties of 1/16 and 1/8 were less well performed than trials of difficulties of ½, one and two (all pcorr < 0.042); trials of difficulty of 1/16 were less well performed than that with a difficulty of ¼ (pcorr < 0.001). For eight-channel vocoded sounds, trials of difficulties of 1/16, 1/8 and 1/4 were less well performed than those of difficulties of one and two (all pcorr < 0.01), and trials of difficulties of 1/16 and 1/8 were less well performed than that of difficulty of 1/2 (both pcorr < 0.042). For four-channel vocoded sounds, trials of difficulties of 1/16, 1/8 and 1/4 were less well performed than that of difficulties of two (all pcorr < 0.001), and trials of difficulty of 1/8 were less well performed than that of difficulties of ½ and one (both pcorr < 0.025). Thus, overall, when fewer channels were used for the vocoded sounds, their discrimination became harder even with large physical differences.
For the identical trials, we also analyzed the percentage of correct responses (Figure 3A). None of the tested models explained the data better than the null model (BF10 < 1). However, there was a small specific effect of the interaction of Modality and Sound Type (BFinclusion = 2.5), but no other significant specific effects (BFinclusion < 0.7). None of the post-hoc tests were significant, and the largest numerical difference between the audio-only and audio-visual trials was for vocoded sounds with four channels.
3.1.2. Cochlear Implant Listeners Compared to Normal-Hearing Participants
The null model and the model with Modality, Group, and the interaction between the two factors (BF10 = 0.9) explained almost equally well the data, with the null model being only 1.1 times more likely to explain the data than the model with all of the effects and interactions (Table 2, Figure 4A). This was confirmed by only a small specific effect of the interaction between Modality and Group (BFinclusion = 1.37), and no other significant specific effects (BFinclusion < 0.9). The data pattern was clearer with separated analyses for the same and different trials, as seen below.
In addition, we analyzed the percentage of correct responses for the different trials with the additional factor of Difficulty (six change sizes) (Figure 5A). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Group, Difficulty, and the interaction between the two (BF10 = 1.24 × 1016). It was 1.6 times better than the model with the main effect of Difficulty (BF10 = 7.68 × 1015), 3.4 times better than the model with the main effects of Group and Difficulty (BF10 = 3.61 × 1015), and 5.8 times better than the model with the main effects of Group, Difficulty, Modality and the interaction between Group and Difficulty (BF10 = 2.14 × 1015). All of the other models were at least 10 times less likely (BF10 < 1.2 × 1015). This was confirmed by a decisive specific effect of Difficulty (BFinclusion = ∞) and a weak specific effect of the interaction between Group and Difficulty (BFinclusion = 2.5). The other specific effects showed no significant evidence (BFinclusion < 0.8). According to t-tests with Holm–Bonferroni correction, over the two groups, trials of difficulty of 1/16 tone were less well categorized than all of other trials (all pcorr < 0.004), trials of difficulty of 1/8 tone were less well categorized than trials of difficulties of 1 and 2 tones (both pcorr < 0.012). In CI users, we found the same pattern of responses: trials of difficulties 1/16 and 1/8 were less well categorized than other difficulties (all pcorr < 0.025). In NH participants, only trials of difficulty of 1/16 tone were less well categorized than trials of difficulty of ¼, ½, 1 and 2 tones (all pcorr < 0.026).
We also analyzed the percentage of correct responses for identical trials (Figure 3A). The null model and the model with Modality, Group, and the interaction between the two factors (BF10 = 1.6) explained almost equally well the data, with the null model being only 1.6 times less likely to explain the data than the model with all of the effects and interactions (Table 2). This was confirmed by a small specific effect of the interaction between Modality and Group (BFinclusion = 2.3), and Group (BFinclusion = 1.5) and no other significant specific effects (BFinclusion < 0.8). According to the t-tests with Holm–Bonferroni correction, for audiovisual trials, CI users tended to have enhanced scores compared to NH listeners (pcorr = 0.07).
3.2. DCI Test
3.2.1. Normal-Hearing Participants and Vocoded Sounds
After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Modality (BF10 = 2.78 × 106) (Figure 2A). It was 8.9 times better than the model with the main effect of Modality and Sound Type (BF10 = 3.12 × 105), and 43 times better than the model with the two main effects of Sound Type and Modality, and their interaction (BF10 = 6.42 × 104). The model with the main effect of Sound Type only showed no significant evidence (BF10 = 0.1) (Table 2). This was confirmed by a decisive specific effect of Modality (BFinclusion = 1.92 × 106) only. The other specific effects showed no significant evidence (BFinclusion < 0.09).
In addition, we analyzed the percentage of correct responses with the additional factor of Difficulty (seven change sizes, see Methods) (Figure 3B). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Modality, Difficulty, and the interaction between the two (BF10 = 6.36 × 1023). It was 41.3 times better than the model with the main effects of Modality, Sound Type, Difficulty, and the interaction between Modality and Difficulty (BF10 = 1.54 × 1022), and 468 times better than the model with the main effects of Modality, Sound Type, Difficulty, and the interaction between Sound Type and Modality, and between Modality and Difficulty (BF10 = 1.36 × 1021). All of the other models were at least 1870 times less likely (BF10 < 3.4 × 1020). This was confirmed by the decisive specific effects of Modality (BFinclusion = 4.6 × 1014), Difficulty (BFinclusion = 3.8 × 104), and the interaction between Modality and Difficulty (BFinclusion = 4020). Other specific effects showed no significant evidence (BFinclusion < 0.01). According to the t-tests with Holm–Bonferroni correction, trials of difficulties of 0.5 and 1 tone were less well categorized than trials of difficulties of 2.5, 3 and 3.5 tones (all pcorr < 0.05). Audiovisual trials were specifically better categorized than auditory trials for difficulty levels of 0.5, 1, 1.5 and 3 tones (all pcorr < 0.001).
3.2.2. Cochlear Implant Listeners Compared to Normal-Hearing Participants
After the comparison to the null model, the best model showing decisive evidence was the one with the main effect of Modality (BF10 = 4178.2) (Figure 4A). It was 1.5 times better than the model with the main effects of Modality and Group (BF10 = 2708.5), and two times better than the model with the two main effects of Group and Modality, and their interaction (BF10 = 2113.5). The model with the main effect of Group only showed no significant evidence (BF10 = 0.5) (Table 3). This was confirmed by a decisive specific effect of Modality (BFinclusion = 3952.3) only. The other specific effects showed no significant evidence (BFinclusion < 1.2). As the three first models were very close, we report the post-hoc comparisons for the interaction. According to the t-tests with Holm–Bonferroni correction, both groups had better scores in the audiovisual trials compared to the auditory trials (pcorr < 0.001 for CI users and pcorr = 0.013 for NH). The CI users were significantly better in the audiovisual trials compared to the NH participants in the auditory trials (pcorr = 0.007), but not when comparing their performance in the auditory trials to the NH’s performance in the auditory trials. Note that on average across the audiovisual and auditory trials, the CI users thus performed better than the NH participants.
Table 3.
Models | P(M) | P(M|data) | BFM | BF10 | Error % | |
---|---|---|---|---|---|---|
PCD | Null model (incl. subject) | 0.2 | 0.284 | 1.586 | 1.000 | |
Modality + Group + Modality × Group | 0.2 | 0.255 | 1.368 | 0.897 | 1.732 | |
Modality | 0.2 | 0.204 | 1.026 | 0.719 | 1.035 | |
Group | 0.2 | 0.150 | 0.706 | 0.528 | 0.943 | |
Modality+ Group | 0.2 | 0.107 | 0.480 | 0.377 | 2.265 | |
DCI | Null model (incl. subject) | 0.2 | 1.111 × 10−4 | 4.444 × 10−4 | 1.000 | |
Modality | 0.2 | 0.464 | 3.465 | 4178.231 | 1.488 | |
Modality + Group | 0.2 | 0.301 | 1.721 | 2708.463 | 1.336 | |
Modality + Group + Modality × Group | 0.2 | 0.235 | 1.227 | 2113.544 | 1.824 | |
Group | 0.2 | 5.756 × 10−5 | 2.303 × 10−4 | 0.518 | 0.660 | |
STM | Null model (incl. subject) | 0.2 | 0.226 | 1.170 | 1.000 | |
Modality | 0.2 | 0.394 | 2.605 | 1.743 | 0.945 | |
Modality + Group | 0.2 | 0.195 | 0.967 | 0.860 | 1.415 | |
Group | 0.2 | 0.105 | 0.470 | 0.465 | 0.647 | |
Modality + Group + Modality × Group | 0.2 | 0.079 | 0.345 | 0.351 | 1.767 | |
AS-total time | Null model (incl. subject) | 0.2 | 1.367 × 10−17 | 5.469 × 10−17 | 1.000 | |
Perception + Group + Perception × Group | 0.2 | 0.878 | 28.803 | 6.422 × 1016 | 1.531 | |
Percept | 0.2 | 0.090 | 0.394 | 6.558 × 1015 | 0.953 | |
Perception + Group | 0.2 | 0.032 | 0.133 | 2.360 × 1015 | 1.629 | |
Group | 0.2 | 4.768 × 10−18 | 1.907 × 10−17 | 0.349 | 1.044 | |
AS-frequency | Null model | 0.5 | 0.163 | 0.195 | 1.000 | |
Group | 0.5 | 0.837 | 5.125 | 5.125 | 7.765 × 10−4 | |
EMO-recognition | Null model (incl. subject) | 0.2 | 2.095 × 10−4 | 8.380 × 10−4 | 1.000 | |
Emotion + Group + Emotion × Group | 0.2 | 0.860 | 24.653 | 4107.699 | 2.294 | |
Emotion + Group | 0.2 | 0.123 | 0.563 | 588.756 | 0.879 | |
Emotion | 0.2 | 0.014 | 0.058 | 67.967 | 0.347 | |
Group | 0.2 | 0.002 | 0.007 | 8.747 | 2.277 | |
EMO-intensity | Null model (incl. subject) | 0.2 | 0.295 | 1.673 | 1.000 | |
Emotion | 0.2 | 0.315 | 1.838 | 1.067 | 0.787 | |
Emotion + Group | 0.2 | 0.174 | 0.843 | 0.590 | 0.956 | |
Group | 0.2 | 0.161 | 0.769 | 0.547 | 0.868 | |
Emotion + Group + Emotion × Group | 0.2 | 0.055 | 0.233 | 0.187 | 1.238 |
In addition, we analyzed the percentage of correct responses with the factor of Difficulty (seven change sizes, see Methods) (Figure 5B). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Modality and Difficulty, and the interaction between the two (BF10 = 2.28 × 1013). It was 1.52 times better than the model with the main effect of Modality (BF10 = 1.5 × 1013); 1.8 times better than the model with the main effects of Modality, Group, Difficulty, the interaction between Group and Modality, and between Modality and Difficulty (BF10 = 1.26 × 1013); and 1.9 times better than the model with the main effects of Modality, Group, Difficulty, and the interaction between Modality and Difficulty (BF10 = 1.19 × 1013). All of the other models were at least 2.7 times less likely (BF10 < 8.4 × 1012). This was confirmed by a decisive specific effect of Modality (BFinclusion = 1.87 × 1013) and a small positive effect of the interaction between Modality and Difficulty (BFinclusion = 3). The other specific effects showed no significant evidence (BFinclusion < 0.77). According to the t-tests with Holm–Bonferroni correction, the trials of difficulty of 0.5 tone had a tendency to be less well categorized than the trials of difficulties of 2.5 tones (pcorr = 0.14). The audiovisual trials were specifically better categorized than the auditory trials for difficulty levels of 0.5, 1 and 1.5 (all pcorr < 0.013).
3.3. STM Test
3.3.1. Normal-Hearing Participants and Vocoded Sounds
None of the tested models explained the data (percentage of correct responses) better than the null model (BF10 < 0.5) (Table 2, Figure 2A). This was confirmed by there being no significant specific effects (BFinclusion < 0.4).
In addition, we analyzed the percentage of Hits minus the percentage of false alarms in NH participants with all sound types, in audio or audiovisual conditions (Figure 2B). None of the tested models explained the data better than the null model (BF10 < 0.6). This was confirmed by there being no significant specific effects (BFinclusion < 0.5).
3.3.2. Cochlear Implant Listeners Compared to Normal-Hearing Participants
The null model and the model with Modality (BF10 = 1.7) explained the data almost equally well, with the null model being only 1.7 times less likely to explain the data than the model with all of the effects and interactions (Table 3, Figure 4A). This was confirmed by a small specific effect of Modality (BFinclusion = 1.3), and no other significant specific effects (BFinclusion < 0.4).
For the analysis of the percentage of Hits minus the percentage of False Alarms (Figure 4B), after comparison to the null model, the best model showing positive evidence was the one with the main effect of Modality (BF10 = 3). The other models showed no significant evidence (BF10 < 1.5). This was confirmed by the small evidence of Modality only (BFinclusion = 2.2), and the other specific effects showed no evidence (BFinclusion < 0.5). The performance in the audiovisual trials was better than performance in the auditory trials in both groups.
3.4. AS Test
3.4.1. Normal-Hearing Participants and Vocoded Sounds
Total time spent in one or two stream percepts (Figure 2C). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Perception (one or two streams) and Sound Type, and the interaction between the two (BF10 = 3.28 × 1048). This model was 7.2 times more likely than the model with the main effect of Perception (BF10 = 4.3 × 1047), and 94 times more likely than the model with the two main effects of Perception and Sound Type (BF10 = 3.48 × 1046). The model with the main effect of Sound Type showed no evidence (BF10 = 0.07) (Table 2). This was confirmed by a decisive specific effect of Perception (BFinclusion = 9.8 × 1013), a strong specific effect of the interaction between Perception and Sound Type (BFinclusion = 28), and a positive specific effect of Sound Type (BFinclusion = 5.1). According to t-tests with Holm–Bonferroni correction, less time was spent in the perception of one stream compared to two streams for all types of sounds (all pcorr < 0.001). There was a tendency to spend less time in the perception of one stream with four-channel vocoded sounds compared to original sounds (pcorr = 0.069), and a tendency to spend more time in the perception of two streams with four-channel vocoded sounds compared to 16-channel vocoded sounds (pcorr = 0.069).
Mean frequency at the change of perception (Figure 2D). After comparison to the null model, the model with the main effects of Sound Type showed strong evidence (BF10 = 12.85) (Table 2). According to the t-tests with Holm–Bonferroni correction, there was a tendency for a higher mean frequency with eight-channel vocoded sounds compared to 16-channel vocoded sounds (pcorr = 0.1).
3.4.2. Cochlear Implant Listeners Compared to Normal-Hearing Participants
Total time spent in one or two streams percepts (Figure 4C). After comparison to the null model, the best model showing decisive evidence was the one with the main effects of Perception (one or two streams) and Group, and the interaction between the two (BF10 = 6.4 × 1016). This model was 9.7 times more likely than the model with the main effect of Perception (BF10 = 6.6 × 1015), and 27.8 times more likely than the model with the two main effects of Perception and Group (BF10 = 2.4 × 1015). The model with only the main effect of Group showed no evidence (BF10 = 0.3) (Table 3). This was confirmed by a decisive specific effect of Perception (BFinclusion = ∞), a strong specific effect on the interaction between Perception and Group (BFinclusion = 28.8), and a positive specific effect of Group (BFinclusion = 6.8). According to the t-tests with Holm–Bonferroni correction, both groups spent less time in the perception of one stream compared to two streams (both pcorr < 0.001). The CI users spent significantly less time in the perception of two streams compared to the NH participants (pcorr = 0.034), and more time in the perception of one stream compared to the NH participants (pcorr = 0.024).
Mean frequency at the change of perception (Figure 4D). After comparison to the null model, the model with the main effect of Group showed positive evidence (BF10 = 5.1) (Table 3). The frequency difference between the A and B sounds at which switches between perceptions occurred was higher for the CI users’ group (111.6 Hz) than the NH participants’ group (49.8 Hz).
3.5. EMO Test
3.5.1. Normal-Hearing Participants and Vocoded Sounds
Emotion categorization (Figure 2E). After comparison to the null model, the best model showing decisive evidence was the one with the two main effects of Sound Type and Emotion, and their interaction (BF10 = 5.68 × 1019). This model was 7.8 times more likely than the model with the two main effects of Sound Type and Emotion (BF10 = 7.3 × 1018), and 14,200 times more likely than the model with the main effect of Sound Type (BF10 = 4 × 1015). The model with the main effects of Emotion showed strong evidence (BF10 = 19.9) (Table 2). This was confirmed by a decisive specific effect of Emotion (BFinclusion = 1.06 × 104) and Sound Type (BFinclusion = 3.3 × 1014), and a strong specific effect of the interaction between Emotion and Sound Type (BFinclusion = 31.1). According to the t-tests with Holm–Bonferroni correction, Anger was significantly better recognized than Joy and Neutrality (pcorr = 0.019 and 0.008 respectively). The original sounds were significantly better recognized than the vocoded sounds with 16, eight and four channels (all pcorr < 0.003). The vocoded sounds with 16 channels were significantly better recognized than the ones with eight and four channels (both pcorr < 0.001), and vocoded sounds with eight channels were better recognized than the ones with four channels (pcorr = 0.016). The original and vocoded sounds with 16 channels were significantly better recognized than the vocoded sounds with four channels for Joy, Sadness and Neutrality (all pcorr < 0.003). The original sounds and 16-channel vocoded sounds were significantly better recognized than the eight-channels vocoded sounds for Joy (pcorr = 0.023). The original sounds were significantly better recognized than the eight-channel vocoded sounds for Neutrality (pcorr = 0.011).
The confusion matrices (Table S2) showed that with vocoded sounds, fear was often confused with anger, which was never the case with the original sounds; indeed, with the original sounds, fear was more confused with sadness. Moreover, with the vocoded sounds, sadness was often confused with neutrality, whereas this confusion was not present with the original sounds.
Intensity ratings (Figure 2F). After comparison to the null model, the best model showing decisive evidence was the one with the two main effects of Sound Type and Emotion, and their interaction (BF10 = 2.8 × 105). This model was 1202 times more likely than the model with the two main effects of Sound Type and Emotion (BF10 = 233.6), and 7777 times more likely than the model with the main effect of Sound Type (BF10 = 36.2). The model with the main effect of Emotion showed small positive evidence (BF10 = 3.7) (Table 2). This was confirmed by a decisive specific effect of Emotion (BFinclusion = 4973), Sound Type (BFinclusion = 3.4 × 104), and the interaction between Emotion and Sound Type (BFinclusion = 4038). According to the t-tests with Holm–Bonferroni correction, Anger was rated as being significantly more intense than Fear and Sadness (pcorr = 0.04 and 0.004 respectively). Joy was rated as being more intense than Sadness (pcorr = 0.005). Vocoded sounds with 16-channels tended to be rated higher than four-channel vocoded sounds (pcorr = 0.078). Vocoded sounds with 16 channels were rated significantly higher than vocoded sounds with four channels for Joy and Sadness (both pcorr < 0.001).
3.5.2. Cochlear Implant Listeners Compared to Normal-Hearing Participants
Emotion categorization (Figure 4E). After comparison to the null model, the best model showing decisive evidence was the one with the two main effects of Group and Emotion, and their interaction (BF10 = 4107.7). This model was seven times more likely than the model with the two main effects of Group and Emotion (BF10 = 588.8), and 60 times more likely than the model with the main effect of Emotion (BF10 = 68). The model with the main effect of Group showed positive evidence (BF10 = 8.7) (Table 3). This was confirmed by a decisive specific effect of Emotion (BFinclusion = 325.9), a strong specific effect of Group (BFinclusion = 45.5), and of the interaction between Emotion and Group (BFinclusion = 24.7). According to the t-tests with Holm–Bonferroni correction, Fear was significantly less recognized than Anger and Neutrality (all pcorr < 0.001). CI had lower recognition scores compared to NH for Joy (t(8) = 4.3 pcorr = 0.004) and for Sadness (t(8) = 3.5 pcorr = 0.038) (other pcorr > 0.8).
The confusion matrices (Table S3) showed that in CI users, joy was often confused with sadness, which was never the case for the control participants. Moreover, in the CI users, sadness was often confused with neutrality; this confusion was not observed for NH participants.
Intensity ratings (Figure 4F). After comparison to the null model, all of the models showed no significant evidence (BF10 < 1.1) (Table 3). This was confirmed by there being no significant specific effects (BFinclusion < 0.8).
Overall, the CI users showed prominent deficits in three out of the five listening tasks compared to the NH participants: Pitch Change Detection, Auditory Stream segregation and Emotional prosody recognition. In contrast, the CI users reached similar performance levels for the pitch Direction Change Identification task, as well as the pitch sequence Short-Term Memory task in comparison to the NH participants.
The NH participants with vocoded sounds showed deficits in three out of the five listening tasks: Pitch Change Detection, Auditory Stream segregation and Emotional prosody recognition (and their associated intensity ratings). These deficits partly correlated to the number of channels of the vocoder used.
In order to assess the potential differences between the unilateral and bilateral CI users, we compared the results of the two subgroups (unilateral vs bilateral CI users, see Table S4 for the individual results). No significant difference was found between the two groups for any of the tasks (all BF10 < 3 for models with the Group effect).
3.6. Relationships between the Tasks: Cochlear Implant Listeners Compared to Normal-Hearing Participants in Pitch Tasks (PCD, DCI and STM)
After comparison to the null model, the best model showing positive evidence was the one with the main effect of Task (BF10 = 7.25) (Figure S1). It was 1.3 times better than the model with the two main effects of Group and Task, and their interaction (BF10 = 5.45), and 2.1 times better than the model with the two main effects of Group and Task (BF10 = 3.42). The model with the main effect of Group was not significant (BF10 = 0.42). This was confirmed by a positive specific effect of Task (BFinclusion = 7.55), and a small specific effect of the interaction between Task and Group (BFinclusion = 1.8), and no other significant specific effect (BFinclusion < 0.8). According to the t-tests with Holm–Bonferroni correction, PCD showed greater recognition scores than DCI (pcorr = 0.018) and STM (pcorr = 0.018). In the control group, the PCD task showed greater recognition scores than the DCI (pcorr = 0.005). This was not the case in the CI users, reflecting the more homogenous results across the tasks in this group, with higher scores in DCI and lower scores in the PCD compared to the NH participants (see Figure S1).
4. Discussion
Overall, our study demonstrated that the rapid assessment of non-verbal auditory perception can be performed for NH participants and CI users. The CI users showed prominent deficits in three out of the five listening tasks compared to the NH participants: Pitch Change Detection, Auditory Stream segregation and Emotion (prosody) recognition. In contrast, they reached similar performance levels for the pitch Direction Change Identification task, as well as the pitch sequence Short-Term Memory task in comparison to the NH participants. This pattern of perceptual deficits of CI users was mostly mimicked in the NH participants with vocoded sounds, with a deficit partly correlated to the number of channels of the vocoder used. Both groups seemed to benefit from visual cues in the pitch tasks, but these effects were small and did not differ between the participant groups.
4.1. Patterns of Non-verbal Auditory Perception Deficits in CI Users and in NH Participants Hearing Vocoded Sounds
In the PCD test, which was the most basic task related to pitch perception in the battery, the CI users demonstrated a deficit compared to the NH participants for the trials with the smallest pitch differences. The CI users’ performance level was comparable to the performance level of the NH participants listening to vocoded sounds (Figure 2A and Figure 3A). These deficits were more pronounced for the difficult trials: the smaller the size change was, the bigger the deficits of the CI users and NH participants with vocoded sounds. These results were expected because CI and vocoded sounds only give a partial and degraded information about the pitch of the sound [7,8]. As we used roving pitches in the PCD task, we prevented frequency-related training and showed that the deficit was not specific to one frequency in particular.
In the DCI task, the participants did not need to recognize the pitch per se, but rather needed to distinguish a pitch difference and to infer a direction on this difference. The CI users showed no deficit on this task compared to the NH participants who listened either to the original sounds or the vocoded sounds. This could be linked to the fact that, compared to the PCD task, the participants were not asked to detect a pitch change, but rather to make a decision on a relationship between two tones, recognizing the contour. Previous work has demonstrated that CI users had remaining implicit pitch processing [32], which could have helped in this task. Moreover, in the DCI task, the pitch change sizes were bigger than in the PCD task. However, this task was not easier than the PCD. Indeed, for the NH participants, it was more difficult to correctly identify the direction of a pitch change (DCI) than to detect a pitch change (PCD).
As pitch is essential for music perception, and CI users often complain about their poor musical appreciation [23,24,28], we used a short music-like task to test their perception of melody information in the context of a short-term memory (STM) task. Surprisingly, the CI users did not show a deficit on this task, and performed similarly to NH participants with original sounds. Note that the NH participants did not show a deficit on this task with vocoded sounds either. Even though hearing loss can decrease cognitive abilities [98], various studies have demonstrated that compensating this hearing loss, in particular with CI, can improve cognitive performance [98,99,100,101]. The observed data pattern here suggests that the short-term memory abilities of CI users were sufficient to perform well at this memory task, even though the difference of melodies was based on a change on the pitch dimension, an acoustic dimension which is difficult to process for them. Moreover, as the to-be-detected difference included a change of contour, this feature could have helped CI users to detect this difference, similarly to the DCI task. Indeed, previous work on melodic contour identification (MCI task) revealed that CI users are able to use melodic contour information [102,103,104]. Their recognition scores are similar to NH participants when the number of semitones between notes is large enough [102], and they improve with the number of years of musical experience [105]. Overall, it appears that when the task is not purely pitch-perception centered, CI users can perform as well as NH participants. This could explain why, despite their poor pitch discrimination abilities, CI users are able to listen to and enjoy music [29,104]. It might be argued that the patients only succeeded because they were not impaired severely enough. However, the CI users tested here experienced difficulties with the PCD task and the two tasks AS and EMO, which are both more related to their everyday life perception.
In agreement with previous reports [42,51,52,53,54], our AS task detected the deficit of streaming segregation in CI users. Indeed, the CI users spent more time in the perception of one stream, meaning that their segregation was not performed as efficiently as that of the NH participants. This was confirmed by an increase of the frequency at the change of perception. Interestingly, these results were also found in NH participants with vocoded sounds, but only for the 4-channels vocoded sounds, i.e., the most degraded sounds. This reflects the deficit of CI users in everyday life in the segregation of two auditory sources [14,52,53], and more generally to hear signals in noise. However, both NH listeners and CI users spend considerable amounts of time in noisy situations, mostly trying to understand sounds in background noise [106]. In order to reflect this deficit, streaming segregation tasks have often been used, as they reflect a way of understanding hearing-in-noise apart from speech comprehension per se. Some studies suggest that these tasks only show a need for more time to make a decision and subjective uncertainty in CI users compared to NH participants [51]. However, in order to account for this, we decided to make several back and forth presentations of the same pitch differences between the A and B sounds of the ABA triplets [50], and to measure perception over the entire sequence. Here, we can assume that the subjective uncertainty would be compensated over the time of the sequence. Recent studies have also shown that this decreased streaming in CI users was not only due to an increased decision time, as it depends on electrode separation, but also on tone repetition time [107,108,109], similarly to NH participants [47]. Moreover, even if this kind of perceptual task is rather simple, the performance correlates with speech perception in noise [42,109]. Taken together, these findings suggest that a task using streaming segregation is a simple and rapid way to evaluate the hearing-in-noise abilities of CI users independently of their phonological skills.
As pitch is important for non-verbal auditory cues in speech, we also evaluated the perception of emotional prosody in our participants. Using a simple paradigm of the emotion categorization of short sentences, we demonstrated a deficit of emotion recognition in CI users compared to NH participants. This deficit was particularly pronounced for joyful and sad sentences, which CI users tended to confound more frequently than did NH participants. Interestingly, the same pattern of deficit was found in NH participants with vocoded sentences, with a deficit related to the number of vocoded channels, and also a specific deficit for neutrality. This demonstrates efficiently the poor perception of emotional prosody in CI users that has already been documented before [28,37,38,110], and allows for a better characterization of this deficit depending on the emotion. Moreover, here, the double paradigm using both emotion categorization and intensity ratings allowed us to show that the perception of emotional prosody is not fully disrupted in CI users. The intensity ratings of the CI users and NH participants did not differ. This result reflects the potential capacities of CI users to correctly perceive emotional prosody on an implicit level. Indeed, in order to rate the intensity of an emotion, no conscious representation of a given stimulus was required. The intensity ratings are more linked to lower activation levels of the representation of this stimulus. Thus, the results of the intensity ratings in CI users reflect a potential implicit preservation of pitch processing in the case of emotional prosody [32], as was already demonstrated in another deficit of pitch perception: congenital amusia [76,111,112,113]. Interestingly, the pattern of results was different for NH participants with vocoded sentences. Indeed, these participants showed decreased intensity ratings for joy and sadness with four-channel vocoded sentences. This could be explained by the fact that NH participants are not used to hearing degraded speech, and use preferentially explicit strategies to detect and judge intensity of an emotion. As vocoded sentences could seem a little dehumanized for NH participants, they might have assumed that the emotions were less sincere and intense, especially for joy, which is generally rather intense, and for sadness, which can be confounded with neutral. This is also in line with previous work that demonstrated the limitations of the vocoder strategy to simulate the CI in NH participants, revealing that vocoded sounds did not give the same results as CI regarding music appreciation [104].
In each of the presented tasks, the sounds were presented at equal RMS. Therefore, it cannot be excluded that the CI users might have used cues other than pitch, such as loudness or timbre cues, to perform the tasks. However, as each of these tasks was designed to reflect the capacities of CI users to process pitch-changing stimuli in ecological situations, we chose to only vary one cue (frequency) and not to use roving of intensity in these tasks, which also avoided interference and overly difficult tasks. Note that if CI users were able to use cues other than pitch, this would reduce the between-group differences if anything. For the PCD task, one could argue that loudness could provide sufficient information to perform the task for CI users. However, this remained the only pitch-related task which revealed a deficit for these participants, so loudness cues were not sufficient. For the DCI task, loudness cues were not expected to be correlated to pitch direction cues given the frequency roving across trials, hence these loudness cues were likely to be insufficient to perform the task. In conclusion, we can assume that pitch cues, which were manipulated in our tasks, were indeed used by the CI participants.
Both bimodal and bilateral CI users were recruited in the present study, as this assessment tool was designed to be performed in various populations with hearing deficits. All of the participants retained their regular device settings (CIs or CI plus conventional hearing aid), as we wanted to reproduce natural listening conditions for each participant, to have the reflection of difficulties participants could experience in their everyday life. We acknowledge that the residual hearing of bimodal users could have helped these patients to perform the tasks. However, we suggest that it was not sufficient to overcome all of their deficits, as we still observed reduced performances of CI users compared to NH participants. Moreover, the additional analyses of the performance of bimodal versus bilateral CI users revealed no differences between the two groups. In the future, it would be interesting to specifically test bimodal users with and without their hearing aids in order to ascertain the specific deficit relating to their CI, and to compare their performance with bilateral users, as it has already been shown that their perception can differ [19,31,33,114].
There was an age difference between the two groups of participants (the NH listeners were significantly younger than the CI users). This difference could have played a role in the difference of performance observed in the PCD task for small pitch changes, for the AS test, or for prosody recognition. However, we could argue that this difference of age was not the main factor, as no difference of group was observed for DCI or STM, and no significant correlation was observed between the age of the CI users and their performance. Furthermore, the deficits in the PCD task, the AS task, and the EMO task were elicited in the NH listeners by the vocoding of the sounds. Moreover, even if the cognitive abilities of participants were not impaired in general (there were no diagnosed psychiatric or neurological disorders), we cannot exclude that age-related central auditory processing deficits could have added some deficit in the CI users. Indeed, age-related hearing loss has been shown to have an effect on the auditory networks and the cognitive abilities of participants, such as working memory abilities, attentive functions, and speech perception [46]. In our present study, it is worth pointing out that for some of our tasks, the CI users performed as well as the NH participants, suggesting that the effect of age on auditory central cognitive abilities and the performances tested here might be rather minor (no effect of the age as a covariate in linear regression models was visible in the CI users’ performance). In order to fully dissociate the effect of CIs and the potential effect of age-related cognitive disorders, it would be interesting to test older normal-hearing participants with the tests of the present battery.
Overall, the present listening tests allow for the rapid (30 min) and easy (using a touchscreen tablet) characterization of several facets of non-verbal auditory perception in CI users and NH participants. The results were not found to depend on the time of use of the CI, as no correlation was found between the time of implantation and the performances of CI users (Pearsons’ correlation, all r < 0.442, all p > 0.2). We suggest, however, that the type of implant, sound processor, or the type of implantation (longer electrodes in particular) could influence the results of participants, as this affects pitch perception [115], which would need to be tested in larger-scale studies. The battery includes basic pitch detection and categorization tasks, but also tasks which are closer to the listening experience of everyday life, such as emotional prosody perception and the challenge of sound source segregation. In the population of CI users tested presently, DCI and STM tasks showed similar pattern of results and did not reveal any deficit. However, as this assessment tool was designed to be used for several pathological populations (and not only CI users with no cognitive deficits), we still recommend performing the entire testing session with the five tasks in order to fully uncover the potential deficits of the tested participants.
4.2. Benefit of Audiovisual Cues for Non-Verbal Auditory Tasks
In the present study, we also investigated a potential enhancement of pitch perception by additional visual stimulation. In the PCD task, we used non-task–informative visual cues which provided only temporal cues about the tones (and in particular, their onset and duration). Indeed, these visual cues did not provide the correct answer to perform the task, but rather gave information about the onset of the sounds. As the visual cues were not informative regarding the pitch of the tones, the visual information with its adequate timing might boost participants’ dynamic attendance to the onsets of the tones, as the tones were presented in a regular, isochronous sequence in the PCD task [116,117,118]. We found only weak evidence, if any, for an improvement of performance in audio-visual trials compared to audio-only trials in CI users. However, the trend was in the expected direction, notably based on a previous study in control and amusic participants, which incorporated a larger number of trials [63].
In the STM task, the visual cues were informative for the pitch dimension, but they were present only during the first melody and the retention delay. Hence, they did not inform them about the correct answer (which was the case in the DCI task). For the STM task, we observed a general improvement of the scores with audiovisual trials compared to the auditory trials in both populations. This result demonstrates that informative visual cues can help participants to enhance auditory perception and encoding, leading to better memory recognition performance, even though these cues were not present during the presentation of S2. In the DCI task, the visual cues were fully informative: they were sufficient to carry out the task without listening to the audio. The results revealed an improvement with audiovisual trials compared to auditory trials in both groups, and for NH participants with vocoded sounds. The visual benefit was even more pronounced in the CI users, who even outperformed the NH participants in the auditory trials. Even if the DCI task could be carried out using only visual cues, they were only presented in half of the trials. The purpose of integrating audio-visual trials in the DCI task was not to provide any specific information, but rather for the sake of completeness, in order to obtain visual information on half of the trials for the three pitch-based tasks. The CI users seemed to better integrate multisensory information and benefit more strongly from this integration than did the NH participants [64,66,69]. Indeed, the scores of the NH participants with audiovisual trials were not at the ceiling, suggesting that these participants might still be relying on auditory perception to carry out the task (as they were requested to do) and made less use of the visual information.
In conclusion, the results from these three tasks showed that visual cues could boost the performance of the participants in pitch perception tasks, both for CI users and NH participants. This suggests that even non-informative visual information (as in the PCD task) or partially informative information (as in the STM task) could help the participants to process pitch information. However, it would require more trials and longer sessions to assess these effects more completely, as they were only small effects in the current data set.
Over the past few years, many training and rehabilitation strategies have been developed to enhance the pitch perception abilities in CI users. For instance, some training focused on auditory musical training [34,39,105,119,120,121,122,123,124,125]. Most of these trainings used the Melodic Contour Identification task [102,105], in which a subject hears a melody, and has to recognize its contour and select the correct response among several contour visual representations. While most of these studies looked at the effect of the training of speech perception [119,120,122], some have demonstrated some improvement of intentional [34] or emotional prosody recognition in CI users [39]. The other studies proposed training strategies based on multi-sensorial integration, which was previously demonstrated to be enhanced in CI users [64]. For example, audio-motor integration [126] using electrotactile stimulation showed some minor improvement in speech perception [127]. Based on the previous results in CI users [64,65,66,128] and on the present results of our audiovisual tasks, it seems that audiovisual training could be a good strategy to enhance the auditory perception in CI users [129]. The present results are in favor of a training strategy based on informative cues, as demonstrated in the DCI and STM tasks. These informational visual cues could benefit pitch perception, especially when the difficulty is increased. Nevertheless, a training strategy based on non-informative cues could be enough to allow for multisensory integration and to enhance pitch perception. In the present study, CI users might even have benefited from a small training effect with visual cues [74], as their results in the DCI or STM tasks were similar to those of the NH participants. Indeed, these visual cues could have helped them to understand the meaning of the task (Up or Down for DCI), and could even give them a hearing strategy to perform the STM task, notably based on contour information [130]. In conclusion, these perception data provide some interesting insights for further training strategies in CI users, as well as for other populations with pitch perception deficits (e.g., individuals with congenital amusia).
5. Conclusions
Overall, the findings of the present study suggest that the five listening tests of PCD, DCI, STM, AS, and EMO can be used to characterize non-verbal auditory perception in participants with hearing difficulties (here CI users), as well as in NH participants with vocoded sounds. This assessment allows for the rapid (30 min) detection of pitch difficulties with auditory non-verbal sounds, as well as emotional prosody processing and stream segregation capacities, to be used easily in research studies. In order to obtain the full normal value range, it would be interesting to run this battery on a large sample of NH participants with various age and socioeconomic backgrounds in future studies. Moreover, the three pitch tasks using visual cues allowed us to better characterize multisensory integration in NH participants and CI users. These results could be a good starting point to devise a new audiovisual training procedure for participants with pitch perception difficulties [122], notably regarding the types of audiovisual cues to use (informative and non-informative). The five listening tests can then be used to monitor the enhancement of pitch perception abilities in participants (i.e., pre/post-training assessment; Pralus et al., in progress).
Acknowledgments
We would like to kindly thank Anjali Bhatara for her help with the prosody stimuli, Lesly Fornoni for her help to recruit the NH participants, Julie Gatel for her help to recruit the CI users, and Jackson Graves for the English spell-checking.
Supplementary Materials
The following are available online at https://www.mdpi.com/article/10.3390/jcm10102093/s1, Figure S1: Results of the NH participants and CI users in the PCD, DCI, and STM tasks (auditory trials only). Table S1: Information on cochlear implant(s) for bilateral and unilateral CI users. Table S2: Percent of answer types for each intended emotion for EMO task, averaged over the 10 NH participants for each Sound Types (Original, vocoded 16, 8, or 4 channels). Table S3: Percent of answer types for each intended emotion for EMO task, averaged over the 10 participants for each group (NH participants and CI users). Table S4: Individual results of CI users.
Author Contributions
Conceptualization, A.P., R.H., A.M., N.G., E.T., A.C. and B.T.; Methodology, A.P., R.H., A.C. and B.T.; Software, P.-E.A., P.B.; Validation, A.P., A.C. and B.T.; Formal Analysis, A.P., F.C., A.C. and B.T.; Investigation, A.P., A.C. and B.T.; Resources, A.P., R.H., E.T., B.T. and A.C.; Data Curation, A.P. and F.C.; Writing―Original Draft Preparation, A.P., A.C. and B.T.; Writing―Review and Editing, R.H., F.C., P.-E.A., A.M., P.B., N.G. and E.T.; Supervision, E.T., A.C. and B.T.; Project Administration, E.T., A.C. and B.T.; Funding Acquisition, A.C. and B.T. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the ‘‘Fédération pour la Recherche sur le Cerveau (FRC)”. This work was conducted in the framework of the LabEx CeLyA (‘‘Centre Lyonnais d’Acoustique”, ANR-10-LABX-0060) and of the LabEx Cortex (‘‘Construction, Function and Cognitive Function and Rehabilitation of the Cortex”, ANR-11-LABX-0042) of the Université de Lyon, within the program ‘‘Investissements d’avenir” (ANR-16-IDEX-0005) operated by the French National Research Agency (ANR).
Institutional Review Board Statement
The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Ethics Committee of CPP Ile de France VI, ID RCD 2018-A02670-55.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The conditions of our ethics approval do not permit public archiving of anonymized study data. Readers seeking access to the data should contact A. Pralus. Access will be granted to named individuals in accordance with ethical procedures governing the reuse of clinical data, including completion of a formal data sharing agreement and approval of the local ethics committee.
Conflicts of Interest
The authors declare no conflict of interest.
Highlights
Five listening tests were used to assess non-verbal auditory perception.
CI users showed deficits in pitch discrimination, emotional prosody, and streaming.
Similar deficits were observed in NH listeners with vocoded sounds.
Visual cues can enhance CI users’ performance in pitch perception tasks.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Lolli S.L., Lewenstein A.D., Basurto J., Winnik S., Loui P. Sound frequency affects speech emotion perception: Results from congenital amusia. Front. Psychol. 2015;6:1340. doi: 10.3389/fpsyg.2015.01340. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Bregman A.S., McAdams S. Auditory Scene Analysis: The Perceptual Organization of Sound. J. Acoust. Soc. Am. 1994;95:1177–1178. doi: 10.1121/1.408434. [DOI] [Google Scholar]
- 3.McAdams S. Segregation of concurrent sounds. I: Effects of frequency modulation coherence. J. Acoust. Soc. Am. 1989;86:2148–2159. doi: 10.1121/1.398475. [DOI] [PubMed] [Google Scholar]
- 4.Pattisapu P., Lindquist N.R., Appelbaum E.N., Silva R.C., Vrabec J.T., Sweeney A.D. A Systematic Review of Cochlear Implant Outcomes in Prelingually-deafened, Late-implanted Patients. Otol. Neurotol. 2020;41:444–451. doi: 10.1097/MAO.0000000000002555. [DOI] [PubMed] [Google Scholar]
- 5.Sharma S.D., Cushing S.L., Papsin B.C., Gordon K.A. Hearing and speech benefits of cochlear implantation in children: A review of the literature. Int. J. Pediatr. Otorhinolaryngol. 2020;133:109984. doi: 10.1016/j.ijporl.2020.109984. [DOI] [PubMed] [Google Scholar]
- 6.Torppa R., Huotilainen M. Why and how music can be used to rehabilitate and develop speech and language skills in hearing-impaired children. Hear. Res. 2019;380:108–122. doi: 10.1016/j.heares.2019.06.003. [DOI] [PubMed] [Google Scholar]
- 7.Glennon E., Svirsky M.A., Froemke R.C. Auditory cortical plasticity in cochlear implant users. Curr. Opin. Neurobiol. 2020;60:108–114. doi: 10.1016/j.conb.2019.11.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Lehmann A., Paquette S. Cross-domain processing of musical and vocal emotions in cochlear implant users. Audit. Cogn. Neurosci. 2015;9:343. doi: 10.3389/fnins.2015.00343. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.McDermott H.J. Music Perception with Cochlear Implants: A Review. Trends Amplif. 2004;8:49–82. doi: 10.1177/108471380400800203. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Limb C.J., Roy A.T. Technological, biological, and acoustical constraints to music perception in cochlear implant users. Hear. Res. 2014;308:13–26. doi: 10.1016/j.heares.2013.04.009. [DOI] [PubMed] [Google Scholar]
- 11.Roy A.T., Jiradejvong P., Carver C., Limb C.J. Assessment of Sound Quality Perception in Cochlear Implant Users During Music Listening. Otol. Neurotol. 2012;33:319–327. doi: 10.1097/MAO.0b013e31824296a9. [DOI] [PubMed] [Google Scholar]
- 12.Roy A.T., Jiradejvong P., Carver C., Limb C.J. Musical Sound Quality Impairments in Cochlear Implant (CI) Users as a Function of Limited High-Frequency Perception. Trends Amplif. 2012;16:191–200. doi: 10.1177/1084713812465493. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Loizou P.C. Speech processing in vocoder-centric cochlear implants. Adv. Otorhinolaryngol. 2006;64:109–143. doi: 10.1159/000094648. [DOI] [PubMed] [Google Scholar]
- 14.Oxenham A.J. Pitch Perception and Auditory Stream Segregation: Implications for Hearing Loss and Cochlear Implants. Trends Amplif. 2008;12:316–331. doi: 10.1177/1084713808325881. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Marozeau J., Innes-Brown H., Blamey P.J. The acoustic and perceptual cues affecting melody segregation for listeners with a cochlear implant. Front. Psychol. 2013;4:790. doi: 10.3389/fpsyg.2013.00790. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Hopyan T., Manno F.A.M., Papsin B.C., Gordon K.A. Sad and happy emotion discrimination in music by children with cochlear implants. Child Neuropsychol. J. Norm. Abnorm. Dev. Child. Adolesc. 2016;22:366–380. doi: 10.1080/09297049.2014.992400. [DOI] [PubMed] [Google Scholar]
- 17.Petersen B., Andersen A.S.F., Haumann N.T., Højlund A., Dietz M.J., Michel F., Riis S.K., Brattico E., Vuust P. The CI MuMuFe—A New MMN Paradigm for Measuring Music Discrimination in Electric Hearing. Front. Neurosci. 2020;14:2. doi: 10.3389/fnins.2020.00002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Sharp A., Delcenserie A., Champoux F. Auditory Event-Related Potentials Associated with Music Perception in Cochlear Implant Users. Front. Neurosci. 2018;12 doi: 10.3389/fnins.2018.00538. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Spitzer E.R., Galvin J.J., Friedmann D.R., Landsberger D.M. Melodic interval perception with acoustic and electric hearing in bimodal and single-sided deaf cochlear implant listeners. Hear. Res. 2021;400:108136. doi: 10.1016/j.heares.2020.108136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Spangmose S., Hjortkjær J., Marozeau J. Perception of Musical Tension in Cochlear Implant Listeners. Front. Neurosci. 2019;13 doi: 10.3389/fnins.2019.00987. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Phillips-Silver J., Toiviainen P., Gosselin N., Turgeon C., Lepore F., Peretz I. Cochlear implant users move in time to the beat of drum music. Hear. Res. 2015;321:25–34. doi: 10.1016/j.heares.2014.12.007. [DOI] [PubMed] [Google Scholar]
- 22.Jiam N.T., Limb C.J. Rhythm processing in cochlear implant−mediated music perception. Ann. N. Y. Acad. Sci. 2019;1453:22–28. doi: 10.1111/nyas.14130. [DOI] [PubMed] [Google Scholar]
- 23.Riley P.E., Ruhl D.S., Camacho M., Tolisano A.M. Music Appreciation after Cochlear Implantation in Adult Patients: A Systematic Review. Otolaryngol. Neck Surg. 2018;158:1002–1010. doi: 10.1177/0194599818760559. [DOI] [PubMed] [Google Scholar]
- 24.Zhou Q., Gu X., Liu B. The music quality feeling and music perception of adult cochlear implant recipients. J. Clin. Otorhinolaryngol. Head Neck Surg. 2019;33:47–51. doi: 10.13201/j.issn.1001-1781.2019.01.010. [DOI] [PubMed] [Google Scholar]
- 25.Ambert-Dahan E., Giraud A., Sterkers O., Samson S. Judgment of musical emotions after cochlear implantation in adults with progressive deafness. Front. Psychol. 2015;6:181. doi: 10.3389/fpsyg.2015.00181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Shirvani S., Jafari Z., Sheibanizadeh A., Motasaddi Zarandy M., Jalaie S. Emotional perception of music in children with unilateral cochlear implants. Iran. J. Otorhinolaryngol. 2014;26:225–233. [PMC free article] [PubMed] [Google Scholar]
- 27.Shirvani S., Jafari Z., Motasaddi Zarandi M., Jalaie S., Mohagheghi H., Tale M.R. Emotional Perception of Music in Children With Bimodal Fitting and Unilateral Cochlear Implant. Ann. Otol. Rhinol. Laryngol. 2016;125:470–477. doi: 10.1177/0003489415619943. [DOI] [PubMed] [Google Scholar]
- 28.Paquette S., Ahmed G.D., Goffi-Gomez M.V., Hoshino A.C.H., Peretz I., Lehmann A. Musical and vocal emotion perception for cochlear implants users. Hear. Res. 2018 doi: 10.1016/j.heares.2018.08.009. [DOI] [PubMed] [Google Scholar]
- 29.Fuller C., Başkent D., Free R. Early Deafened, Late Implanted Cochlear Implant Users Appreciate Music More than and Identify Music as Well as Postlingual Users. Front. Neurosci. 2019;13 doi: 10.3389/fnins.2019.01050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Migirov L., Kronenberg J., Henkin Y. Self-reported listening habits and enjoyment of music among adult cochlear implant recipients. Ann. Otol. Rhinol. Laryngol. 2009;118:350–355. doi: 10.1177/000348940911800506. [DOI] [PubMed] [Google Scholar]
- 31.Giannantonio S., Polonenko M.J., Papsin B.C., Paludetti G., Gordon K.A. Experience Changes How Emotion in Music Is Judged: Evidence from Children Listening with Bilateral Cochlear Implants, Bimodal Devices, and Normal Hearing. PLoS ONE. 2015;10:e0136685. doi: 10.1371/journal.pone.0136685. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Tillmann B., Poulin-Charronnat B., Gaudrain E., Akhoun I., Delbé C., Truy E., Collet L. Implicit Processing of Pitch in Postlingually Deafened Cochlear Implant Users. Front. Psychol. 2019;10 doi: 10.3389/fpsyg.2019.01990. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Buechner A., Krueger B., Klawitter S., Zimmermann D., Fredelake S., Holube I. The perception of the stereo effect in bilateral and bimodal cochlear implant users and its contribution to music enjoyment. PLoS ONE. 2020;15:e0235435. doi: 10.1371/journal.pone.0235435. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Lo C.Y., McMahon C.M., Looi V., Thompson W.F. Melodic Contour Training and Its Effect on Speech in Noise, Consonant Discrimination, and Prosody Perception for Cochlear Implant Recipients. Behav. Neurol. 2015;2015:352869. doi: 10.1155/2015/352869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Peng S.-C., Tomblin J.B., Turner C.W. Production and perception of speech intonation in pediatric cochlear implant recipients and individuals with normal hearing. Ear Hear. 2008;29:336–351. doi: 10.1097/AUD.0b013e318168d94d. [DOI] [PubMed] [Google Scholar]
- 36.Torppa R., Faulkner A., Huotilainen M., Jaervikivi J., Lipsanen J., Laasonen M., Vainio M. The perception of prosody and associated auditory cues in early-implanted children: The role of auditory working memory and musical activities. Int. J. Audiol. 2014;53:182–191. doi: 10.3109/14992027.2013.872302. [DOI] [PubMed] [Google Scholar]
- 37.Pak C.L., Katz W.F. Recognition of emotional prosody by Mandarin-speaking adults with cochlear implants. J. Acoust. Soc. Am. 2019;146:EL165–EL171. doi: 10.1121/1.5122192. [DOI] [PubMed] [Google Scholar]
- 38.Deroche M.L.D., Felezeu M., Paquette S., Zeitouni A., Lehmann A. Neurophysiological Differences in Emotional Processing by Cochlear Implant Users, Extending Beyond the Realm of Speech. Ear Hear. 2019 doi: 10.1097/AUD.0000000000000701. [DOI] [PubMed] [Google Scholar]
- 39.Good A., Gordon K.A., Papsin B.C., Nespoli G., Hopyan T., Peretz I., Russo F.A. Benefits of Music Training for Perception of Emotional Speech Prosody in Deaf Children With Cochlear Implants. Ear Hear. 2017;38:455–464. doi: 10.1097/AUD.0000000000000402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Bugannim Y., Roth D.A.-E., Zechoval D., Kishon-Rabin L. Training of Speech Perception in Noise in Pre-Lingual Hearing Impaired Adults with Cochlear Implants Compared with Normal Hearing Adults. Otol. Neurotol. 2019;40:e316–e325. doi: 10.1097/MAO.0000000000002128. [DOI] [PubMed] [Google Scholar]
- 41.Choi J.E., Moon I.J., Kim E.Y., Park H.-S., Kim B.K., Chung W.-H., Cho Y.-S., Brown C.J., Hong S.H. Sound Localization and Speech Perception in Noise of Pediatric Cochlear Implant Recipients: Bimodal Fitting Versus Bilateral Cochlear Implants. Ear Hear. 2017;38:426–440. doi: 10.1097/AUD.0000000000000401. [DOI] [PubMed] [Google Scholar]
- 42.Hong R.S., Turner C.W. Pure-tone auditory stream segregation and speech perception in noise in cochlear implant recipients. J. Acoust. Soc. Am. 2006;120:360–374. doi: 10.1121/1.2204450. [DOI] [PubMed] [Google Scholar]
- 43.Döge J., Baumann U., Weissgerber T., Rader T. Single-sided Deafness: Impact of Cochlear Implantation on Speech Perception in Complex Noise and on Auditory Localization Accuracy. Otol. Neurotol. 2017;38 doi: 10.1097/MAO.0000000000001520. [DOI] [PubMed] [Google Scholar]
- 44.Kaandorp M.W., Smits C., Merkus P., Festen J.M., Goverts S.T. Lexical-Access Ability and Cognitive Predictors of Speech Recognition in Noise in Adult Cochlear Implant Users. Trends Hear. 2017;21:2331216517743887. doi: 10.1177/2331216517743887. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Coffey E.B.J., Arseneau-Bruneau I., Zhang X., Zatorre R.J. The Music-In-Noise Task (MINT): A Tool for Dissecting Complex Auditory Perception. Front. Neurosci. 2019;13 doi: 10.3389/fnins.2019.00199. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Sardone R., Battista P., Panza F., Lozupone M., Griseta C., Castellana F., Capozzo R., Ruccia M., Resta E., Seripa D., et al. The Age-Related Central Auditory Processing Disorder: Silent Impairment of the Cognitive Ear. Front. Neurosci. 2019;13:619. doi: 10.3389/fnins.2019.00619. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Van Noorden L.P.A.S. Ph.D. Thesis. Eindhoven University of Technology; Eindhoven, The Netherlands: 1975. Temporal Coherence in the Perception of Tone Sequences. [Google Scholar]
- 48.Nie Y., Nelson P.B. Auditory stream segregation using amplitude modulated bandpass noise. Front. Psychol. 2015;6 doi: 10.3389/fpsyg.2015.01151. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Gaudrain E., Grimault N., Healy E.W., Béra J.-C. Streaming of vowel sequences based on fundamental frequency in a cochlear-implant simulation. J. Acoust. Soc. Am. 2008;124:3076–3087. doi: 10.1121/1.2988289. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Grimault N., Bacon S.P., Micheyl C. Auditory stream segregation on the basis of amplitude-modulation rate. J. Acoust. Soc. Am. 2002;111:1340–1348. doi: 10.1121/1.1452740. [DOI] [PubMed] [Google Scholar]
- 51.Böckmann-Barthel M., Deike S., Brechmann A., Ziese M., Verhey J.L. Time course of auditory streaming: Do CI users differ from normal-hearing listeners? Front. Psychol. 2014;5 doi: 10.3389/fpsyg.2014.00775. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Cooper H.R., Roberts B. Auditory stream segregation of tone sequences in cochlear implant listeners. Hear. Res. 2007;225:11–24. doi: 10.1016/j.heares.2006.11.010. [DOI] [PubMed] [Google Scholar]
- 53.Cooper H.R., Roberts B. Auditory stream segregation in cochlear implant listeners: Measures based on temporal discrimination and interleaved melody recognition. J. Acoust. Soc. Am. 2009;126:1975. doi: 10.1121/1.3203210. [DOI] [PubMed] [Google Scholar]
- 54.Hong R.S., Turner C.W. Sequential stream segregation using temporal periodicity cues in cochlear implant recipients. J. Acoust. Soc. Am. 2009;126:291–299. doi: 10.1121/1.3140592. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Rouger J., Lagleyre S., Démonet J., Fraysse B., Deguine O., Barone P. Evolution of crossmodal reorganization of the voice area in cochlear-implanted deaf patients. Hum. Brain Mapp. 2012;33:1929–1940. doi: 10.1002/hbm.21331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Strelnikov K., Marx M., Lagleyre S., Fraysse B., Deguine O., Barone P. PET-imaging of brain plasticity after cochlear implantation. Hear. Res. 2015;322:180–187. doi: 10.1016/j.heares.2014.10.001. [DOI] [PubMed] [Google Scholar]
- 57.Strelnikov K., Rouger J., Demonet J., Lagleyre S., Fraysse B., Deguine O., Barone P. Visual activity predicts auditory recovery from deafness after adult cochlear implantation. Brain J. Neurol. 2013;136:3682–3695. doi: 10.1093/brain/awt274. [DOI] [PubMed] [Google Scholar]
- 58.McGurk H., MacDonald J. Hearing lips and seeing voices. Nature. 1976;264:746–748. doi: 10.1038/264746a0. [DOI] [PubMed] [Google Scholar]
- 59.Frassinetti F., Bolognini N., Bottari D., Bonora A., Làdavas E. Audiovisual integration in patients with visual deficit. J. Cogn. Neurosci. 2005;17:1442–1452. doi: 10.1162/0898929054985446. [DOI] [PubMed] [Google Scholar]
- 60.Grasso P.A., Làdavas E., Bertini C. Compensatory Recovery after Multisensory Stimulation in Hemianopic Patients: Behavioral and Neurophysiological Components. Front. Syst. Neurosci. 2016;10:45. doi: 10.3389/fnsys.2016.00045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Passamonti C., Bertini C., Làdavas E. Audio-visual stimulation improves oculomotor patterns in patients with hemianopia. Neuropsychologia. 2009;47:546–555. doi: 10.1016/j.neuropsychologia.2008.10.008. [DOI] [PubMed] [Google Scholar]
- 62.Caclin A., Bouchet P., Djoulah F., Pirat E., Pernier J., Giard M.-H. Auditory enhancement of visual perception at threshold depends on visual abilities. Brain Res. 2011;1396:35–44. doi: 10.1016/j.brainres.2011.04.016. [DOI] [PubMed] [Google Scholar]
- 63.Albouy P., Lévêque Y., Hyde K.L., Bouchet P., Tillmann B., Caclin A. Boosting pitch encoding with audiovisual interactions in congenital amusia. Neuropsychologia. 2015;67:111–120. doi: 10.1016/j.neuropsychologia.2014.12.006. [DOI] [PubMed] [Google Scholar]
- 64.Rouger J., Lagleyre S., Fraysse B., Deneve S., Deguine O., Barone P. Evidence that cochlear-implanted deaf patients are better multisensory integrators. Proc. Natl. Acad. Sci. USA. 2007;104:7295–7300. doi: 10.1073/pnas.0609419104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Rouger J., Fraysse B., Deguine O., Barone P. McGurk effects in cochlear-implanted deaf subjects. Brain Res. 2008;1188:87–99. doi: 10.1016/j.brainres.2007.10.049. [DOI] [PubMed] [Google Scholar]
- 66.Strelnikov K., Rouger J., Lagleyre S., Fraysse B., Démonet J.-F., Déguine O., Barone P. Increased audiovisual integration in cochlear-implanted deaf patients: Independent components analysis of longitudinal positron emission tomography data. Eur. J. Neurosci. 2015;41:677–685. doi: 10.1111/ejn.12827. [DOI] [PubMed] [Google Scholar]
- 67.Strelnikov K., Rouger J., Barone P., Deguine O. Role of speechreading in audiovisual interactions during the recovery of speech comprehension in deaf adults with cochlear implants. Scand. J. Psychol. 2009;50:437–444. doi: 10.1111/j.1467-9450.2009.00741.x. [DOI] [PubMed] [Google Scholar]
- 68.Waddington E., Jaekel B.N., Tinnemore A.R., Gordon-Salant S., Goupell M.J. Recognition of Accented Speech by Cochlear-Implant Listeners: Benefit of Audiovisual Cues. Ear Hear. 2020 doi: 10.1097/AUD.0000000000000842. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Barone P., Chambaudie L., Strelnikov K., Fraysse B., Marx M., Belin P., Deguine O. Crossmodal interactions during non-linguistic auditory processing in cochlear-implanted deaf patients. Cortex J. Devoted Study Nerv. Syst. Behav. 2016;83:259–270. doi: 10.1016/j.cortex.2016.08.005. [DOI] [PubMed] [Google Scholar]
- 70.Owens E., Kessler D.K., Raggio M.W., Schubert E.D. Analysis and revision of the minimal auditory capabilities (MAC) battery. Ear Hear. 1985;6:280–290. doi: 10.1097/00003446-198511000-00002. [DOI] [PubMed] [Google Scholar]
- 71.Nilsson M., Soli S.D., Sullivan J.A. Development of the Hearing In Noise Test for the measurement of speech reception thresholds in quiet and in noise. J. Acoust. Soc. Am. 1994;95:1085–1099. doi: 10.1121/1.408469. [DOI] [PubMed] [Google Scholar]
- 72.Nilsson M.J., McCaw V.M., Soli S. Minimum Speech Test Battery for Adult Cochlear Implant Users. House Ear Institute; Los Angeles, CA, USA: 1996. [Google Scholar]
- 73.Shafiro V., Hebb M., Walker C., Oh J., Hsiao Y., Brown K., Sheft S., Li Y., Vasil K., Moberly A.C. Development of the Basic Auditory Skills Evaluation Battery for Online Testing of Cochlear Implant Listeners. Am. J. Audiol. 2020;29:577–590. doi: 10.1044/2020_AJA-19-00083. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Pralus A., Lévêque Y., Blain S., Bouchet P., Tillmann B., Caclin A. Training pitch perception in cochlear implant users via audiovisual integration: A pilot study in normal-hearing controls. Implant. Congr. Montr. 2019 [Google Scholar]
- 75.Tillmann B., Schulze K., Foxton J.M. Congenital amusia: A short-term memory deficit for non-verbal, but not verbal sounds. Brain Cogn. 2009;71:259–264. doi: 10.1016/j.bandc.2009.08.003. [DOI] [PubMed] [Google Scholar]
- 76.Pralus A., Fornoni L., Bouet R., Gomot M., Bhatara A., Tillmann B., Caclin A. Emotional prosody in congenital amusia: Impaired and spared processes. Neuropsychologia. 2019:107234. doi: 10.1016/j.neuropsychologia.2019.107234. [DOI] [PubMed] [Google Scholar]
- 77.Tillmann B., Albouy P., Caclin A. Congenital amusias. Handb. Clin. Neurol. 2015;129:589–605. doi: 10.1016/B978-0-444-62630-1.00033-0. [DOI] [PubMed] [Google Scholar]
- 78.Albouy P., Mattout J., Bouet R., Maby E., Sanchez G., Aguera P.-E., Daligault S., Delpuech C., Bertrand O., Caclin A., et al. Impaired pitch perception and memory in congenital amusia: The deficit starts in the auditory cortex. Brain J. Neurol. 2013;136:1639–1661. doi: 10.1093/brain/awt082. [DOI] [PubMed] [Google Scholar]
- 79.Albouy P., Cousineau M., Caclin A., Tillmann B., Peretz I. Impaired encoding of rapid pitch information underlies perception and memory deficits in congenital amusia. Sci. Rep. 2016;6:18861. doi: 10.1038/srep18861. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Graves J.E., Pralus A., Fornoni L., Oxenham A.J., Caclin A., Tillmann B. Short- and long-term memory for pitch and non-pitch contours: Insights from congenital amusia. Brain Cogn. 2019;136:103614. doi: 10.1016/j.bandc.2019.103614. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Tillmann B., Lévêque Y., Fornoni L., Albouy P., Caclin A. Impaired short-term memory for pitch in congenital amusia. Brain Res. 2016;1640:251–263. doi: 10.1016/j.brainres.2015.10.035. [DOI] [PubMed] [Google Scholar]
- 82.Albouy P., Caclin A., Norman-Haignere S.V., Lévêque Y., Peretz I., Tillmann B., Zatorre R.J. Decoding Task-Related Functional Brain Imaging Data to Identify Developmental Disorders: The Case of Congenital Amusia. Front. Neurosci. 2019;13 doi: 10.3389/fnins.2019.01165. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Albouy P., Mattout J., Sanchez G., Tillmann B., Caclin A. Altered retrieval of melodic information in congenital amusia: Insights from dynamic causal modeling of MEG data. Front. Hum. Neurosci. 2015;9:20. doi: 10.3389/fnhum.2015.00020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Hyde K.L., Peretz I. Brains That Are out of Tune but in Time. Psychol. Sci. 2004;15:356–360. doi: 10.1111/j.0956-7976.2004.00683.x. [DOI] [PubMed] [Google Scholar]
- 85.Loui P., Guenther F.H., Mathys C., Schlaug G. Action-Perception Mismatch in Tone-Deafness. Curr. Biol. 2008;18:R331–R332. doi: 10.1016/j.cub.2008.02.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Williamson V.J., Stewart L. Memory for pitch in congenital amusia: Beyond a fine-grained pitch discrimination problem. Memory. 2010;18:657–669. doi: 10.1080/09658211.2010.501339. [DOI] [PubMed] [Google Scholar]
- 87.Lima C.F., Brancatisano O., Fancourt A., Müllensiefen D., Scott S.K., Warren J.D., Stewart L. Impaired socio-emotional processing in a developmental music disorder. Sci. Rep. 2016;6 doi: 10.1038/srep34911. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Patel A.D., Wong M., Foxton J., Lochy A., Peretz I. Speech Intonation Perception Deficits in Musical Tone Deafness (congenital Amusia) Music Percept. Interdiscip. J. 2008;25:357–368. doi: 10.1525/mp.2008.25.4.357. [DOI] [Google Scholar]
- 89.Thompson W.F., Marin M.M., Stewart L. Reduced sensitivity to emotional prosody in congenital amusia rekindles the musical protolanguage hypothesis. Proc. Natl. Acad. Sci. USA. 2012;109:19027–19032. doi: 10.1073/pnas.1210344109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Butera I.M., Stevenson R.A., Mangus B.D., Woynaroski T.G., Gifford R.H., Wallace M.T. Audiovisual Temporal Processing in Postlingually Deafened Adults with Cochlear Implants. Sci. Rep. 2018;8:11345. doi: 10.1038/s41598-018-29598-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Arnoldner C., Kaider A., Hamzavi J. The role of intensity upon pitch perception in cochlear implant recipients. Laryngoscope. 2006;116:1760–1765. doi: 10.1097/01.mlg.0000228214.02606.42. [DOI] [PubMed] [Google Scholar]
- 92.Arnoldner C., Riss D., Kaider A., Mair A., Wagenblast J., Baumgartner W.-D., Gstöttner W., Hamzavi J.-S. The intensity-pitch relation revisited: Monopolar versus bipolar cochlear stimulation. Laryngoscope. 2008;118:1630–1636. doi: 10.1097/MLG.0b013e3181799715. [DOI] [PubMed] [Google Scholar]
- 93.Hirel C., Nighoghossian N., Lévêque Y., Hannoun S., Fornoni L., Daligault S., Bouchet P., Jung J., Tillmann B., Caclin A. Verbal and musical short-term memory: Variety of auditory disorders after stroke. Brain Cogn. 2017;113:10–22. doi: 10.1016/j.bandc.2017.01.003. [DOI] [PubMed] [Google Scholar]
- 94.Massida Z., Belin P., James C., Rouger J., Fraysse B., Barone P., Deguine O. Voice discrimination in cochlear-implanted deaf subjects. Hear. Res. 2011;275:120–129. doi: 10.1016/j.heares.2010.12.010. [DOI] [PubMed] [Google Scholar]
- 95.Lee M.D., Wagenmakers E.-J. Bayesian Cognitive Modeling: A Practical Course; Cambridge Core—Psychology Research Methods and Statistics. Cambridge University Press; Cambridge, UK: 2014. [Google Scholar]
- 96.Wagenmakers E.-J., Marsman M., Jamil T., Ly A., Verhagen J., Love J., Selker R., Gronau Q.F., Šmíra M., Epskamp S., et al. Bayesian inference for psychology. Part I: Theoretical advantages and practical ramifications. Psychon. Bull. Rev. 2018;25:35–57. doi: 10.3758/s13423-017-1343-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Wagenmakers E.-J., Love J., Marsman M., Jamil T., Ly A., Verhagen J., Selker R., Gronau Q.F., Dropmann D., Boutin B., et al. Bayesian inference for psychology. Part II: Example applications with JASP. Psychon. Bull. Rev. 2017:1–19. doi: 10.3758/s13423-017-1323-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Amieva H., Ouvrard C. Does Treating Hearing Loss in Older Adults Improve Cognitive Outcomes? A Review. J. Clin. Med. 2020;9:805. doi: 10.3390/jcm9030805. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Claes A.J., Van de Heyning P., Gilles A., Rompaey V.V., Mertens G. Cognitive outcomes after cochlear implantation in older adults: A systematic review. Cochlear Implants Int. 2018;19:239–254. doi: 10.1080/14670100.2018.1484328. [DOI] [PubMed] [Google Scholar]
- 100.Mosnier I., Bebear J.-P., Marx M., Fraysse B., Truy E., Lina-Granade G., Mondain M., Sterkers-Artières F., Bordure P., Robier A., et al. Improvement of cognitive function after cochlear implantation in elderly patients. JAMA Otolaryngol. Head Neck Surg. 2015;141:442–450. doi: 10.1001/jamaoto.2015.129. [DOI] [PubMed] [Google Scholar]
- 101.Völter C., Götze L., Dazert S., Falkenstein M., Thomas J.P. Can cochlear implantation improve neurocognition in the aging population? Clin. Interv. Aging. 2018;13:701–712. doi: 10.2147/CIA.S160517. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Galvin J.J., 3rd, Fu Q.J., Nogaki G. Melodic contour identification by cochlear implant listeners., Melodic Contour Identification by Cochlear Implant Listeners. Ear Hear. 2007;28:302–319. doi: 10.1097/01.aud.0000261689.35445.20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Luo X., Masterson M.E., Wu C.-C. Contour identification with pitch and loudness cues using cochlear implants. J. Acoust. Soc. Am. 2014;135:EL8–EL14. doi: 10.1121/1.4832915. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Wright R., Uchanski R.M. Music Perception and Appraisal: Cochlear Implant Users and Simulated CI Listening. J. Am. Acad. Audiol. 2012;23:350–379. doi: 10.3766/jaaa.23.5.6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.Galvin J.J., 3rd, Fu Q.J., Shannon R.V. Melodic contour identification and music perception by cochlear implant users., Melodic Contour Identification and Music Perception by Cochlear Implant Users. Ann. N. Y. Acad. Sci. 2009;1169:518–533. doi: 10.1111/j.1749-6632.2009.04551.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Busch T., Vanpoucke F., van Wieringen A. Auditory Environment Across the Life Span of Cochlear Implant Users: Insights From Data Logging. J. Speech Lang. Hear. Res. JSLHR. 2017;60:1362–1377. doi: 10.1044/2016_JSLHR-H-16-0162. [DOI] [PubMed] [Google Scholar]
- 107.Paredes-Gallardo A., Madsen S.M.K., Dau T., Marozeau J. The Role of Temporal Cues in Voluntary Stream Segregation for Cochlear Implant Users. Trends Hear. 2018;22:2331216518773226. doi: 10.1177/2331216518773226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Paredes-Gallardo A., Madsen S.M.K., Dau T., Marozeau J. The Role of Place Cues in Voluntary Stream Segregation for Cochlear Implant Users. Trends Hear. 2018;22:2331216517750262. doi: 10.1177/2331216517750262. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Saki N., Nikakhlagh S., Mirmomeni G., Bayat A. Perceptual organization of sequential stimuli in cochlear implant listeners: A temporal processing approach. Int. Tinnitus J. 2019;23:37–41. doi: 10.5935/0946-5448.20190007. [DOI] [PubMed] [Google Scholar]
- 110.Everhardt M.K., Sarampalis A., Coler M., Baskent D., Lowie W. Meta-Analysis on the Identification of Linguistic and Emotional Prosody in Cochlear Implant Users and Vocoder Simulations. Ear Hear. 2020 doi: 10.1097/AUD.0000000000000863. [DOI] [PubMed] [Google Scholar]
- 111.Frühholz S., Ceravolo L., Grandjean D. Specific Brain Networks during Explicit and Implicit Decoding of Emotional Prosody. Cereb. Cortex. 2012;22:1107–1117. doi: 10.1093/cercor/bhr184. [DOI] [PubMed] [Google Scholar]
- 112.Stewart L. Characterizing congenital amusia. Q. J. Exp. Psychol. 2011;64:625–638. doi: 10.1080/17470218.2011.552730. [DOI] [PubMed] [Google Scholar]
- 113.Tillmann B., Lalitte P., Albouy P., Caclin A., Bigand E. Discrimination of tonal and atonal music in congenital amusia: The advantage of implicit tasks. Neuropsychologia. 2016;85:10–18. doi: 10.1016/j.neuropsychologia.2016.02.027. [DOI] [PubMed] [Google Scholar]
- 114.Blamey P.J., Maat B., Başkent D., Mawman D., Burke E., Dillier N., Beynon A., Kleine-Punte A., Govaerts P.J., Skarzynski P.H., et al. A Retrospective Multicenter Study Comparing Speech Perception Outcomes for Bilateral Implantation and Bimodal Rehabilitation. Ear Hear. 2015;36:408–416. doi: 10.1097/AUD.0000000000000150. [DOI] [PubMed] [Google Scholar]
- 115.Canfarotta M.W., Dillon M.T., Buss E., Pillsbury H.C., Brown K.D., O’Connell B.P. Frequency-to-Place Mismatch: Characterizing Variability and the Influence on Speech Perception Outcomes in Cochlear Implant Recipients. Ear Hear. 2020 doi: 10.1097/AUD.0000000000000864. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116.Fiveash A., Schön D., Canette L.-H., Morillon B., Bedoin N., Tillmann B. A stimulus-brain coupling analysis of regular and irregular rhythms in adults with dyslexia and controls. Brain Cogn. 2020;140:105531. doi: 10.1016/j.bandc.2020.105531. [DOI] [PubMed] [Google Scholar]
- 117.Jones M.R. Time, our lost dimension: Toward a new theory of perception, attention, and memory. Psychol. Rev. 1976;83:323–355. doi: 10.1037/0033-295X.83.5.323. [DOI] [PubMed] [Google Scholar]
- 118.Kuroda T., Ono F., Kadota H. Dynamic Attending Binds Time and Rhythm Perception. Brain Nerve Shinkei Kenkyu No Shinpo. 2017;69:1195–1202. doi: 10.11477/mf.1416200895. [DOI] [PubMed] [Google Scholar]
- 119.Barlow N., Purdy S.C., Sharma M., Giles E., Narne V. The Effect of Short-Term Auditory Training on Speech in Noise Perception and Cortical Auditory Evoked Potentials in Adults with Cochlear Implants. Semin. Hear. 2016;37:84–98. doi: 10.1055/s-0035-1570335. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 120.Cheng X., Liu Y., Shu Y., Tao D.-D., Wang B., Yuan Y., Galvin J.J., Fu Q.-J., Chen B. Music Training Can Improve Music and Speech Perception in Pediatric Mandarin-Speaking Cochlear Implant Users. Trends Hear. 2018;22:2331216518759214. doi: 10.1177/2331216518759214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121.Firestone G.M., McGuire K., Liang C., Zhang N., Blankenship C.M., Xiang J., Zhang F. A Preliminary Study of the Effects of Attentive Music Listening on Cochlear Implant Users’ Speech Perception, Quality of Life, and Behavioral and Objective Measures of Frequency Change Detection. Front. Hum. Neurosci. 2020;14 doi: 10.3389/fnhum.2020.00110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Fuller C.D., Galvin J.J., Maat B., Başkent D., Free R.H. Comparison of Two Music Training Approaches on Music and Speech Perception in Cochlear Implant Users. Trends Hear. 2018;22:2331216518765379. doi: 10.1177/2331216518765379. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123.Lerousseau J.P., Hidalgo C., Schön D. Musical Training for Auditory Rehabilitation in Hearing Loss. J. Clin. Med. 2020;9:1058. doi: 10.3390/jcm9041058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124.Patel A.D. Can nonlinguistic musical training change the way the brain processes speech? The expanded OPERA hypothesis. Hear. Res. 2014;308:98–108. doi: 10.1016/j.heares.2013.08.011. [DOI] [PubMed] [Google Scholar]
- 125.Shukor N.F.A., Lee J., Seo Y.J., Han W. Efficacy of Music Training in Hearing Aid and Cochlear Implant Users: A Systematic Review and Meta-Analysis. Clin. Exp. Otorhinolaryngol. 2020 doi: 10.21053/ceo.2020.00101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.Chari D.A., Barrett K.C., Patel A.D., Colgrove T.R., Jiradejvong P., Jacobs L.Y., Limb C.J. Impact of Auditory-Motor Musical Training on Melodic Pattern Recognition in Cochlear Implant Users. Otol. Neurotol. 2020;41:e422. doi: 10.1097/MAO.0000000000002525. [DOI] [PubMed] [Google Scholar]
- 127.Huang J., Sheffield B., Lin P., Zeng F.-G. Electro-Tactile Stimulation Enhances Cochlear Implant Speech Recognition in Noise. Sci. Rep. 2017;7 doi: 10.1038/s41598-017-02429-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 128.Innes-Brown H., Marozeau J., Blamey P. The Effect of Visual Cues on Difficulty Ratings for Segregation of Musical Streams in Listeners with Impaired Hearing. PLoS ONE. 2011;6 doi: 10.1371/journal.pone.0029327. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 129.Sato T., Yabushita T., Sakamoto S., Katori Y., Kawase T. In-home auditory training using audiovisual stimuli on a tablet computer: Feasibility and preliminary results. Auris. Nasus. Larynx. 2019 doi: 10.1016/j.anl.2019.09.006. [DOI] [PubMed] [Google Scholar]
- 130.Talamini F., Vigl J., Doerr E., Grassi M., Carretti B. Auditory and Visual Mental Imagery in Musicians and Nonmusicians. OSF Prepr. 2020 doi: 10.31219/osf.io/vphr8. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The conditions of our ethics approval do not permit public archiving of anonymized study data. Readers seeking access to the data should contact A. Pralus. Access will be granted to named individuals in accordance with ethical procedures governing the reuse of clinical data, including completion of a formal data sharing agreement and approval of the local ethics committee.