Abstract
Previous positron emission tomography (PET) studies have shown that various cortical areas are activated to process speech signal in cochlear implant (CI) users. Nonetheless, differences in task dimension among studies and low statistical power preclude from understanding sound processing mechanism in CI users. Hence, we performed activation likelihood estimation meta‐analysis of PET studies in CI users and normal hearing (NH) controls to compare the two groups. Eight studies (58 CI subjects/92 peak coordinates; 45 NH subjects/40 peak coordinates) were included and analyzed, retrieving areas significantly activated by lexical and nonlexical stimuli. For lexical and nonlexical stimuli, both groups showed activations in the components of the dual‐stream model such as bilateral superior temporal gyrus/sulcus, middle temporal gyrus, left posterior inferior frontal gyrus, and left insula. However, CI users displayed additional unique activation patterns by lexical and nonlexical stimuli. That is, for the lexical stimuli, significant activations were observed in areas comprising salience network (SN), also known as the intrinsic alertness network, such as the left dorsal anterior cingulate cortex (dACC), left insula, and right supplementary motor area in the CI user group. Also, for the nonlexical stimuli, CI users activated areas comprising SN such as the right insula and left dACC. Previous episodic observations on lexical stimuli processing using the dual auditory stream in CI users were reconfirmed in this study. However, this study also suggests that dual‐stream auditory processing in CI users may need supports from the SN. In other words, CI users need to pay extra attention to cope with degraded auditory signal provided by the implant. Hum Brain Mapp 36:1982–1994, 2015. © 2015 Wiley Periodicals, Inc.
Keywords: cochlear implant, positron emission tomography, brain, meta‐analysis
INTRODUCTION
Cochlear implantation (CI), surgical insertion of electrodes into the cochlea, is so far the only available medical technique that can successfully replace a deficient sensory modality in humans. CI has been established as a standard treatment option for subjects with profound hearing loss, and thus more than 200,000 recipients including 80,000 children have been rehabilitated with CI [Kral and O'Donoghue, 2010; Zeitler et al., 2011]. From neuroscientists' viewpoint, CI also provides a unique opportunity to study cortical change associated with profound sensorineural hearing loss and restoration of the auditory modality [Giraud et al., 2001; Lazard et al., 2014; Song et al., 2013b].
However, even for a proficient CI user, poorly represented temporal fine structure and limited spectral cues delivered by the implant, yield impoverished input to the auditory cortex when compared with normal acoustic stimulation [Kral and O'Donoghue, 2010]. Thus, CI recipients show various performances. This large range of variability in post‐CI outcomes has long been investigated. Preoperative hypometabolism in the auditory cortex [Lee et al., 2001, 2007a] and postoperative hypermetabolism in the visual cortex [Strelnikov et al., 2013] have been suggested to be positively correlated with speech performance.
In this regard, researchers have been investigating brain activation patterns in deaf subjects with CI using functional imaging techniques, mostly positron emission tomography (PET). PET is optimal functional neuroimaging tool for CI users [Song et al., 2014b] as compared with other methodologies such as functional magnetic resonance imaging, magneto‐ or electroencephalography [Song et al., 2013a, 2014a, 2015]. In the early era of PET studies on CI users, most studies adopted regions of interest (ROIs) such as primary/secondary auditory cortices (A1/A2), Broca's/Wernicke's areas for statistical analysis and found relative increase of regional cerebral blood flow (rCBF) of these areas for sound stimuli as in normal hearing (NH) controls [Naito et al., 1995; Okazawa et al., 1996]. However, progress in imaging analysis techniques and accumulation of knowledge that various nonauditory brain regions other than aforementioned ROIs are activated during sound perception and interpretation in NH people [Hickok and Poeppel, 2007; Hickok et al., 2011; Rauschecker, 2011], studies on CI subjects have implemented data‐driven whole brain approach. Thus, various nonauditory multisensory brain areas have been revealed to be activated by both lexical and nonlexical stimuli [Giraud and Truy, 2002; Giraud et al., 2001].
Hitherto individual PET studies in CI subjects vary in tasks (i.e., phoneme, word, sentence, time‐reversed sentence, foreign language, environmental sound, and white noise) and are limited by statistical power and sensitivity due to small number of enrolled subjects. As an option to overcome these limitations, function‐location meta‐analysis permits to retrieve the most consistent activation areas as well as to compare results across different task dimensions in a standardized fashion [Laird et al., 2005].
Hence, we performed a meta‐analysis of PET studies on CI subjects by use of a coordinate‐based technique (activation likelihood estimation; ALE) [Song et al., 2012; Turkeltaub et al., 2002] to get a better understanding of the characteristic features of cortical activation patterns for both lexical and nonlexical auditory stimuli processing in CI relative to NH subjects.
MATERIALS AND METHODS
Selection Criteria and Included Studies
To identify all studies available, multiple PubMed searches on PET studies on CI according to Preferred Reporting Items for Systematic reviews and Meta‐Analyses (PRISMA) guidelines [Moher et al., 2009] were performed. Keywords used in the search were: “positron emission tomography” and “cochlear implant” with activated limits: article types other than review; human species; English language.
The inclusion criteria were that studies
Published in a peer‐reviewed journal,
Used a task involving passive hearing of auditory‐only stimulus,
Enrolled adult CI subjects (>18 years) who had bilateral postlingual profound hearing loss
Enrolled CI subjects who were implanted unilaterally,
Enrolled CI subjects who showed good performance enough to regain speech performance in everyday life (scores for word discrimination >60% or for sentence comprehension >70%) to reduce performance‐related confound. Only two of the included studies included “poor performers,”
Based on a data‐driven whole‐brain approach for meta‐analysis. Studies based on selected ROIs were excluded,
Were reported in standard stereotaxic spaces such as Montreal Neurological Institute (MNI) [Collins et al., 1994] or Talairach and Tournoux space [Talairach and Tornoux, 1988]
Were driven by categorical contrasts rather than correlation analyses, and
Reported a t‐value ≥3 or a Z‐score ≥2.33 to ensure comparable specificity [Friebel et al., 2011].
Of 63 initially retrieved studies, eight met our inclusion criteria [Coez et al., 2008; Giraud and Truy, 2002; Giraud et al., 2001; Mortensen et al., 2006; Naito et al., 2000; Nishimura et al., 2000; Song et al., in press; Wong et al., 1999]. Of these eight studies, six [Coez et al., 2008; Giraud and Truy, 2002; Giraud et al., 2001; Naito et al., 2000; Song et al., in press; Wong et al., 1999] included NH controls that underwent the same PET experiment paradigms, and thus these six studies were used for ALE meta‐analyses on the cortical activation pattern of sound processing in NH controls. In total, 58 CI subjects with a total of 92 peak coordinates (Table 1) and 45 NH subjects with a total of 40 peak coordinates (Table 2) were included for the current meta‐analysis. The literature search, selection, and compilation of coordinates for the contrast were performed independently by two investigators (J‐J.S and S.V), and the meta‐analysis was performed after confirming that the selected articles for final inclusion were concordant between the two.
Table 1.
Patients included in the ALE meta‐analyses
| Study number | Author | Year | Patients | CI side | Handedness | Contrast | Side | Anatomic Site |
|---|---|---|---|---|---|---|---|---|
| 1 | Wong | 1999 | 5 | 5R | R | Sentence–baseline | R | STG/MTG/TTG |
| L | STG/MTG | |||||||
| Word–baseline | R | STG | ||||||
| L | STG/MTG | |||||||
| Time–reversed sentence–baseline | L | MTG | ||||||
| L | STG | |||||||
| R | Cerebellum | |||||||
| Sentence–time–reversed sentence | R | MTG | ||||||
| 2 | Nishimura | 2000 | 6 | NA | R | Word–baseline | L | Temporal lobe/STG |
| 3 | Naito | 2000 | 12 | 6R, 6L | R | Sentence–baseline | R | STG/MTG |
| R | SMA | |||||||
| L | STG/MTG | |||||||
| L | IFG, post | |||||||
| L | HG | |||||||
| L | ACC | |||||||
| White noise–baseline | R | SMG | ||||||
| R | STG/MTG | |||||||
| R | Thalamus/putamen | |||||||
| R | Cerebellum | |||||||
| R | IFG, post | |||||||
| R | SMA | |||||||
| L | HG | |||||||
| L | STG/MTG | |||||||
| L | IFG, post | |||||||
| L | ACC | |||||||
| 4 | Giraud | 2001 | 6 | 4R, 2L | R | Speech–environmental | L | ITG, post |
| 5 | Giraud | 2002 | 6 | 4R, 2L | NA | French–Norwegian | R | TOP junction |
| R | FG | |||||||
| R | OG | |||||||
| L | SPC | |||||||
| L | ITG, post | |||||||
| L | PFC, inf | |||||||
| 6 | Mortensen | 2006 | 7 | 2R, 5L | NA | White noise–baseline | R | STG |
| L | STG | |||||||
| Word–baseline | R | STG, ant/post | ||||||
| L | STG | |||||||
| L | PFC, inf | |||||||
| Speech–baseline | R | STG | ||||||
| R | cerebellum | |||||||
| L | STG | |||||||
| L | PreCG/SFG | |||||||
| Speech–white noise | R | STG, ant/post | ||||||
| R | cerebellum | |||||||
| L | temporal pole | |||||||
| L | STG, ant/post | |||||||
| L | PFC, inf | |||||||
| Baseline–white noise | R | Cerebellum | ||||||
| L | Cerebellum | |||||||
| Baseline–speech | R | MFG | ||||||
| R | Cuneus/ precuneus | |||||||
| R | Cerebellum | |||||||
| L | Cerebellum | |||||||
| L | Cuneus | |||||||
| white noise −speech | R | MTG | ||||||
| R | Precuneus | |||||||
| R | PHG | |||||||
| L | Lingual gyrus | |||||||
| 7 | Coez | 2008 | 6 | 2R, 4L | R | nonvoice −baseline | R | STS, middle/post |
| R | Distal FG | |||||||
| L | STS, ant/middle/post | |||||||
| Voice–nonvoice | R | STS, middle/post | ||||||
| R | STG | |||||||
| L | STS, ant/middle | |||||||
| 8 | Song | 2014 | 10 | 6R, 4L | R | Word–baseline | R | STG |
| L | MTG | |||||||
| L | STG | |||||||
| L | Temporal pole |
CI, cochlear implant; R, right; L, left; NA, not available; STG, superior temporal gyrus; MTG, middle temporal gyrus; TTG, transverse temporal gyrus; SMA, supplementary motor area; HG, hippocampal gyrus; ACC, anterior cingulate cortex; SMG, supramarginal gyrus; IFG, inferior frontal gyrus; post, posterior; ITG, inferior temporal gyrus; TOP, temporo‐occipito‐parietal; FG, fusiform gyrus; OG, occipital gyrus; SPC, superior parietal cortex; ant, anterior; PFC, prefrontal cortex; inf, inferior; PreCG, precentral gyrus; SFG, superior frontal gyrus; MFG, middle frontal gyrus; PHG, parahippocampal gyrus; STS, superior temporal sulcus; FG, frontal gyrus.
Table 2.
Normal controls included in the ALE meta‐analyses
| Study number | Author | Year | Controls | Handedness | Contrast | Side | Anatomic Site |
|---|---|---|---|---|---|---|---|
| 1 | Wong | 1999 | 5 | R | Word–baseline | R | STG/MTG |
| L | STG | ||||||
| Sentence–baseline | R | MTG | |||||
| L | STG/MTG | ||||||
| Sentence–reversed sentence | L | IFG | |||||
| R | SMG | ||||||
| 2 | Naito | 2000 | 12 | R | Sentence–baseline | R | STG/MTG |
| L | STG | ||||||
| L | IFG, post | ||||||
| L | HG | ||||||
| L | Caudate nucleus | ||||||
| White noise–baseline | R | TTG | |||||
| R | Putamen | ||||||
| R | HG | ||||||
| L | ITG | ||||||
| L | IFG, post | ||||||
| L | ACC | ||||||
| L | Cerebellum | ||||||
| 3 | Giraud | 2001 | 6 | R | Speech–environmental | L | Wernickes' area |
| 4 | Giraud | 2002 | 6 | NA | French–Norwegian | L | ITG, post |
| L | SPC | ||||||
| R | OG | ||||||
| 5 | Coez | 2008 | 6 | R | Nonvoice–baseline | R | STS, ant/post |
| R | Uncus | ||||||
| L | STS, ant/post | ||||||
| Voice–nonvoice | R | STS, middle/post | |||||
| L | STS, middle/post | ||||||
| 6 | Song | 2014 | 10 | R | Word–baseline | R | STG |
| R | MTG | ||||||
| R | Thalamus | ||||||
| L | STG | ||||||
| L | MTG | ||||||
| L | Thalamus |
CI, cochlear implant; R, right; L, left; NA, not available; STG, superior temporal gyrus; MTG, middle temporal gyrus; IFG: inferior frontal gyrus; SMG, supramarginal gyrus; post, posterior; HG, hippocampal gyrus; TTG, transverse temporal gyrus; ITG, inferior temporal gyrus; ACC, anterior cingulate cortex; SPC, superior parietal cortex; OG, occipital gyrus; ant, anterior.
Meta‐Analysis Algorithm
To determine the concurrence in reported coordinates across the included studies, we conducted ALE meta‐analyses for three contrasts, namely “lexical stimuli–baseline,” “nonlexical stimuli–baseline,” and “lexical stimuli–nonlexical stimuli,” both in CI users and NH controls using the software Brainmap GingerALE (http://brainmap.org/ale/index.html) desktop application ver. 2.3.1. Because only one study [Mortensen et al., 2006] provided results from the contrast “nonlexical stimuli–lexical stimuli,” we could not conduct ALE meta‐analysis for this contrast. The method we used for our meta‐analysis is the latest revised version of the ALE approach [Turkeltaub et al., 2012] for coordinate‐based meta‐analysis of neuroimaging results [Eickhoff et al., 2009; Laird et al., 2005; Turkeltaub et al., 2002]. Turkeltaub's nonadditive ALE algorithm minimizes both within‐experiment and within‐group effects, and thus, optimizes the degree to which ALE values represent concordance of statistically significant findings across independent studies included [Eickhoff et al., 2012; Turkeltaub et al., 2012]. For each analysis, we used Talairach and Tournoux's coordinate space [Talairach and Tornoux, 1988] and more conservative gray matter mask in the GingerALE preferences menu. The FWHM values were subject‐based [Eickhoff et al., 2009], with no additional FWHM.
The input for the first meta‐analysis consisted of the coordinates of brain regions that were activated in response to lexical stimuli (words or sentences) compared to baseline images. For the first analysis, studies 1, 2, 3, 6, and 8 (40 CI subjects with 40 foci) were adopted for the CI user group and studies 1, 2, and 6 (27 NH controls with 25 foci) were adopted for the NH control group (Table 3). For the second analysis, coordinates that were significantly modulated by nonlexical stimuli such as white noise or nonvoice sounds as compared with baseline were analyzed. For the second analysis, studies 1, 3, 6, and 7 (30 CI subjects with 20 foci) were used for the CI user group and studies 1, 2, and 5 (23 NH controls with 18 foci) were used for the NH control group (Table 4). The third meta‐analysis included coordinates of brain regions that were activated more by lexical stimuli than by nonlexical stimuli. For this third analysis, studies 1, 4, 5, and 6 (24 CI subjects with 20 foci) were used for the CI user group and studies 1, 3, 4, and 5 (23 NH controls with 9 foci) were used for the NH control group (Table 5).
Table 3.
Summary of the spatial location and extent of ALE values for the contrast (lexical stimuli–baseline) in the cochlear implant patients and normal hearing controls (FDR corrected P < .05, voxel size > 50)
| Weighted Center | |||||||
|---|---|---|---|---|---|---|---|
| Side | BA | Region | Volume (mm3) | x | y | z | Max ALE Value |
| Cochlear implant patients (contributed studies: studies 1, 2, 3, 6, and 8 of Table I ) | |||||||
| L | 22 | STG/STS | 3,528 | −57 | −25 | 3 | 0.014 |
| L | 22 | MTG | i.above | 0.013 | |||
| L | 22 | STG/STS | i.above | 0.010 | |||
| R | 22 | STG/STS | 1,080 | 57 | −24 | 4 | 0.012 |
| R | 38 | STG | 1,008 | 52 | 1 | −7 | 0.011 |
| L | 22 | STG | 368 | −48 | −6 | −4 | 0.009 |
| R | MTG | 112 | 50 | −31 | 0 | 0.007 | |
| L | 38 | STG | 104 | −53 | 5 | −7 | 0.007 |
| L | 21 | MTG | 80 | −59 | −3 | −3 | 0.007 |
| L | 28 | PHG | 56 | −30 | 6 | −20 | 0.007 |
| L | 13 | Insula | 56 | −44 | 10 | 12 | 0.007 |
| L | 32 | ACC | 56 | −6 | 16 | 40 | 0.007 |
| R | 6 | SMA | 56 | 4 | 8 | 52 | 0.007 |
| Normal hearing controls (contributed studies: studies 1, 2, and 6 of Table II ) | |||||||
| R | 21 | MTG | 552 | 57 | −8 | −5 | 0.011 |
| R | 22 | STG/STS | 368 | 58 | −25 | 1 | 0.009 |
| L | 22 | STG | 200 | −49 | −10 | 0 | 0.008 |
| L | 22 | MTG | 72 | −50 | −36 | 4 | 0.007 |
| R | 38 | STG | 64 | 48 | 2 | −8 | 0.007 |
| L | MTG | 64 | −65 | −26 | 0 | 0.007 | |
| L | 13 | Insula | 56 | −38 | −20 | −8 | 0.007 |
| L | 44 | pIFG | 56 | −42 | 14 | 8 | 0.007 |
| L | Caudate body | 56 | −14 | 20 | 8 | 0.007 | |
| L | Caudate tail | 56 | −26 | −34 | 12 | 0.007 | |
FDR, false detection rate; L, left; R, right; BA, Brodmann area; STG, superior temporal gyrus; STS, superior temporal sulcus; MTG, middle temporal gyrus; ACC, anterior cingulate cortex; SMA, supplementary motor area; pIFG, posterior inferior frontal gyrus.
Table 4.
Summary of the spatial location and extent of ALE values for the contrast (nonlexical stimuli–baseline) in the cochlear implant patients and normal hearing controls (FDR corrected P < .05, voxel size > 50)
| Weighted Center | |||||||
|---|---|---|---|---|---|---|---|
| Side | BA | Region | Volume (mm3) | x | y | z | Max ALE Value |
| Cochlear implant patients (contributed studies: studies 1, 3, 6, and 7 of Table I ) | |||||||
| L | 22 | STG/STS | 480 | −55 | −26 | 4 | 0.008 |
| L | 22 | STG | 248 | −48 | 2 | −5 | 0.007 |
| R | Cbll culmen | 160 | 26 | −56 | −24 | 0.007 | |
| R | Cbll culmen | 152 | 6 | −42 | −20 | 0.007 | |
| R | 13 | Insula | 152 | 34 | −30 | 26 | 0.007 |
| R | Caudate body | 96 | 15 | 12 | 11 | 0.007 | |
| L | 24 | dACC | 96 | −3 | 4 | 35 | 0.007 |
| L | Cbll culmen | 80 | −11 | −42 | −2 | 0.007 | |
| R | Putamen | 80 | 24 | 4 | 7 | 0.007 | |
| R | Cbll vermis | 64 | 1 | −73 | −15 | 0.007 | |
| R | Thalamic VLN | 64 | 17 | −15 | 17 | 0.007 | |
| Normal hearing controls (contributed studies: studies 1, 2, and 5 of Table II ) | |||||||
| R | 21 | STG/STS | 448 | 55 | −9 | −4 | 0.009 |
| L | Cbll culmen | 152 | −6 | −48 | −6 | 0.007 | |
| R | Caudate tail | 96 | 31 | −36 | 5 | 0.007 | |
| R | Putamen | 96 | 15 | 3 | 10 | 0.007 | |
| R | Caudate tail | 96 | 33 | −44 | 15 | 0.007 | |
| R | 36 | PHG | 80 | 33 | −30 | −12 | 0.007 |
| L | Cbll culmen | 80 | −1 | −38 | −6 | 0.007 | |
FDR, false detection rate; L, left; R, right; BA, Brodmann area; STG, superior temporal gyrus; STS, superior temporal sulcus; Cbll, cerebellar; dACC, dorsal anterior cingulate cortex; VLN, ventrolateral nucleus; MTG, middle temporal gyrus; FG, fusiform gyrus; PHG, parahippocampal gyrus.
Table 5.
Summary of the spatial location and extent of ALE values for the contrast (lexical stimuli–nonlexical stimuli) in the cochlear implant patients and normal hearing controls (FDR corrected P < .05, voxel size > 50)
| Weighted Center | |||||||
|---|---|---|---|---|---|---|---|
| Side | BA | Region | Volume (mm3) | x | y | z | Max ALE Value |
| Cochlear implant patients (contributed studies: studies 1, 4, 5, and 6 of Table I ) | |||||||
| R | 22 | STG/STS | 792 | 60 | −24 | 1 | 0.010 |
| L | 21 | MTG | 744 | −60 | −10 | −3 | 0.007 |
| L | 22 | STG/STS | 176 | −60 | −25 | 2 | 0.006 |
| R | Cbll uvula | 80 | 25 | −80 | −24 | 0.005 | |
| L | 38 | STG | 80 | −38 | 12 | −22 | 0.006 |
| L | 47 | pIFG | 80 | −46 | 17 | −4 | 0.005 |
| R | 22 | STG | 80 | 55 | −10 | −3 | 0.005 |
| L | 37 | Fusiform G | 56 | −46 | −46 | −14 | 0.005 |
| L | 22 | STG | 56 | −34 | 24 | −14 | 0.005 |
| L | 37 | MTG | 56 | −42 | −58 | −4 | 0.005 |
| Normal hearing controls (contributed studies: studies 1, 3, 4, and 5 of Table II ) | |||||||
| L | 21 | MTG | 160 | −64 | −30 | −2 | 0.005 |
| R | 18 | Fusiform G | 56 | 22 | −90 | −14 | 0.005 |
| L | 19 | ITG | 56 | −44 | −58 | −6 | 0.005 |
| R | 21 | STS | 56 | 62 | −30 | 0 | 0.005 |
| L | 22 | STS | 56 | −62 | −14 | 0 | 0.005 |
| L | 5 | PoCG | 56 | −34 | −40 | 58 | 0.005 |
FDR, false detection rate; L, left; R, right; BA, Brodmann area; STG, superior temporal gyrus; STS, superior temporal sulcus; pIFG, posterior inferior frontal gyrus; MTG, middle temporal gyrus; Cbll, cerebellar; ITG, inferior temporal gyrus; FG, fusiform gyrus; PoCG, postcentral gyrus.
For the resulting ALE map, a statistical threshold of P < 0.05, false discovery rate (FDR)‐corrected for multiple comparisons [Laird et al., 2005] and a minimum cluster size of 50 mm3 were adopted. Statistically significant voxels obtained by the current ALE meta‐analyses represent the concordance of the investigated effect across the studies. ALE results were then overlaid onto a standardized individual anatomical T1‐template (http://www.brainmap.org/ale/Colin1.1.nii) and cluster centers were anatomically delineated using MRIcron software (http://www.sph.sc.edu/comd/rorden/mricron/) [Rorden et al., 2007]. To describe correct location of each cluster center, all the peak coordinates and their designated locations both as brain regions and Brodmann areas (BAs) were reconfirmed using Talairach and Tournoux's atlas [Talairach and Tornoux, 1988].
RESULTS
Significant ALE Clusters for the Contrast (Lexical Stimuli–Resting State)
The results of the first analysis, contrasting lexical stimuli and baseline resting state both in CI users and NH controls are summarized in Figure 1 and Table 3. For the lexical stimuli condition, both the CI user‐ and NH control groups revealed activation of bilateral primary auditory‐ and secondary auditory cortices (A1 and A2), namely superior‐ and middle temporal gyri (STG and MTG, BAs 21 and 22) and superior temporal sulcus (STS). In CI users, significant activations for lexical stimuli were observed in the left insula (BA 13), left dorsal anterior cingulate cortex (dACC, BA 32), left parahippocampal gyrus (PHG, BA 28), and the right supplementary motor area (SMA, BA 6; Fig. 1, upper panels). In contrast, the NH group displayed increased rCBF in the left insula (BA 13), left posterior inferior frontal gyrus (pIFG, BA 44), and the caudate body and tail (Fig. 1, lower panels).
Figure 1.

Main activation effects in cochlear implant (CI) subjects (upper panel) and normal hearing (NH) controls (lower panel) under lexical stimuli (FDR corrected P < 0.05, k > 50 voxels). [Color figure can be viewed in the online issue, which is available at http://wileyonlinelibrary.com.]
Significant ALE Clusters for the Contrast (Nonlexical Stimuli–Resting State)
By the second analysis contrasting nonlexical sound stimuli and silent baseline, the CI user group showed significant increases in rCBF in the left A2 (BA 22) and the NH control group in the right A2 (BA 21; Fig. 2 and Table 4). Also, both groups displayed increased rCBF in the cerebellum (the CI user group in the cerebellar culmen and vermis; the NH control group in the cerebellar culmen). Of note, the CI user group activated the right insula (BA 13), right thalamic ventrolateral nucleus VLN, right caudate body, and the left dACC (BA 24) under the nonlexical stimuli condition (Fig. 2, upper panels), while the NH control group revealed activations in the right caudate tail and right PHG (BA 36; Fig. 2, lower panels).
Figure 2.

Main activation effects in CI subjects (upper panel) and NH controls (lower panel) under lexical stimuli (FDR corrected P < 0.05, k > 50 voxels). [Color figure can be viewed in the online issue, which is available at http://wileyonlinelibrary.com.]
Significant ALE Clusters for the Contrast (Lexical Stimuli–Nonlexical Stimuli)
The third contrast, lexical stimuli minus nonlexical stimuli, revealed significant increases in rCBF in the bilateral A1/A2 and the fusiform gyrus in both the CI user‐ and NH control groups (Fig. 3 and Table 5). The CI user group uniquely activated the left inferior frontal gyrus (IFG, BA 47) more under the lexical stimuli than under the nonlexical stimuli (Fig. 3, upper panels). Also, the CI users showed increased activation in the cerebellar uvula under the lexical stimuli as compared with nonlexical stimuli. In contrast, the NH controls displayed increased activation in the postcentral gyrus (PoCG, BA 5) under the lexical stimuli as compared with nonlexical stimuli (Fig. 3, lower panels).
Figure 3.

Main activation effects in CI subjects (upper panel) and NH controls (lower panel) for the contrast “lexical stimuli–nonlexical stimuli (FDR corrected P < 0.05, k > 50 voxels). [Color figure can be viewed in the online issue, which is available at http://wileyonlinelibrary.com.]
DISCUSSION
To understand the mechanism of speech processing in NH individuals, researchers have endeavored to develop integrative models for speech perception. The dual‐stream model [Hickok and Poeppel, 2007; Hickok et al., 2011], a representative model of speech processing, holds that early stages of speech perception occurs in auditory regions consisting of the dorsal STG and STS bilaterally and then diverges into two broad streams: the ventral stream, a bilaterally organized structure consisting of MTG and inferior temporal sulcus that directly processes speech signals for comprehension, and the dorsal stream, a strongly left‐hemisphere dominant structure consisting of parietal–temporal junction, posterior IFG, premotor cortex, and anterior insula that maps acoustic speech signals to frontal lobe articulatory networks [Hickok and Poeppel, 2000, 2004, 2007].
As in NH subjects, researchers have been investigating the cortical organization of speech processing in CI users. Parts of cortical areas that were suggested to be involved in the dual‐stream model in NH subjects also showed increased activity by auditory stimuli in CI users [Giraud et al., 2000], visually evoked phonological and environmental sound representations [Lazard et al., 2013], or speechreading [Lee et al., 2007b]. Also, an fMRI study in postlingual deafened subjects using rhyming task on written regular words has indicated that subjects who rely on a dorsal phonological route will become good CI performers while ventral temporo‐frontal route‐dependent subjects will become poor CI performers [Lazard et al., 2010].
However, these observations in CI users do not preclude the possibility of different speech processing strategy, especially considering deafness‐induced plastic changes and intrinsically distorted and impoverished auditory inputs provided by the CI [Lazard et al., 2012]. Indeed, by performing the current meta‐analysis, we corroborated different speech and nonspeech sound processing patterns in CI users.
Activation of the Dual Streams for Speech Processing in CI Users as in NH Controls
For lexical stimuli, the NH control group showed increased rCBF in areas comprising the phonological network (bilateral STS), spectrotemporal analysis (STG), articulatory network (left insula and pIFG), and lexical interface (MTG) of the dual‐stream speech processing system [Hickok and Poeppel, 2004, 2007] (Fig. 1 and Table 3). For the contrast “lexical stimuli–nonlexical stimuli,” NH controls displayed increased activations in areas comprising the phonological network (bilateral STS), lexical interface (MTG), and sensorimotor interface (PoCG; Fig. 3 and Table 5). The NH control group in the current meta‐analysis, therefore, used both ventral and dorsal language processing routes for lexical stimuli.
The CI user group also showed increased rCBF in areas comprising the phonological network (bilateral STS), spectrotemporal analysis (STG), articulatory network (left insula), and lexical interface (MTG) of the dual‐stream speech processing system for lexical stimuli (Fig. 1 and Table 3). Also, the CI user group showed increased activations in areas at the phonological network (bilateral STS), spectrotemporal analysis (STG), articulatory network (left pIFG), lexical interface (left posterior MTG), and combinatorial network (left anterior MTG) for the contrast “lexical stimuli–nonlexical stimuli” (Fig. 3 and Table 5). By observing these relative activations, utilization of the dual‐stream speech processing system for lexical stimuli in CI users was reconfirmed, and these results are in line with previous reports [Giraud and Lee, 2007; Lazard et al., 2010; Lee et al., 2007a]. In summary, these episodic reports on the utilization of dual auditory stream for speech processing in CI users were reconfirmed in this study in a meta‐analytic level.
Meanwhile, nonlexical stimuli increased cerebellar areas both in the CI user group and NH group (Fig. 2 and Table 4). As the cerebellum may be involved in purely sensory auditory processing [Petacchi et al., 2005], cerebellar activations may designate similar sound processing strategies for nonlexical stimuli in both groups.
One important discrepancy between the current and previous studies should be addressed. Previous reports have suggested an important role of the right posterior temporal gyrus in processing nonspeech sound not only in NH subjects [Halpern and Zatorre, 1999] but also in CI users [Lazard et al., 2011]. In contrast, the current meta‐analysis found right posterior temporal activation only in the NH control group, not in the CI user group. This may be attributed to the differences in task dimension (sound imagery vs. real sound presentation) or sample size‐related differences in statistical power. Future studies should be performed to reevaluate this discrepancy.
CI Users Need Attention Control to Process Speech Sound: Activation of Salience Network for Lexical Stimuli
Adding to the utilization of the dual‐stream speech processing system, CI users demonstrated unique activation patterns that were not observed in NH controls both for lexical and nonlexical stimuli.
For the lexical stimuli, significant activations were observed in areas at the left dACC, left insula, and right SMA in the CI user group. These three areas are left‐sided components of previously reported bilateral salience network (SN) [Bonnelle et al., 2012; Seeley et al., 2007], a network activated when behaviorally relevant information is processed [Seeley et al., 2007; Sridharan et al., 2008]. The SN operates to control dynamic changes of activity in other cortical networks [Sridharan et al., 2008], and the role of the SN as a mediator of the function of other networks seems to be most evident when a rapid change in behavior is required. In particular, the integrity of the SN is crucial for the efficient regulation of activity in the default mode network [Bonnelle et al., 2012] that is important for efficient behavioral performance. Considering impoverished auditory input processed by the CI, the activation of the SN for lexical stimuli may indicate that salience is attached to the stimuli, both top down, that is, behaviorally relevant [Seeley et al., 2007], and bottom up, that is, distinctive [Fecteau and Munoz, 2006; Serences and Yantis, 2006], to process distorted speech cues in CI users. In NH people listening to acoustically degraded words using vocoder, the activity in the dACC was mainly affected only when there was a behavioral task involved [Obleser and Weisz, 2012]. This is in line with the results of the current study in that CI users in the included studies were exposed to degraded lexical auditory input, and of five studies used for the contrast “lexical stimuli–baseline,” four used tasks such as button‐pressing [Wong et al., 1999], mouse‐clicking [Song et al., in press], or answering to questions after each measurement [Mortensen et al., 2006; Naito et al., 2000] to maintain attention level of the participants. The activation of the SN may also explain markedly broad activation of the A1/A2 in CI users for lexical stimuli as compared with NH controls (Fig. 1), which is not prominent for nonlexical stimuli. That is, probably because understanding lexical stimuli is more crucial as well as attention‐consuming in CI users than in NH controls, attention control mediated by the SN in CI users may have resulted in broader activation of the A1/A2 as compared with NH controls who could process lexical stimuli with less effort. Additionally, as a corollary to effortful interpretation of distorted lexical signal, the left PHG in the CI user group may have been activated to retrieve semantic memory [Kirwan et al., 2009; Whatmough and Chertkow, 2007].
The included studies used different types of lexical stimuli (i.e., sentences or words). Different types of lexical stimuli may have affected the result to some extent, because cortical regions such as the anterior temporal lobe are known to be specifically involved in sentence‐level processing in NH people [Hickok and Poeppel, 2007]. However, as in other ALE meta‐analyses that used different types of stimuli [Duerden and Albanese, 2013; van der Laan et al., 2011], different types of lexical stimuli used in this study may add information to previously performed individual studies from a holistic point of view.
CI Users Process Nonspeech Sound with Extra Effort: Activation of SN for Nonlexical Stimuli
Meanwhile, for nonlexical stimuli, CI users activated subareas of the SN such as the right insula and left dACC. As in the lexical stimuli condition, CI users in this study may have activated the SN to detect and process nonlexical stimuli with poorly represented spectral and temporal cues. Together with the unique activation of the SN for lexical stimuli, the CI user group demonstrated much effortful processing of both lexical and nonlexical stimuli as compared with the NH control group.
Of note, CI users showed activations significant in areas at the left A2 but the NH control group at the right A2 for nonlexical stimuli (Fig. 2). This may show that CI users have troubles differentiating lexical and nonlexical stimuli during low level processing due to degraded sound input. The unique activation of the left IFG for the contrast “lexical stimuli–nonlexical stimuli” in the CI user group segregation may also prove that the differentiation between lexical and nonlexical sound is appearing at higher processing levels in CI users. In this regard, the insula and dACC may have been activated to better detect sound stimuli, although the stimuli presented were nonlexical.
Strength and Limitations of this Study
To the best of our knowledge, this is the first study that used a coordinate‐based ALE meta‐analytic approach to systematically determine consistency across PET studies on speech and nonspeech sound perception in CI subjects. By minimizing study‐to‐study disparities with regard to paradigm, we attempted to see common neural correlates for sound processing in CI users. The principal strength of this quantitative meta‐analysis is that it is based on multiple peer‐reviewed studies, in our case with 58 CI patients with 92 peak coordinates as well as 45 NH subjects with 40 coordinates. Thus, the results from the present CI‐related brain activation maps are more robust than those of any individual imaging study on sound perception in CI users. Also, by comparing the results of CI users with those of NH controls, we have reconfirmed similarities and disclosed disparities in the two groups with regard to speech and nonspeech sound processing strategy.
However, limitations of this study should be addressed. First, a limitation of the ALE analysis is that it does not take into account the level of statistical significance and the cluster size. However, it is unlikely that the variation in statistical thresholds has otherwise significantly biased the obtained results because false positives from a single study are averaged out when multiple studies are combined. Second, because only two studies [Coez et al., 2008; Mortensen et al., 2006] included poor performers, we could only investigate neural correlates of good performers. Also, we could not adopt a contrast “nonlexical stimuli–lexical stimuli” because only one study [Mortensen et al., 2006] provided results from the contrast. Future meta‐analysis of poor performers and comparison to the results of this study may enable us to further understand sound processing mechanisms and critical factors for good performance in CI users, and future meta‐analysis using the contrast “nonlexical stimuli–lexical stimuli” may further prove the role of alertness in CI‐assisted sound processing. Third, the effect of the side of CI could not be explored because all but one study [Wong et al., 1999] included both right‐ and left‐CI users. Although previous researchers reported that neither the side of CI nor the number of active electrodes had a significant effect on the brain regions activated in previous studies [Giraud et al., 2001; Naito et al., 2000], future meta‐analysis comparing the effect of the side of CI on cortical activity may give us additional information. Furthermore, subtraction analyses such as “CI (lexical–baseline)–NH (lexical–baseline)” or “CI (nonlexical–baseline)–NH (nonlexical–baseline)” did not show significant foci of relative activation because the number of included studies was not enough for yielding significant results [Eickhoff et al., 2011]. To further explore the differences between the two groups using subtraction analyses, future follow‐up meta‐analyses on larger number of studies should be performed. Finally, a longitudinal study with serial functional imaging study may further elucidate the change in the level of attention needed to process sound provided by CI.
CONCLUSION
Taken together, previous episodic observations on lexical stimuli processing using the dual auditory stream in CI users were reconfirmed in a meta‐analytic level by this study. However, this study also suggests that dual‐stream auditory processing in CI users may need support from the SN. In other words, CI users need to pay extra attention to cope with degraded auditory signal provided by the implant, although they use the same dual‐stream pathways as NH peers do.
REFERENCES
- Bonnelle V, Ham TE, Leech R, Kinnunen KM, Mehta MA, Greenwood RJ, Sharp DJ (2012): Salience network integrity predicts default mode network function after traumatic brain injury. Proc Natl Acad Sci USA 109:4690–4695. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Coez A, Zilbovicius M, Ferrary E, Bouccara D, Mosnier I, Ambert‐Dahan E, Bizaguet E, Syrota A, Samson Y, Sterkers O (2008): Cochlear implant benefits in deafness rehabilitation: PET study of temporal voice activations. J Nucl Med 49:60–67. [DOI] [PubMed] [Google Scholar]
- Collins DL, Neelin P, Peters TM, Evans AC (1994): Automatic 3D intersubject registration of MR volumetric data in standardized Talairach space. J Comput Assist Tomogr 18:192–205. [PubMed] [Google Scholar]
- Duerden EG, Albanese MC (2013): Localization of pain‐related brain activation: A meta‐analysis of neuroimaging data. Hum Brain Mapp 34:109–149. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eickhoff SB, Laird AR, Grefkes C, Wang LE, Zilles K, Fox PT (2009): Coordinate‐based activation likelihood estimation meta‐analysis of neuroimaging data: A random‐effects approach based on empirical estimates of spatial uncertainty. Hum Brain Mapp 30:2907–2926. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eickhoff SB, Bzdok D, Laird AR, Roski C, Caspers S, Zilles K, Fox PT (2011): Co‐activation patterns distinguish cortical modules, their connectivity and functional differentiation. Neuroimage 57:938–949. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eickhoff SB, Bzdok D, Laird AR, Kurth F, Fox PT (2012): Activation likelihood estimation meta‐analysis revisited. Neuroimage 59:2349–2361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fecteau JH, Munoz DP (2006): Salience, relevance, and firing: A priority map for target selection. Trends Cogn Sci 10:382–390. [DOI] [PubMed] [Google Scholar]
- Friebel U, Eickhoff SB, Lotze M (2011): Coordinate‐based meta‐analysis of experimentally induced and chronic persistent neuropathic pain. Neuroimage 58:1070–1080. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Giraud AL, Truy E (2002): The contribution of visual areas to speech comprehension: A PET study in cochlear implants patients and normal‐hearing subjects. Neuropsychologia 40:1562–1569. [DOI] [PubMed] [Google Scholar]
- Giraud AL, Lee HJ (2007): Predicting cochlear implant outcome from brain organisation in the deaf. Restor Neurol Neurosci 25:381–390. [PubMed] [Google Scholar]
- Giraud AL, Truy E, Frackowiak RS, Gregoire MC, Pujol JF, Collet L (2000): Differential recruitment of the speech processing system in healthy subjects and rehabilitated cochlear implant patients. Brain 123 (Pt 7):1391–1402. [DOI] [PubMed] [Google Scholar]
- Giraud AL, Price CJ, Graham JM, Frackowiak RS (2001): Functional plasticity of language‐related brain areas after cochlear implantation. Brain 124:1307–1316. [DOI] [PubMed] [Google Scholar]
- Halpern AR, Zatorre RJ (1999): When that tune runs through your head: A PET investigation of auditory imagery for familiar melodies. Cereb Cortex 9:697–704. [DOI] [PubMed] [Google Scholar]
- Hickok G, Poeppel D (2000): Towards a functional neuroanatomy of speech perception. Trends Cogn Sci 4:131–138. [DOI] [PubMed] [Google Scholar]
- Hickok G, Poeppel D (2004): Dorsal and ventral streams: A framework for understanding aspects of the functional anatomy of language. Cognition 92:67–99. [DOI] [PubMed] [Google Scholar]
- Hickok G, Poeppel D (2007): The cortical organization of speech processing. Nat Rev Neurosci 8:393–402. [DOI] [PubMed] [Google Scholar]
- Hickok G, Houde J, Rong F (2011): Sensorimotor integration in speech processing: Computational basis and neural organization. Neuron 69:407–422. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kirwan CB, Shrager Y, Squire LR (2009): Medial temporal lobe activity can distinguish between old and new stimuli independently of overt behavioral choice. Proc Natl Acad Sci USA 106:14617–14621. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kral A, O'Donoghue GM (2010): Profound deafness in childhood. N Engl J Med 363:1438–1450. [DOI] [PubMed] [Google Scholar]
- Laird AR, Fox PM, Price CJ, Glahn DC, Uecker AM, Lancaster JL, Turkeltaub PE, Kochunov P, Fox PT (2005): ALE meta‐analysis: Controlling the false discovery rate and performing statistical contrasts. Hum Brain Mapp 25:155–164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lazard DS, Lee HJ, Gaebler M, Kell CA, Truy E, Giraud AL (2010): Phonological processing in post‐lingual deafness and cochlear implant outcome. Neuroimage 49:3443–3451. [DOI] [PubMed] [Google Scholar]
- Lazard DS, Giraud AL, Truy E, Lee HJ (2011): Evolution of non‐speech sound memory in postlingual deafness: Implications for cochlear implant rehabilitation. Neuropsychologia 49:2475–2482. [DOI] [PubMed] [Google Scholar]
- Lazard DS, Marozeau J, McDermott HJ (2012): The sound sensation of apical electric stimulation in cochlear implant recipients with contralateral residual hearing. PLoS One 7:e38687. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lazard DS, Lee HJ, Truy E, Giraud AL (2013): Bilateral reorganization of posterior temporal cortices in post‐lingual deafness and its relation to cochlear implant outcome. Hum Brain Mapp 34:1208–1219. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lazard DS, Innes‐Brown H, Barone P (2014): Adaptation of the communicative brain to post‐lingual deafness. Evidence from functional imaging. Hear Res 307:136–143. [DOI] [PubMed] [Google Scholar]
- Lee DS, Lee JS, Oh SH, Kim SK, Kim JW, Chung JK, Lee MC, Kim CS (2001): Cross‐modal plasticity and cochlear implants. Nature 409:149–150. [DOI] [PubMed] [Google Scholar]
- Lee HJ, Giraud AL, Kang E, Oh SH, Kang H, Kim CS, Lee DS (2007a): Cortical activity at rest predicts cochlear implantation outcome. Cereb Cortex 17:909–917. [DOI] [PubMed] [Google Scholar]
- Lee HJ, Truy E, Mamou G, Sappey‐Marinier D, Giraud AL (2007b): Visual speech circuits in profound acquired deafness: A possible role for latent multimodal connectivity. Brain 130:2929–2941. [DOI] [PubMed] [Google Scholar]
- Moher D, Liberati A, Tetzlaff J, Altman DG (2009): Preferred reporting items for systematic reviews and meta‐analyses: The PRISMA statement. PLoS Med 6:e1000097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mortensen MV, Mirz F, Gjedde A (2006): Restored speech comprehension linked to activity in left inferior prefrontal and right temporal cortices in postlingual deafness. Neuroimage 31:842–852. [DOI] [PubMed] [Google Scholar]
- Naito Y, Okazawa H, Honjo I, Hirano S, Takahashi H, Shiomi Y, Hoji W, Kawano M, Ishizu K, Yonekura Y (1995): Cortical activation with sound stimulation in cochlear implant users demonstrated by positron emission tomography. Brain Res Cogn Brain Res 2:207–214. [DOI] [PubMed] [Google Scholar]
- Naito Y, Tateya I, Fujiki N, Hirano S, Ishizu K, Nagahama Y, Fukuyama H, Kojima H (2000): Increased cortical activation during hearing of speech in cochlear implant users. Hear Res 143:139–146. [DOI] [PubMed] [Google Scholar]
- Nishimura H, Doi K, Iwaki T, Hashikawa K, Oku N, Teratani T, Hasegawa T, Watanabe A, Nishimura T, Kubo T (2000): Neural plasticity detected in short‐ and long‐term cochlear implant users using PET. Neuroreport 11:811–815. [DOI] [PubMed] [Google Scholar]
- Obleser J, Weisz N (2012): Suppressed alpha oscillations predict intelligibility of speech and its acoustic details. Cereb Cortex 22:2466–2477. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Okazawa H, Naito Y, Yonekura Y, Sadato N, Hirano S, Nishizawa S, Magata Y, Ishizu K, Tamaki N, Honjo I and others (1996): Cochlear implant efficiency in pre‐ and postlingually deaf subjects. A study with H2(15)O and PET. Brain 119 (Pt 4):1297–1306. [DOI] [PubMed] [Google Scholar]
- Petacchi A, Laird AR, Fox PT, Bower JM (2005): Cerebellum and auditory function: An ALE meta‐analysis of functional neuroimaging studies. Hum Brain Mapp 25:118–128. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rauschecker JP (2011): An expanded role for the dorsal auditory pathway in sensorimotor control and integration. Hear Res 271:16–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rorden C, Karnath HO, Bonilha L (2007): Improving lesion‐symptom mapping. J Cogn Neurosci 19:1081–1088. [DOI] [PubMed] [Google Scholar]
- Seeley WW, Menon V, Schatzberg AF, Keller J, Glover GH, Kenna H, Reiss AL, Greicius MD (2007): Dissociable intrinsic connectivity networks for salience processing and executive control. J Neurosci 27:2349–2356. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Serences JT, Yantis S (2006): Selective visual attention and perceptual coherence. Trends Cogn Sci 10:38–45. [DOI] [PubMed] [Google Scholar]
- Song JJ, De Ridder D, Van de Heyning P, Vanneste S (2012): Mapping tinnitus‐related brain activation: An activation‐likelihood estimation metaanalysis of PET studies. J Nucl Med 53:1550–1557. [DOI] [PubMed] [Google Scholar]
- Song JJ, De Ridder D, Schlee W, Van de Heyning P, Vanneste S (2013a): "Distressed aging": The differences in brain activity between early‐ and late‐onset tinnitus. Neurobiol Aging 34:1853–1863. [DOI] [PubMed] [Google Scholar]
- Song JJ, Punte AK, De Ridder D, Vanneste S, Van de Heyning P (2013b): Neural substrates predicting improvement of tinnitus after cochlear implantation in patients with single‐sided deafness. Hear Res 299:1–9. [DOI] [PubMed] [Google Scholar]
- Song JJ, De Ridder D, Weisz N, Schlee W, Van de Heyning P, Vanneste S (2014a): Hyperacusis‐associated pathological resting‐state brain oscillations in the tinnitus brain: A hyperresponsiveness network with paradoxically inactive auditory cortex. Brain Struct Funct 219:1113–1128. [DOI] [PubMed] [Google Scholar]
- Song JJ, Mertens G, Deleye S, Staelens S, Ceyssens S, Gilles A, de Bodt M, Vanneste S, De Ridder D, Kim E and others (2014b): Neural Substrates of Conversion Deafness in a Cochlear Implant Patient: A Molecular Imaging Study Using H215O‐PET. Otol Neurotol 35:1780–1784. [DOI] [PubMed] [Google Scholar]
- Song JJ, Vanneste S, Schlee W, Van de Heyning P, De Ridder D (2015): Onset‐related differences in neural substrates of tinnitus‐related distress: The anterior cingulate cortex in late‐onset tinnitus, and the frontal cortex in early‐onset tinnitus. Brain Struct Funct 220:571–584. [DOI] [PubMed] [Google Scholar]
- Song JJ, Lee HJ, Kang H, Lee DS, Chang SO, Oh SH: Effects of congruent and incongruent visual cues on speech perception and brain activity in cochlear implant users. Brain Struct Funct (in press). [DOI] [PubMed] [Google Scholar]
- Sridharan D, Levitin DJ, Menon V (2008): A critical role for the right fronto‐insular cortex in switching between central‐executive and default‐mode networks. Proc Natl Acad Sci USA 105:12569–12574. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Strelnikov K, Rouger J, Demonet JF, Lagleyre S, Fraysse B, Deguine O, Barone P (2013): Visual activity predicts auditory recovery from deafness after adult cochlear implantation. Brain 136:3682–3695. [DOI] [PubMed] [Google Scholar]
- Talairach J, Tornoux P. 1988. Co‐Planar Stereotaxic Atlas of the Human Brain: 3‐Dimensional Proportional System: An Approach to Cerebral Imaging. Stuttgart: Georg Thieme.
- Turkeltaub PE, Eden GF, Jones KM, Zeffiro TA (2002): Meta‐analysis of the functional neuroanatomy of single‐word reading: Method and validation. Neuroimage 16:765–780. [DOI] [PubMed] [Google Scholar]
- Turkeltaub PE, Eickhoff SB, Laird AR, Fox M, Wiener M, Fox P (2012): Minimizing within‐experiment and within‐group effects in Activation Likelihood Estimation meta‐analyses. Hum Brain Mapp 33:1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- van der Laan LN, de Ridder DT, Viergever MA, Smeets PA (2011): The first taste is always with the eyes: A meta‐analysis on the neural correlates of processing visual food cues. Neuroimage 55:296–303. [DOI] [PubMed] [Google Scholar]
- Whatmough C, Chertkow H (2007): rCBF to the hippocampal complex covaries with superior semantic memory retrieval. Behav Brain Res 181:262–269. [DOI] [PubMed] [Google Scholar]
- Wong D, Miyamoto RT, Pisoni DB, Sehgal M, Hutchins GD (1999): PET imaging of cochlear‐implant and normal‐hearing subjects listening to speech and nonspeech. Hear Res 132:34–42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zeitler DM, Wang KH, Prasad RS, Wang EY, Roland JT (2011): Flat‐panel computed tomography versus multislice computed tomography to evaluate cochlear implant positioning. Cochlear Implants Int 12:216–222. [DOI] [PubMed] [Google Scholar]
