Skip to main content
Social Cognitive and Affective Neuroscience logoLink to Social Cognitive and Affective Neuroscience
. 2013 Sep 11;9(9):1395–1403. doi: 10.1093/scan/nst124

Perception of affective and linguistic prosody: an ALE meta-analysis of neuroimaging studies

Michel Belyk 1,, Steven Brown 1
PMCID: PMC4158374  PMID: 23934416

Abstract

Prosody refers to the melodic and rhythmic aspects of speech. Two forms of prosody are typically distinguished: ‘affective prosody’ refers to the expression of emotion in speech, whereas ‘linguistic prosody’ relates to the intonation of sentences, including the specification of focus within sentences and stress within polysyllabic words. While these two processes are united by their use of vocal pitch modulation, they are functionally distinct. In order to examine the localization and lateralization of speech prosody in the brain, we performed two voxel-based meta-analyses of neuroimaging studies of the perception of affective and linguistic prosody. There was substantial sharing of brain activations between analyses, particularly in right-hemisphere auditory areas. However, a major point of divergence was observed in the inferior frontal gyrus: affective prosody was more likely to activate Brodmann area 47, while linguistic prosody was more likely to activate the ventral part of area 44.

Keywords: affective prosody; linguistic prosody; speech; emotion; ALE meta-analysis; brain imaging, inferior frontal gyrus

INTRODUCTION

Prosody comes from the Greek prosodia, meaning ‘sung to music’ (Pearsall et al., 2005). Speech prosody therefore refers to the song-like vocal modulations that accompany speech. For this reason, it is often considered to be ‘the music of speech’ (Wennerstrom, 2001). The pitch modulations associated with speech prosody convey two broad categories of information. On the one hand, pitch modulations convey information about a speaker’s emotional state (Fairbanks and Pronovost, 1938), what has been referred to as ‘emotional’ or ‘affective’ prosody (Monrad-Krohn, 1947). On the other hand, they provide cues regarding syntax and pragmatics (Beach, 1991), what has been referred to as ‘intrinsic’ or ‘linguistic’ prosody (Monrad-Krohn, 1947). While these two types of prosody are functionally distinct, they rely on a common set of acoustic cues related to pitch, loudness, tempo and voice quality (Fonagy, 1978; Juslin and Laukka, 2003). The sharing of acoustic parameters by these two processes suggests that they might rely on a common system for the perception of pitch but that this pitch information may be fed into distinct systems for processing either emotion (affective prosody) or syntax/pragmatics (linguistic prosody).

Affective prosody conveys a speaker’s emotional state largely through global changes in pitch height and loudness, although other acoustic features also serve to disambiguate emotional states (Banse and Sherer, 1996). Emotional expressions can take the form of ‘affect bursts’ (Schröder, 2003) that have emotional but not semantic meaning (e.g. ‘Yuck!’) or can occur concurrently with normal speech. Affective prosody conveys a broad range of emotional states (Sauter and Scott, 2007) that can be recognized across cultures without prior experience (Sauter et al., 2010; Scherer et al., 2001), much like facial expressions (Ekman et al.,1969).

Linguistic prosody uses local increases in pitch height and/or loudness to signal features like word stress (e.g. CONtent vs conTENT; Gay, 1978), sentence focus (e.g. two WHITE shirts vs TWO white shirts; Ladd and Morton, 1997), segmentation of the speech stream into phrases (Jusczyk et al., 1992), broad pragmatic categories of utterances (modality), such as declarative vs interrogative sentences (Xu and Xu, 2005), and the standard intonational melodies that are used as part of mother-infant communication (Fernald, 1992) as well as communication between adults. The conventions of linguistic prosody vary across languages and are important contributors to the melody and rhythm of speech. It is for this reason that deviations from standard prosody contribute to the impression of a foreign accent (de Mareüil and Vieru-Dimulescu, 2006).

Early investigations into the neural basis of speech prosody analyzed neurological cases of patients suffering from strokes. These studies focused overwhelmingly on the lateralization of prosody, especially compared to the well-accepted left-hemisphere dominance for the lexicosyntactic aspect of language. A major finding of these early studies was that the perception of affective prosody was impaired in patients with unilateral right-hemisphere lesions (Ross, 1981; Gorelick and Ross, 1987). However, these studies did not examine patients with left-hemisphere lesions, and studies that have since done so have reported deficits in patients with both types of unilateral lesions (Trauner et al., 1996; Pell, 1998). Similarly, deficits in the perception of linguistic prosody have been reported in patients with lesions in both the left (Pell and Baum, 1997) and right (Weintraub et al., 1981) hemispheres. A meta-analysis of this literature revealed that both affective and linguistic prosody are impaired by damage to either hemisphere, although damage to the right hemisphere tends to have a larger impact on affective prosody and the left hemisphere on linguistic prosody (Witteman et al., 2011).

Neurological studies have generally been conducted with patients having a diverse set of lesions and have seldom reported the location of lesions beyond the level of the hemisphere or lobe. Therefore, the neurological literature does not permit an examination of localization hypotheses at a finer scale than the lobe. Interestingly, transcranial magnetic stimulation of healthy individuals can induce deficits in the perception of affective prosody when applied to either the left or right inferior frontal gyrus (IFG; Hoekert et al., 2010).

Neuroimaging studies have been similarly inconclusive with respect to the hemispheric lateralization of prosody perception. The literature has variably reported unilateral or bilateral activations for affective prosody (Wildgruber et al., 2005; Bach et al., 2008; Ethofer et al., 2009) and linguistic prosody (Meyer et al., 2002; Strelnikov et al., 2006). Despite these inconsistencies in lateralization, neuroimaging studies have contributed to the broader localizationist account of prosody perception.

Neural models of affective-prosody perception (Ethofer et al., 2006; Schirmer and Kotz, 2006) suggest that low-level acoustic analyses are performed in the posterior superior temporal gyrus (STG)—in what has been called the ‘emotional voice area’ (Ethofer et al., 2012)—and the superior temporal sulcus (STS). Similarly, more recent models suggest that acoustic processing is performed in the middle part of the STS (mSTS; Belin et al., 2000), that identification of vocally expressed emotions is performed in either the anterior (Kotz and Paulmann, 2011) or posterior (Brück et al., 2011a) STG/STS, and that explicit evaluation of vocally expressed emotions is performed by inferior frontal regions (Wildgruber et al., 2009). Passive perception of prosody reliably activates the STG (Humphries et al., 2005; Dietrich et al., 2008). Posterior temporal areas are proposed to project to inferior frontal regions for explicit evaluation of emotional meaning when such evaluation is task-relevant. While studies of both affective and linguistic prosody routinely report activations in Broca’s area (Gandour et al., 2003a,b), Schirmer and Kotz (2006) proposed that a region anteroventral to Broca’s area—the IFG pars orbitalis (Brodmann area [BA] 47)—may be specifically involved in the perception of affective prosody. A meta-analysis of the imaging literature on the perception of affective prosody supports the involvement of the IFG pars orbitalis when attention is directed towards affective prosody rather than away from it and the IFG pars triangularis (BA 45) whether or not attention is directed towards affective prosody (Witteman et al., 2012).

The perception of prosody stimulates additional regions beyond the superior temporal and inferior frontal gyri (Buchanan et al., 2000; Brück et al., 2011a). Studies of affective and linguistic prosody routinely report activations in speech-related areas—even when contrasted with other speech-perception tasks—including the anterior cingulate cortex (ACC; Doherty et al., 2004; Frühholz et al., 2011), inferior parietal lobule (IPL; Gandour et al., 2003a; Johnstone et al., 2006), anterior insula (Meyer et al., 2002; Ethofer et al., 2009) and basal ganglia (Meyer et al., 2004; Bach et al., 2008).

Given the inconsistencies in both the neurological and neuroimaging literatures, we sought to clarify the localization of prosody perception in the brain by performing a statistical meta-analysis of published neuroimaging studies of affective and linguistic prosody either separately, in contrast, or in conjunction using the ‘activation likelihood estimation’ (ALE) method (Turkeltaub et al., 2002; Eickhoff et al., 2012). The goal was to assess whether these two functions are mediated by shared or distinct brain networks. The major predictions were that these functions should show commonalities in posterior temporal areas that process the acoustic features of vocal pitch, but that differences should be seen in higher-level areas in the frontal lobe that generate distinct interpretations of these pitch modulations.

METHODS

Inclusion criteria

A meta-analysis of published neuroimaging studies of affective and linguistic prosody was performed using ALE meta-analysis (Turkeltaub et al., 2002) in order to compare areas of brain activation across these functions. Published articles were retrieved in February 2012 by searches in the Web of Knowledge database using the search terms ‘prosody + fMRI’ and ‘prosody + PET’. The reference sections of resultant studies were searched for additional studies. Experiments in which subjects made emotional judgments were classified as ‘affective prosody’, while studies in which subjects made judgments based on word stress, focus, syntax or modality were classified as ‘linguistic prosody’.

Our inclusion criteria for the studies were: (i) that brain scanning was performed using either functional magnetic resonance imaging (fMRI) or positron emission tomography (PET); (ii) that papers reported activation foci in the form of standardized stereotaxic coordinates in either Talairach or Montreal Neurological Institute (MNI) space; (iii) that subjects were healthy adults (thereby excluding results from clinical populations); (iv) that subjects made active judgments about the affective or linguistic prosody of auditorily presented speech stimuli; (v) that the analyses included a high-level contrast against a suitable control condition so as to remove the influence of low-level phonological processing (e.g. passive listening or gender discrimination) and (vi) that results from the entire scanned volume were reported (thereby excluding studies reporting region-of-interest analyses only). Due to the large number of studies with only partial brain coverage, we performed a separate analysis with the additional criterion (vii) that the entire brain-volume was imaged (thereby excluding studies with an insufficient field of view to encompass the whole brain). This criterion is discussed further in the ‘Brain coverage’ section.

Our searches yielded 29 independent experiments conducted in German, English, French, Mandarin, Japanese and Russian (see Supplementary Tables 1 and 2 for details). Wherever studies reported multiple experiments from the same group of subjects, the contrasts were included together as a single study. Similarly, for studies that reported the results of more than one subject-group, each group was treated separately, in accordance with the approach of Turkeltaub et al. (2011). Separate analyses were conducted for affective prosody (n = 19 experiments) and linguistic prosody (n = 10). GingerALE 2.1 was used for all analyses and to convert MNI coordinates to Talairach coordinates. The ALE results were registered onto a Talairach-normalized template brain using Mango (ric.uthscsa.edu/mango). All analyses were corrected for multiple comparisons using the False Discovery Rate P < 0.05 and cluster threshold k = 10.

Brain coverage

ALE meta-analysis is an empirical technique for the analysis of brain imaging studies (Turkeltaub et al., 2002). Each focus of activation is modeled as a three-dimensional Gaussian probability distribution whose width is determined by the size of the subject-group so as to reflect increasing uncertainty with decreasing sample size (Eickhoff et al., 2009). Maps of activation likelihoods are created for each study by taking the maximum probability of activation at each voxel. A random-effects analysis tests for the convergence of activations across studies against a null hypothesis of spatially independent brain activations.

Due to the limited brain coverage of many of the studies included in our dataset, we modified the standard ALE method in order to test the null hypothesis of spatially independent brain activations within the brain volume that was imaged in all of the included studies. Standard ALE analyses mask the brain volume to gray matter. Activation foci are unlikely to originate from ventricles or white matter. Therefore, in order to avoid skewing the empirical null distribution—and overestimating any effects—this portion of brain space must be excluded (Eickhoff et al., 2009). Similarly, activation foci cannot originate from outside the field of view for a given study, and so this region must therefore be excluded from the analysis. We therefore further restricted the analyses to the portion of the brain-volume that was imaged in all studies meeting our inclusion criteria. This area extended from z = −6 to z = 38 in Talairach space (see horizontal red lines in Figure 1). A second set of whole-brain analyses was performed to assess convergence beyond the restricted volume of coverage. Whole-brain analyses only included studies that imaged the entire brain volume. This additional inclusion criterion reduced the number of studies to 10 and 4, respectively, for affective and linguistic prosody. For all figures and tables, all 29 experiments contributed to analyses within the restricted range. Only those experiments with full brain coverage contributed to analyses outside this range.

Fig. 1.

Fig. 1

Sagittal sections showing major foci for the individual ALE meta-analyses for affective prosody (red) and linguistic prosody (green) as well as the statistical conjunction of the two (blue). These slices demonstrate the bilateral involvement of inferior frontal regions for affective and linguistic prosody perception. The figure also demonstrates the clear segregation of functions within the inferior frontal gyrus as well as sharing in the right auditory cortex. Red lines demarcate the limits of the ‘restricted’ analysis (z = −6 to z = 38): foci within the red lines were generated by the restricted analysis (which included all studies), while foci outside the red lines were generated by the whole-brain analysis (including only those studies that reported whole-brain coverage).

Conjunctions and contrasts

In addition to separate analyses, we performed a statistical conjunction (Nichols et al., 2005) of the meta-analyses in order to determine which areas, if any, were common to affective and linguistic prosody. Direct contrasts were performed to determine which areas were specific to each of these two functions. Because there were many more studies of affective prosody than linguistic prosody in the dataset—which may bias the results—we also report the number and percentage of studies of affective prosody and linguistic prosody that contribute to each of the ALE foci. Due to the small number of studies covering the whole brain, direct contrasts are reported for the restricted analysis only.

Post hoc analysis of working memory demands based on task-type

The studies included in the meta-analyses used tasks that fall into two broad classes: identification tasks and same/different tasks. Subjects performing an identification task are presented with an auditory stimulus and are required to identify—from a limited set of possible responses—which emotion or intonation is being presented. Subjects performing a same/different task are presented with pairs of stimuli and are required to indicate whether the same emotion or intonation occurs in both presentations. To the extent that the latter task requires subjects to maintain a representation of the first stimulus-presentation long enough to perform a comparison with the second, it may impose greater demands on working memory than an identification task. Among the studies included in our meta-analyses, affective prosody experiments were much more likely to use identification tasks or similar tasks with a low working memory load (16 out of 19), while studies of linguistic prosody were more evenly divided (four and six low and high working memory load, respectively). We therefore compared experiments of linguistic prosody containing putatively low vs high working memory load as, estimated from task demands, in order to account for areas of convergence that may be more reflective of working memory demands than prosody perception per se.

RESULTS

We performed individual ALE analyses of affective and linguistic prosody. Due to the preponderance of studies with functional coverage limited to the perisylvian region alone, we performed two parallel analyses for each function, one restricted to the volume covered by all studies in the dataset (in order to avoid violating the assumptions of the ALE method) and a second, whole-brain analysis exclusively for those studies that reported whole-brain coverage. Results from both the restricted and whole-brain analyses are combined in all figures and tables. Figure 1 presents the location of the major ALE foci for each analysis, and Table 1 provides the Talairach coordinates and cluster sizes for each ALE focus. Results will first be presented for analyses of each function separately, followed by a conjunction of analyses to identify shared regions, and finally direct contrasts to identify regions specific to each function.

Table 1.

Talairach coordinates of likelihood maxima and cluster sizes for individual ALE analyses of affective prosody and linguistic prosody perception, respectively

Affective prosody
Linguistic prosody
Brain region x y z ALE (×103) x y z ALE (×103)
Right hemisphere Frontal lobe
    IFG pars triangularis (BA 45) 46 22 16 21.04
    Insula (BA 13) 54 −36 20 19.41
42 8 12 13.76
    IFG pars orbitalis (BA 47) 48 14 0 18.49 46 20 2 8.61
38 26 0 15.87
    Middle frontal gyrus (BA 9) 48 16 28 11.34 46 14 30 19.94
    SMA (BA 6) 8 18 50 9.04 4 16 48 14.69
8 26 42 6.98
    Middle frontal gyrus (BA 10) 34 36 10 15.34
    IFG pars opercularis (BA 44) 54 8 6 11.49 48 8 15 1.30
Temporal lobe
    aSTG (BA 22) 54 0 4 13.77
    pSTG (BA 22) 48 −24 4 20.68 46 −24 0 12.25
46 −32 4 20.60
56 −44 4 11.98
    Heschl’s gyrus (BA 41) 48 −32 8 11.71
    PHG (BA 28) 16 −10 −12 11.41
    MTG (BA 21) 44 −4 −16 8.95
    Subcallosal gyrus (BA 34) 26 6 −10 8.08
Parietal lobe
    SMG (BA 40/7) 36 −54 46 10.91 36 −58 48 15.74
Subcortical
    Claustrum 26 16 4 13.57
    Cerebellum 18 −64 −16 9.42 2 −70 −10 6.75
Left hemisphere Frontal lobe
    IFG pars orbitalis (BA 47) −40 22 −2 19.34
−44 34 −2 14.64
−50 20 0 13.57
    IFG pars triangularis (BA 45) −46 22 12 13.62
    Anterior insula (BA 13) −32 22 2 16.75 −32 18 6 9.48
    Middle frontal gyrus (BA 9) −40 6 34 16.84
    IFG pars opercularis (BA 44) −42 2 6 12.70 −44 14 10 13.45
Temporal lobe
    aSTG (BA 22) −50 10 2 13.32
    pSTG (BA 22) −48 −46 12 11.73
Parietal lobe
    SMG (BA 40) −30 −50 38 10.87 −30 −50 40 12.35
−36 −44 38 11.86
    SMG (BA 40/7) −46 −56 42 7.70
Subcortical
    Amygdala −18 −6 −12 21.72
    Caudate nucleus −12 −4 14 13.04 −16 16 8 11.23
    Putamen −22 14 −12 8.22
    Cerebellum −28 −60 −21 6.45
    Cerebellum −6 −74 −18 6.75
    Thalamus −8 −6 10 12.60
Midline Frintal lobe
    SMA (BA 6) 0 14 48 11.99
Occipital lobe
    Cuneus (BA 17) 0 −82 8 16.58

Results from both the restricted and whole-brain analysis are combined. IFG: inferior frontal gyrus; MOG: middle occipital gyrus; MTG: middle temporal gyrus; PHG: parahippocampal gyrus; SMA: supplementary motor area; SMG: supramarginal gyrus.

Affective prosody activated both audio-vocal and limbic areas. Audio-vocal activations were observed in right pSTG, bilateral aSTG, supplementary motor area (SMA), IFG pars opercularis (BA 44), pars triangularis (BA 45) and supramarginal gyrus, right middle temporal gyrus (MTG), cerebellum, and middle frontal gyrus (BA 9, BA 10), left caudate nucleus and thalamus. Presumed emotion-related activations were observed in limbic areas, including bilateral IFG pars orbitalis (BA 47), left amygdala, ventral anterior insula and ventral putamen, right parahippocampal gyrus (BA 28) and subcallosal gyrus (BA 34). Importantly, the frontal language areas that are most widely discussed in this literature showed bilateral activity.

In contrast to this limbic profile for affective prosody, linguistic prosody showed ALE foci primarily in speech and language areas. These included bilateral IFG pars opercularis (BA 44), pSTG, supramarginal gyrus (BA 40), middle frontal gyrus, right SMA, IFG pars orbitalis, primary auditory cortex (BA 41) and the left caudate nucleus. Nonlanguage-related foci were observed in the bilateral insula and cerebellum as well as in the right claustrum and primary visual cortex. As with affective prosody, the ALE foci in frontal perisylvian language areas were present bilaterally.

Next, we compared the functions using conjunctions so as to identify areas of overlap vs areas of function-specificity (Figure 1 and Table 2). Conjunction analyses demonstrated that affective prosody shared common areas with linguistic prosody. As predicted, affective and linguistic prosody showed overlapping activations in the right STG (BA 22). Other areas of overlap included the bilateral supramarginal gyrus, right middle frontal gyrus, left insula and midline SMA.

Table 2.

Statistical conjunction demonstrates areas of commonality between affective prosody and linguistic prosody

LP U AP x y z Size (mm3)
Right hemisphere
    Superior temporal gyrus (BA 22) 46 −24 0 799
    Supramrginal gyrus (BA 40/7) 36 −54 46 437
    Middle frontal gyrus (BA 9) 48 16 30 115
Left hemisphere
    Supramrginal gyrus (BA 40) −30 −50 38 27
    Anterior insula (BA 13) −30 20 4 81
Midline
    Supplementary motor area (BA 6) 0 16 48 669

In order to identify regions that were specific to each condition, we performed direct contrasts (see the right panel of Figure 2 and Table 3). Affective prosody had a stronger association with activation in the left IFG pars orbitalis (BA 47) and thalamus as well as right pSTG (BA 22) and MTG (BA 21). Linguistic prosody had a stronger association with activation in the left pSTG, bilateral middle frontal gyrus (BA 9), bilateral IFG pars opercularis (BA 44), right supramarginal gyrus (BA 40), claustrum and midline primary visual cortex.

Fig. 2.

Fig. 2

The left panel shows the ALE foci for affective prosody (AP, red) and linguistic prosody (LP, green) registered onto axial sections. The right panel shows two direct contrasts, and highlights areas unique to each function. Affective prosody is uniquely associated with the IFG pars orbitalis (BA 47), while linguistic prosody is uniquely associated with the ventral IFG pars opercularis (BA 44).

Table 3.

Pairwise contrasts demonstrate areas of activation unique to affective prosody and linguistic prosody

AP > LP x y z Size (mm3) AP studies LP studies
Right hemisphere
    Superior temporal sulcus (BA 22)* 44 −40 0 75 9 (47%) 5 (50%)
    Middle temporal gyrus (BA 21) 44 −44 2 27 5 (26%) 1 (10%)
Left hemisphere
    IFG pars orbitalis (BA 47) −40 28 −6 1063 9 (47% 1 (10%)
    Thalamus −12 −6 16 197 3 (16%) 0 (0%)
LP > AP x y z Size (mm3) AP studies LP studies
Right hemisphere
    Middle frontal gyrus (BA9) 50 8 30 2773 4 (21%) 5 (50%)
    Angular gyrus (BA 40) 50 −32 20 1437 4 (21%) 5 (50%)
    Claustrum** 28 20 6 397 0 (0%) 0 (0%)
    IFG pars opercularis (BA 44) 46 8 16 111 3 (16%) 9 (90%)
Left hemisphere
    Middle frontal gyrus (BA 9) −42 8 30 321 4 (21%) 3 (30%)
    IFG pars opercularis (BA 44) −40 14 8 129 4 (21%) 4 (40%)
    Superior temporal gyrus (22) −50 −42 10 45 4 (21%) 4 (40%)
Midline
    Cuneus (BA 17) 0 −76 10 1349 4 (21%) 3 (30%)

*Peaks from either condition may be differentially localized as suggested by the left panel of Figure 1. **May be mislocalized from nearby Putamen and/or Insula. The number of studies of affective and linguistic prosody contributing to each locus corroborates the ALE results. AP: affective prosody; LP: linguistic prosody.

As an additional analysis, we divided the studies of linguistic-prosody perception into those with putatively high vs low verbal working-memory load, as estimated by task demands (Table 4). Higher working memory load was associated with increased activation in the bilateral middle frontal gyrus (part of the dorsolateral prefrontal cortex) and right STG. Given that the literatures under review here were not orthogonal with respect to working memory demands, differences between individual ALE analyses in the right STG and middle frontal gyri should be interpreted with caution.

Table 4.

Meta-analysis of verbal working memory

High > Low Verbal working memory x y z ALE (103) Same/Different Identification
Right hemisphere
    Superior temporal gyrus (BA 22) 48 −30 12 7.80 5 (83%) 0 (0%)
    Middle frontal gyrus (BA 9) 52 12 32 6.98 4 (67%) 1 (25%)
Left hemisphere
    Middle frontal gyrus (BA 9) −42 14 32 10.96 2 (33%) 0 (0%)

Studies of linguistic prosody were divided into those with high vs low working-memory demands based on task type. The bilateral middle frontal gyrus and right STG are more likely to be reported in studies with high verbal working-memory demands. Contrasts between individual ALE meta-analyses must be interpreted cautiously to avoid falsely attributing foci in these areas to prosody perception. The number of studies with low and high working memory loads contributing to each locus corroborates the ALE results.

DISCUSSION

The primary objective of this study was to use meta-analytic techniques to help clarify which brain regions are reported consistently in studies of affective and linguistic prosody perception in light of inconsistency and confusion in both the neurological and neuroimaging literatures. We examined the functional neuroimaging literatures related to affective and linguistic prosody individually and then jointly using conjunction and contrast methods. The results revealed both shared and distinct components of the networks involved in these processes, reflecting both the perception of vocal-pitch modulation and its functional interpretation.

Our strongest prediction of overlap between the two functions was for auditory association areas in the pSTG. Interestingly, the right pSTG has been dubbed the ‘emotional voice area’ by researchers of affective prosody (Ethofer et al., 2012). In confirmation of this area’s role in emotional voice perception, we observed convergence centered in right Heschl’s gyrus and extending into the pSTG for studies of affective prosody. However, we observed a similar area of convergence bilaterally for linguistic prosody as well as for verbal working memory, and the pSTG is commonly reported in studies of music perception as well (Zatorre et al., 1994; Brown et al., 2004). Indeed, Wiethoff et al. (2008) observed that activation in this region could be explained entirely by the acoustic parameters of the stimuli. The pSTG appears to respond to a variety of types of auditory stimuli and may not be specific to emotional voices. Emotional voices may simply contain a larger degree of pitch modulation than the neutral voices that are typically used as baseline stimuli in many studies of affective prosody.

In addition to demonstrating overlap in right auditory areas, the conjunction analysis revealed convergence across functions in the SMA, a motor structure involved in speech production. Electrical stimulation of the anterior portion of the left (but not right) SMA elicits vocalization (Fried et al., 1991). Lesions to this area can cause aphasic symptoms (Fontaine et al., 2002) and akinetic mutism (Nagaratnam et al., 2004). The individual ALE meta-analyses of affective prosody and linguistic prosody suggested a role of the right SMA in those functions as well. Surprisingly, convergence between studies was not observed in the ACC just ventral to the SMA. Animal models of vocal control demonstrate an important role of the ACC in top-down control of the vocal-motor nuclei in the brain stem (Jürgens, 2002), and ALE meta-analysis of human neuroimaging studies of vocalization show foci in this region for both spoken and sung utterances (Brown et al., 2009). Indeed, several of the studies included in these meta-analyses reported activations in the ACC (Gandour et al., 1998; Doherty et al., 2004; Bach et al., 2008), and yet ALE foci in this region did not reach significance for any analysis. One likely explanation is that much of the ACC lies outside the volume covered in our ‘restricted’ analyses and that our whole-brain analyses had too little power to detect convergence in this area. In addition, the anatomy of the ACC is variable across individuals (Paus et al., 1996), and it is therefore possible that differences between subjects in cingulate anatomy resulted in subtle variability in the localization of foci between studies.

Unlike the result in auditory areas, substantial divergence was observed in inferior frontal regions. More specifically, affective prosody activated the IFG pars orbitalis (BA 47) bilaterally while linguistic prosody activated the IFG pars opercularis (ventral BA 44) bilaterally. The absence of overlap in the inferior frontal region suggests that activations here do not simply relate to some aspect of pitch processing alone, but may instead reflect the different classes of information that listeners extract from affective vs linguistic cues in speech prosody. Notably, Wildgruber et al. (2004) compared affective and linguistic prosody perception directly in an fMRI study. These authors observed a similar localization for both functions in inferior frontal regions.

We observed a small number of areas that were uniquely associated with each function of interest. The IFG pars orbitalis (BA 47), which was associated most strongly with affective prosody, is distinct from adjacent Broca’s area in both cytoarchitecture (Brodmann, 1909/1994) and structural connectivity. The homologous region in macaques, area 47/12, receives projections from both limbic regions and the homologue of Broca’s area (Petrides and Pandya, 2001). Area 47/12 is part of an orbitofrontal network that receives input from sensory areas, including auditory, visual, somatosensory, olfactory, visceral and gustatory cortices, as well as limbic areas such as the amygdala, subiculum, entorhinal cortex and perirhinal cortex. This same network projects to the hypothalamus and periaquaductal gray by way of the ventromedial prefrontal cortex (Price, 1999). Diffusion tensor imaging in humans reveals a similar pattern. The IFG pars orbitalis is connected to auditory and visual areas via the inferior occipitofrontal fasciculus and middle longitudinal fasciculus (Turken and Dronkers, 2011). The frontal operculum adjacent to BA 47 is connected to the amygdala and septal region (Anwander et al., 2007) and plays a role in emotion regulation in conjunction with the amygdalae and nucleus accumbens (Wager et al., 2008). Patients with lesions in this region and the adjacent orbitofrontal cortex have deficits in recognizing emotions in others as well as changes in behavior and subjective emotional experience (Hornak et al., 1996). This region is consistently active when subjects experience particular emotions or when they perceive emotions in either the auditory or visual domain (Lindquist et al., 2012). The IFG pars orbitalis may therefore be well situated to act as an interface between limbic and sensorimotor networks, as would be necessary for affective prosody perception. Indeed, given the diverse sensory information available to this region, it is not surprising that it is involved in the perception of emotional faces and gestures as well (Sprengelmeyer et al., 1998; Lotze et al., 2006).

Linguistic prosody, in contrast, was associated most strongly with the IFG pars opercularis (BA 44). Ventral BA 44 is associated with lexicosyntactic function. Functional MRI studies have shown that syntactic processing activates ventral IFG pars opercularis (Friederici et al., 2000; Heim et al., 2003a). In contrast, other linguistic functions, such as phonological processing, activate a locus in dorsal BA 44 (Heim et al., 2003b; Papoutsi et al., 2009). Given that linguistic prosody plays a role in syntactic disambiguation (Beach, 1991), it is perhaps not surprising that this suprasegmental element of speech shares brain areas with syntactic processing.

Lateralization vs localization

Both the neurological and neuroimaging literatures on the perception of prosody are concerned primarily with the lateralization of function in temporal and frontal language areas. In agreement with this literature, we observed consistent right-hemisphere lateralization in temporal-lobe auditory areas. Importantly, we observed this pattern of lateralization for both affective and linguistic prosodies, constituting a region of overlap between these functions. In contrast, our results did not support a consistent lateralization in the frontal lobe for either affective or linguistic prosody. Our meta-analyses instead demonstrated that bilateral inferior frontal activations were likely to be reported by neuroimaging studies of both functions, although in nonoverlapping regions. While direct contrasts between conditions appeared to support the lateralization of affective prosody to the left IFG, our primary analyses demonstrated that affective prosody perception did in fact activate right inferior frontal regions as well. Both affective and linguistic prosody activated bilateral (although distinct) inferior frontal regions, as demonstrated by Figure 1. However, this does not preclude the interpretation that some functional aspect of the task may influence patterns of lateralization. It has been proposed that one contributor to the frequent, but inconsistent, lateralization of speech prosody, especially in temporal-lobe auditory areas, is that the window of temporal integration of pitch information differs between the two hemispheres (Buchanan et al., 2000) such that the left hemisphere processes relatively fast frequency modulations and the right hemisphere relatively slow modulations (Zatorre, 2001).

Task-type

The middle frontal gyrus (part of the dorsolateral prefrontal cortex) and STG were associated with task-related differences in working-memory load in the linguistic-prosody meta-analysis. Activations in this region were more prominent in studies of linguistic prosody than affective prosody perception. This may be due to a greater proclivity towards experiments with high verbal working-memory demands in that literature. Studies of linguistic prosody used methods with either a high working memory load, namely same/different tasks, or with a low working memory load, namely forced-choice identification tasks. In comparison, studies of affective prosody used primarily tasks with low verbal working-memory demands. This methodological difference might account for the increased likelihood of observing activation in the middle frontal gyrus for linguistic prosody compared with affective prosody. This finding is corroborated by a meta-analysis that explicitly examined verbal working-memory demands (Chein et al., 2002). Note that this analysis was conducted to detect confounds in our primary analyses and should not be taken as an analysis of working memory per se.

Production and perception

To the best of our knowledge, only one study has compared functional activations between perception and production of prosody, and it did so for both linguistic and affective prosody (Aziz-Zadeh et al., 2010). While that study did not observe activation in the IFG pars orbitalis that we described for affective prosody, it did observe activation in the left IFG pars opercularis for the production and perception of both affective and linguistic prosodies. Our meta-analyses revealed ALE foci in this region, although the localization varied for each function. The IFG pars opercularis may be an important point of interaction for affective and linguistic prosody. More specifically, the IFG pars opecularis is purported to be a ‘mirror neuron’ area involved in both the production and perception of actions (Aziz-Zadeh et al., 2006). This area that may be structurally connected with the primary motor cortex (Greenlee et al., 2004; Simonyan et al., 2009). It may therefore constitute an area of convergence for affective and linguistic prosody en route to the motor cortex.

Prosody networks

A number of models have proposed temporo-frontal networks for prosody perception based on the activation patterns for affective prosody (Ethofer et al., 2006; Schirmer and Kotz, 2006). Two recent models have suggested that prosody perception occurs in three stages: (i) acoustic analysis in the voice-selective areas of the mSTS (Belin et al., 2000), (ii) identification of vocally expressed emotion in the aSTG (Kotz and Paulmann, 2011) or pSTG (Brück et al., 2011b) and (iii) explicit evaluation of prosody in the IFG. These models agree with one another in most respects, with the exception of the localization of temporal regions specific for affective voices. While another prosody meta-analysis (Witteman et al., 2012) supported the localization of Brück et al. (2011b) to the pSTG, our results supported the role of both aSTG and pSTG in affective prosody processing. Notably, we observed ALE foci in bilateral aSTG for affective prosody only, not for linguistic prosody. However these foci did not survive a direct contrast between the two functions. Due to the lower power of the linguistic-prosody analysis relative to affective prosody, it cannot be concluded from the data that either the aSTG or pSTG is specific to affective prosody.

Conjunction analysis revealed several areas of common activation between affective and linguistic prosody. Among these were the right auditory association cortex, which is specialized for the fine-grained analysis of pitch (Zatorre and Gandour, 2008), left anterior insula, which is anatomically connected to the entire extent of the IFG (spanning the pars opercularis, pars triangularis and pars orbitalis; Catani et al., 2012), and the somatotopic oro-laryngeal portion of the SMA (Fried et al., 1991). This group of regions is likely involved in audio-vocal functioning generally, rather than prosody specifically.

Affective and linguistic prosody do not generally occur in isolation but rather in parallel with speech. A focus of future research should be to further develop network models of prosody perception and to extend these models to incorporate production with the aim of integrating these networks with extant models of speech. For example, the ‘Directions into Velocities and Articulators’ (DIVA) model (Golfinopoulos et al., 2010) is a well-established model of speech production that describes how intended speech sounds are converted into articulatory movements that ultimately result in the production of speech. Such a set of mechanisms should, in theory, accommodate the production of the pitch-based cues that are used for affective and linguistic prosodies.

The IFG pars opercularis locus observed for linguistic prosody is part of Broca’s area (and Broca’s homolog) and is therefore already a component of most neural models of speech. However, the expression of emotion is acoustically similar whether it occurs without language in the form of affect bursts such as laughter and crying (Schröder, 2003) or with language in the form of affective prosody (Banse and Sherer, 1996). Affective prosody may therefore require the integration of an evolutionarily ancestral subcortical system for affective communication found in monkeys (Jürgens, 2009) with the evolutionarily recent cortical system for speech and language that is found only in humans. We suggest that the IFG pars orbitalis (BA 47) may function as such an interface between emotion and vocalization, although others have proposed that the ACC serves this function as well (Jürgens, 2009). One caveat to this proposal is the suggested role of the IFG pars orbitalis in other functions. This region has previously been reported in neuroimaging studies of both linguistic (Fiez, 1997) and musical semantics (Levitin and Menon, 2003) as well as in pitch memory (Zatorre et al., 1994). Price (1999) noted that the orbital region of the macaque, including BA 47/12, is cytoarchitectonically diverse. Further research is needed to search for potential functional subdivisions within this region.

LIMITATIONS

A potential limitation of our analysis is that our dataset included more studies of affective prosody than linguistic prosody. This unbalanced design may have introduced some bias into the data and limited the inferences that could be made from it. We attempted to mitigate this limitation by checking the number of studies that contribute to each of the foci in our contrasts.

Our analysis of working memory load relied on a small and unbalanced sample of studies of linguistic prosody perception. Furthermore, our division into high and low working memory load was confounded with the distinction between task-driven effects and stimulus-driven effects discussed by Witteman et al. (2012). For these reasons, we stress that our working memory results are provisional and are intended only to aid in the interpretation of the other analyses.

CONCLUSION

We meta-analyzed the literatures on the neural correlates of two pitch-based paralinguistic functions. The results provide mixed support for hemispheric lateralization of speech prosody, with greater lateralization seen in temporal-lobe auditory areas than in frontal-lobe evaluative areas. Instead, the results support a localizationist account based on differentiation of the two prosodic functions in the IFG. Linguistic prosody is associated with a portion of the IFG pars opercularis that is involved in syntactic processing. Affective prosody is associated with the IFG pars orbitalis, which is connected with both limbic and speech-motor areas, making it a good candidate as an interface between emotion and voice.

SUPPLEMENTARY DATA

Supplementary data are available at SCAN online.

Conflict of Interest

None declared.

Supplementary Material

Supplementary Data

Acknowledgments

This work was funded by a grant from the Natural Sciences and Engineering Research Council (NSERC) of Canada to S.B.

REFERENCES

  1. Anwander A, Tittgemeyer M, von Cramon DY, Friederici AD, Knösche TR. Connectivity-based parcellation of Broca’s area. Cerebral Cortex. 2007;17(4):816–25. doi: 10.1093/cercor/bhk034. [DOI] [PubMed] [Google Scholar]
  2. Aziz-Zadeh L, Koski L, Zaidel E, Mazziotta J, Iacoboni M. Lateralization of the human mirror neuron system. Journal of Neuroscience. 2006;26(11):2964–70. doi: 10.1523/JNEUROSCI.2921-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Aziz-Zadeh L, Sheng T, Gheytanchi A. Common premotor regions for the perception and production of prosody and correlations with empathy and prosodic ability. PLoS One. 2010;5(1):e8759. doi: 10.1371/journal.pone.0008759. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bach DR, Grandjean D, Sander D, Herdener M, Strik WK, Seifritz E. The effect of appraisal level on processing of emotional prosody in meaningless speech. Neuroimage. 2008;42(2):919–27. doi: 10.1016/j.neuroimage.2008.05.034. [DOI] [PubMed] [Google Scholar]
  5. Banse R, Sherer K. Acoustic profiles in vocal emotion expression. Journal of Personality and Social Psychology. 1996;70(3):614–36. doi: 10.1037//0022-3514.70.3.614. [DOI] [PubMed] [Google Scholar]
  6. Beach CM. The interpretation of prosodic patterns at points of syntactic structure ambiguity: evidence for cue trading relations. Journal of Memory and Language. 1991;30(6):644–63. [Google Scholar]
  7. Belin P, Zatorre RJ, Lafaille P, Ahad P, Pike B. Voice-selective areas in human auditory cortex. Nature. 2000;6767:309–12. doi: 10.1038/35002078. [DOI] [PubMed] [Google Scholar]
  8. Brodmann K. Description of individual brain maps I. The human brain. In: Garey LJ, editor. Localisation in the Cerebral Cortex. 3rd edn. New York, NY: Springer; 1909/1994. pp. 106–25. [Google Scholar]
  9. Brown S, Laird AR, Pfordresher PQ, Thelen SM, Turkeltaub P, Liotti M. The somatotopy of speech: phonation and articulation in the human motor cortex. Brain and Cognition. 2009;70(1):31–41. doi: 10.1016/j.bandc.2008.12.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Brown S, Martinez M, Parsons L. Passive music listening spontaneously engages limbic and paralimbic systems. Neuroreport. 2004;13:2033–7. doi: 10.1097/00001756-200409150-00008. [DOI] [PubMed] [Google Scholar]
  11. Brück C, Kreifelts B, Kaza E, Lotze M, Wildgruber D. Impact of personality on the cerebral processing of emotional prosody. Neuroimage. 2011a;58(1):259–68. doi: 10.1016/j.neuroimage.2011.06.005. [DOI] [PubMed] [Google Scholar]
  12. Brück C, Kreifelts B, Wildgruber D. Emotional voices in context: a neurobiological model of multimodal affective information processing. Physics of Life Reviews. 2011b;8:383–403. doi: 10.1016/j.plrev.2011.10.002. [DOI] [PubMed] [Google Scholar]
  13. Buchanan TW, Lutz K, Mirzazade S, et al. Recognition of emotional prosody and verbal components of spoken language: an fMRI study. Cognitive Brain Research. 2000;9(3):227–38. doi: 10.1016/s0926-6410(99)00060-9. [DOI] [PubMed] [Google Scholar]
  14. Catani M, Dell’acqua F, Vergani F, et al. Short frontal lobe connections of the human brain. Cortex. 2012;48(2):273–91. doi: 10.1016/j.cortex.2011.12.001. [DOI] [PubMed] [Google Scholar]
  15. Chein JM, Fissell K, Jacobs S, Fiez JA. Functional heterogeneity within Broca’s area during verbal working memory. Physiology and Behaviour. 2002;77:635–39. doi: 10.1016/s0031-9384(02)00899-5. [DOI] [PubMed] [Google Scholar]
  16. de Mareüil PB, Vieru-Dimulescu B. The contribution of prosody to the perception of foreign accent. Phonetica. 2006;63(4):247–67. doi: 10.1159/000097308. [DOI] [PubMed] [Google Scholar]
  17. Dietrich S, Hertrich I, Alter K, Ischebeck A, Ackermann H. Understanding the emotional expression of verbal interjections: a functional MRI study. Neuroreport. 2008;19(18):1751–5. doi: 10.1097/WNR.0b013e3283193e9e. [DOI] [PubMed] [Google Scholar]
  18. Doherty CP, West WC, Dilley LC, Shattuck-Hufnagel S, Caplan D. Question/statement judgments: an fMRI study of intonation processing. Human Brain Mapping. 2004;23(2):85–98. doi: 10.1002/hbm.20042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Eickhoff SB, Bzdok D, Laird AR, Kurth F, Fox PT. Activation likelihood estimation meta-analysis revisited. Neuroimage. 2012;59(3):2349–61. doi: 10.1016/j.neuroimage.2011.09.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Eickhoff SB, Laird AR, Grefkes C, Wang LE, Zilles K, Fox PT. Coordinate-based activation likelihood estimation meta-analysis of neuroimaging data: a random-effects approach based on empirical estimates of spatial uncertainty. Human Brain Mapping. 2009;30(9):2907–26. doi: 10.1002/hbm.20718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Ekman P, Sorenson ER, Friesen WV. Pan-cultural elements in facial displays of emotion. Science. 1969;164(3875):86–8. doi: 10.1126/science.164.3875.86. [DOI] [PubMed] [Google Scholar]
  22. Ethofer T, Anders S, Erb M, et al. Cerebral pathways in processing of affective prosody: a dynamic causal modeling study. Neuroimage. 2006;30(2):580–7. doi: 10.1016/j.neuroimage.2005.09.059. [DOI] [PubMed] [Google Scholar]
  23. Ethofer T, Bretscher J, Gschwind M, Kreifelts B, Wildgruber D, Vuilleumier P. Emotional voice areas: anatomic location, functional properties, and structural connections revealed by combined fMRI/DTI. Cerebral Cortex. 2012;22(1):191–200. doi: 10.1093/cercor/bhr113. [DOI] [PubMed] [Google Scholar]
  24. Ethofer T, Kreifelts B, Wiethoff S, et al. Differential influences of emotion, task, and novelty on brain regions underlying the processing of speech melody. Journal of Cognitive Neuroscience. 2009;21(7):1255–68. doi: 10.1162/jocn.2009.21099. [DOI] [PubMed] [Google Scholar]
  25. Fairbanks G, Provonost W. Vocal pitch during simulated emotion. Science. 1938;88:382–3. doi: 10.1126/science.88.2286.382. [DOI] [PubMed] [Google Scholar]
  26. Fernald A. Meaningful melodies in mothers’ speech to infants. In: Papoušek H, Jürgens U, Papoušek M, editors. Nonverbal Vocal Communication: Comparative and Developmental Approaches. New York, NY: Cambridge University Press; 1992. pp. 262–82. [Google Scholar]
  27. Fiez J. Phonology, semantics, and the role of the left inferior prefrontal cortex. Human Brain Mapping. 1997;5(2):79–83. [PubMed] [Google Scholar]
  28. Fonagy I. A new method of investigating the perception of prosodic features. Language and Speech. 1978;21(1):34–49. doi: 10.1177/002383097802100102. [DOI] [PubMed] [Google Scholar]
  29. Fontaine D, Capelle L, Duffau H. Somatotopy of the supplementary motor area: evidence from correlation of the extent of surgical resection with the clinical patterns of deficit. Neurosurgery. 2002;50(2):297–303. doi: 10.1097/00006123-200202000-00011. [DOI] [PubMed] [Google Scholar]
  30. Fried I, Katz A, McCarthy G, et al. Functional organization of human supplementary motor cortex studied by electrical stimulation. Journal of Neuroscience. 1991;11(11):3656–66. doi: 10.1523/JNEUROSCI.11-11-03656.1991. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Friederici AD, Meyer M, von Cramon DY. Auditory language comprehension: an event-related fMRI study on the processing of syntactic and lexical information. Brain and Language. 2000;75(3):465–77. [PubMed] [Google Scholar]
  32. Frühholz S, Ceravolo L, Grandjean D. Specific brain networks during explicit and implicit decoding of emotional prosody. Cerebral Cortex. 2011;22(5):1107–17. doi: 10.1093/cercor/bhr184. [DOI] [PubMed] [Google Scholar]
  33. Gandour J, Wong D, Hutchins G. Pitch processing in the human brain is influenced by language experience. Neuroreport. 1998;9(9):2115–19. doi: 10.1097/00001756-199806220-00038. [DOI] [PubMed] [Google Scholar]
  34. Gandour J, Dzemidzic M, Wong D, et al. Temporal integration of speech prosody is shaped by language experience: an fMRI study. Brain and Language. 2003a;84(3):318–36. doi: 10.1016/s0093-934x(02)00505-9. [DOI] [PubMed] [Google Scholar]
  35. Gandour J, Wong D, Dzemidzic M, Lowe M, Tong Y, Li X. A cross-linguistic fMRI study of perception of intonation and emotion in Chinese. Human Brain Mapping. 2003b;18(3):149–57. doi: 10.1002/hbm.10088. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Gay T. Physiological and acoustic correlates of perceived stress. Language and Speech. 1978;21(4):347–53. doi: 10.1177/002383097802100409. [DOI] [PubMed] [Google Scholar]
  37. Golfinopoulos E, Tourville J, Guenther F. The integration of large-scale neural networks modeling and functional brain imaging in speech motor control. Neuroimage. 2010;50(3):862–74. doi: 10.1016/j.neuroimage.2009.10.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Gorelick PB, Ross ED. The aprosodias: further functional-anatomical evidence for the organisation of affective language in the right hemisphere. Journal of Neurology and Psychiatry. 1987;50(5):553–60. doi: 10.1136/jnnp.50.5.553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Greenlee JDW, Oya H, Kawasaki H, et al. A functional connection between inferior frontal gyrus and orofacial motor cortex in human. Journal of Neurophysiology. 2004;92(2):1153–64. doi: 10.1152/jn.00609.2003. [DOI] [PubMed] [Google Scholar]
  40. Heim S, Opitz B, Friederici AD. Distributed cortical networks for syntax processing: Broca’s area as the common denominator. Brain and Language. 2003a;85(3):402–8. doi: 10.1016/s0093-934x(03)00068-3. [DOI] [PubMed] [Google Scholar]
  41. Heim S, Opitz B, Müller K, Friederici AD. Phonological processing during language production: fMRI evidence for a shared production-comprehension network. Cognitive Brain Research. 2003b;16(2):285–96. doi: 10.1016/s0926-6410(02)00284-7. [DOI] [PubMed] [Google Scholar]
  42. Hoekert M, Vingerhoets G, Aleman A. Results of a pilot study on the involvement of bilateral inferior frontal gyri in emotional prosody perception: an rTMS study. BMC Neuroscience. 2010;11:1–8. doi: 10.1186/1471-2202-11-93. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Hornak J, Rolls E, Wade D. Face and voice expression identification in patients with emotional and behavioural changes following ventral frontal lobe damage. Neuropsychologia. 1996;34:247–61. doi: 10.1016/0028-3932(95)00106-9. [DOI] [PubMed] [Google Scholar]
  44. Humphries C, Love T, Swinney D, Hickok G. Response of anterior temporal cortex to syntactic and prosodic manipulations during sentence processing. Human Brain Mapping. 2005;26(2):128–38. doi: 10.1002/hbm.20148. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Johnstone T, van Reekum CM, Oakes TR, Davidson RJ. The voice of emotion: an fMRI study of neural responses to angry and happy vocal expressions. Social Cognitive and Affective Neuroscience. 2006;1(3):242–9. doi: 10.1093/scan/nsl027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Jusczyk PW, Hirsh-Pasek K, Nelson DG, Kennedy LJ, Woodward A, Piwoz J. Perception of acoustic correlates of major phrasal units by young infants. Cognitive Psychology. 1992;24(2):252–93. doi: 10.1016/0010-0285(92)90009-q. [DOI] [PubMed] [Google Scholar]
  47. Jürgens U. Neural pathways underlying vocal control. Neuroscience & Biobehavioral Review. 2002;26(2):235–58. doi: 10.1016/s0149-7634(01)00068-9. [DOI] [PubMed] [Google Scholar]
  48. Jürgens U. The neural control of vocalization in mammals: a review. Journal of Voice. 2009;23(1):1–10. doi: 10.1016/j.jvoice.2007.07.005. [DOI] [PubMed] [Google Scholar]
  49. Kotz S, Paulmann S. Emotion, language, brain. Language and Linguistic Compass. 2011;5:108–25. [Google Scholar]
  50. Ladd DR, Morton R. The perception of intonational emphasis: continuous or categorical? Journal of Phonetics. 1997;25(3):313–42. [Google Scholar]
  51. Levitin D, Menon V. Musical structure is processed in ‘language’ areas of the brain: a possible role for Brodmann area 47 in temporal coherence. Neuroimage. 2003;20(4):2142–52. doi: 10.1016/j.neuroimage.2003.08.016. [DOI] [PubMed] [Google Scholar]
  52. Lindquist KA, Wager TD, Kober H, Bliss-Moreau E, Barrett LF. The brain basis of emotion: a meta-analytic review. Behavioral and Brain Sciences. 2012;35:121–43. doi: 10.1017/S0140525X11000446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Lotze M, Heymans U, Birbaumer N, et al. Differential cerebral activation during observation of expressive gestures and motor acts. Neuropsychologia. 2006;44(10):1787–95. doi: 10.1016/j.neuropsychologia.2006.03.016. [DOI] [PubMed] [Google Scholar]
  54. Juslin P, Laukka P. Communication of emotions in vocal expression and music performance: different channels, same code? Psychological Bulletin. 2003;129:770–814. doi: 10.1037/0033-2909.129.5.770. [DOI] [PubMed] [Google Scholar]
  55. Meyer M, Alter K, Friederici AD, Lohmann G, von Cramon DY. fMRI reveals brain regions mediating slow prosodic modulations in spoken sentences. Human Brain Mapping. 2002;17(2):73–88. doi: 10.1002/hbm.10042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Meyer M, Steinhauer K, Alter K, Friederici AD, von Cramon DY. Brain activity varies with modulation of dynamic pitch variance in sentence melody. Brain and Language. 2004;89(2):277–89. doi: 10.1016/S0093-934X(03)00350-X. [DOI] [PubMed] [Google Scholar]
  57. Monrad-Krohn GH. The prosodic quality of speech and its disorders. Acta Psychiatrica Scandinavica. 1947;3–4:255–69. [Google Scholar]
  58. Nagaratnam N, Nagaratnam K, Ng K, Diu P. Akinetic mutism following stroke. Journal of Clinical Neuroscience. 2004;11(1):25–30. doi: 10.1016/j.jocn.2003.04.002. [DOI] [PubMed] [Google Scholar]
  59. Nichols T, Brett M, Andersson J, Wager T, Poline J-B. Valid conjunction inference with the minimum statistic. Neuroimage. 2005;25(3):653–60. doi: 10.1016/j.neuroimage.2004.12.005. [DOI] [PubMed] [Google Scholar]
  60. Pearsall J, Hanks P, Soanes C, Stevenson A, editors. Oxford English Dictionary. 2nd edn. 2005. revised. New York, NY: Oxford University Press. [Google Scholar]
  61. Papoutsi M, de Zwart JA, Jansma JM, Pickering MJ, Bednar JA, Horwitz B. From phonemes to articulatory codes: an fMRI study of the role of Broca’s area in speech production. Cerebral Cortex. 2009;19(9):2156–65. doi: 10.1093/cercor/bhn239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Paus T, Tomaiuolo F, Otaky N, et al. Human cingulate and paracingulate sulci: pattern, variability, asymmetry, and probabilistic map. Cerebral Cortex. 1996;6:207–14. doi: 10.1093/cercor/6.2.207. [DOI] [PubMed] [Google Scholar]
  63. Pell MD. Recognition of prosody following unilateral brain lesion: influence of functional and structural attributes of prosodic contours. Neuropsychologia. 1998;36(8):701–15. doi: 10.1016/s0028-3932(98)00008-6. [DOI] [PubMed] [Google Scholar]
  64. Pell MD, Baum SR. The ability to perceive and comprehend intonation in linguistic and affective contexts by brain-damaged adults. Brain and Language. 1997;57(1):80–99. doi: 10.1006/brln.1997.1638. [DOI] [PubMed] [Google Scholar]
  65. Petrides M, Pandya DN. Comparative cytoarchitectonic analysis of the human and the macaque ventrolateral prefrontal cortex and corticocortical connection patterns in the monkey. Neuroscience. 2001;16:291–310. doi: 10.1046/j.1460-9568.2001.02090.x. [DOI] [PubMed] [Google Scholar]
  66. Price JL. Prefrontal cortical networks related to visceral function and mood. Annals of the New York Academy of Sciences. 1999;877:383–96. doi: 10.1111/j.1749-6632.1999.tb09278.x. [DOI] [PubMed] [Google Scholar]
  67. Ross ED. The aprosodias: functional-anatomic organization of the affective components of language in the right hemisphere. Archives of Neurology. 1981;38(9):561–9. doi: 10.1001/archneur.1981.00510090055006. [DOI] [PubMed] [Google Scholar]
  68. Sauter DA, Eisner F, Ekman P, Scott SK. Cross-cultural recognition of basic emotions through nonverbal emotional vocalizations. Proceedings of the National Academy of Sciences. 2010;107(6):2408–12. doi: 10.1073/pnas.0908239106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Sauter DA, Scott SK. More than one kind of happiness: Can we recognize vocal expressions of different positive states? Motivation and Emotion. 2007;31(3):192–9. [Google Scholar]
  70. Scherer K, Banse R, Wallbott H. Emotion inferences from vocal expression correlate across languages and cultures. Journal of Cross Cultural Psychology. 2001;32(1):76–92. [Google Scholar]
  71. Schirmer A, Kotz SA. Beyond the right hemisphere: brain mechanisms mediating vocal emotional processing. Trends in Cognitive Science. 2006;10(1):24–30. doi: 10.1016/j.tics.2005.11.009. [DOI] [PubMed] [Google Scholar]
  72. Schröder M. Experimental study of affect bursts. Speech Communication. 2003;40:99–116. [Google Scholar]
  73. Simonyan K, Ostuni J, Ludlow CL, Horwitz B. Functional but not structural networks of the human laryngeal motor cortex show left hemispheric lateralization during syllable but not breathing production. Journal of Neuroscience. 2009;29(47):14912–23. doi: 10.1523/JNEUROSCI.4897-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Sprengelmeyer R, Rausch M, Eysel UT, Przuntek H. Neural structures associated with recognition of facial expressions of basic emotions. Philosophical Transactions of the Royal Society B. 1998;265(1409):1927–31. doi: 10.1098/rspb.1998.0522. [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Strelnikov KN, Vorobyev VA, Chernigovskaya TV, Medvedev SV. Prosodic clues to syntactic processing: a PET and ERP study. Neuroimage. 2006;29(4):1127–34. doi: 10.1016/j.neuroimage.2005.08.021. [DOI] [PubMed] [Google Scholar]
  76. Trauner D, Ballantyne S, Friedland S, Chase C. Disorders of affective and linguistic prosody in children after early unilateral brain damage. Annals of Neurology. 1996;39(3):361–7. doi: 10.1002/ana.410390313. [DOI] [PubMed] [Google Scholar]
  77. Turkeltaub PE, Eden GF, Jones KM, Zeffiro TA. Meta-analysis of the functional neuroanatomy of single-word reading: method and validation. Neuroimage. 2002;16(3):765–80. doi: 10.1006/nimg.2002.1131. [DOI] [PubMed] [Google Scholar]
  78. Turkeltaub PE, Eickhoff SB, Laird AR, Fox M, Wiener M, Fox P. Minimizing within-experiment and within-group effects in activation likelihood estimation meta-analyses. Human Brain Mapping. 2011;33:1–13. doi: 10.1002/hbm.21186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Turken A, Dronkers N. The neural architecture of the language comprehension network: converging evidence from lesion and connectivity analyses. Frontiers in Systems Neuroscience. 2011;5:1–20. doi: 10.3389/fnsys.2011.00001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Wager TD, Davidson ML, Hughes BL, Lindquist MA, Ochsner KN. Prefrontal-subcortical pathways mediating successful emotion regulation. Neuron. 2008;59:1037–50. doi: 10.1016/j.neuron.2008.09.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Weintraub S, Mesulam M, Kramer L. Disturbances in prosody: a right-hemisphere contribution to language. Archives of Neurology. 1981;38:742–44. doi: 10.1001/archneur.1981.00510120042004. [DOI] [PubMed] [Google Scholar]
  82. Wennerstrom A. The Music of Everyday Speech: Prosody and Discourse Analysis. New York, NY: Oxford University Press; 2001. [Google Scholar]
  83. Wiethoff S, Wildgruber D, Kreifelts B, et al. Cerebral processing of emotional prosody: influence of acoustic parameters and arousal. Neuroimage. 2008;39(2):885–93. doi: 10.1016/j.neuroimage.2007.09.028. [DOI] [PubMed] [Google Scholar]
  84. Wildgruber D, Hertrich I, Riecker A, et al. Distinct frontal regions subserve evaluation of linguistic and affective aspects of intonation. Cerebral Cortex. 2004;14I:1384–9. doi: 10.1093/cercor/bhh099. [DOI] [PubMed] [Google Scholar]
  85. Wildgruber D, Riecker A, Hertrich I, et al. Identification of emotional intonation evaluated by fMRI. Neuroimage. 2005;24(4):1233–41. doi: 10.1016/j.neuroimage.2004.10.034. [DOI] [PubMed] [Google Scholar]
  86. Wildgruber D, Ethofer T, Grandjean D, Kreifelts B. A cerebral network model of speech prosody comprehension. International Journal of Speech-Language Pathology. 2009;11:277–81. [Google Scholar]
  87. Witteman J, van Heuven VJP, Schiller N. Hearing feelings: a quantitative meta-analysis on the neuroimaging literature of emotional prosody perception. Neuropsychologia. 2012;50:2752–63. doi: 10.1016/j.neuropsychologia.2012.07.026. [DOI] [PubMed] [Google Scholar]
  88. Witteman J, van Ijzendoorn M, van de Velde D, van Heuven VJP, Schiler NO. The nature of hemispheric specialization for linguistic and emotional prosodic perception: a meta-analysis of the lesion literature. Neuropsychologia. 2011;49:3722–38. doi: 10.1016/j.neuropsychologia.2011.09.028. [DOI] [PubMed] [Google Scholar]
  89. Xu Y, Xu CX. Phonetic realization of focus in English declarative intonation. Journal of Phonetics. 2005;33(2):159–97. [Google Scholar]
  90. Zatorre RJ. Neural specializations for tonal processing. Annals of the New York Academy of Sciences. 2001;930:193–210. doi: 10.1111/j.1749-6632.2001.tb05734.x. [DOI] [PubMed] [Google Scholar]
  91. Zatorre RJ, Evans AC, Meyer E. Neural mechanisms underlying melodic perception and memory for pitch. Journal of Neuroscience. 1994;14(4):1908–19. doi: 10.1523/JNEUROSCI.14-04-01908.1994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Zatorre RJ, Gandour JT. Neural specializations for speech and pitch: moving beyond the dichotomies. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences. 2008;363(1493):1087–104. doi: 10.1098/rstb.2007.2161. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Data

Articles from Social Cognitive and Affective Neuroscience are provided here courtesy of Oxford University Press

RESOURCES