Abstract
In Western music, the major mode is typically used to convey excited, happy, bright or martial emotions, whereas the minor mode typically conveys subdued, sad or dark emotions. Recent studies indicate that the differences between these modes parallel differences between the prosodic and spectral characteristics of voiced speech sounds uttered in corresponding emotional states. Here we ask whether tonality and emotion are similarly linked in an Eastern musical tradition. The results show that the tonal relationships used to express positive/excited and negative/subdued emotions in classical South Indian music are much the same as those used in Western music. Moreover, tonal variations in the prosody of English and Tamil speech uttered in different emotional states are parallel to the tonal trends in music. These results are consistent with the hypothesis that the association between musical tonality and emotion is based on universal vocal characteristics of different affective states.
Introduction
Different sets of tones and tone-relationships are used to convey particular emotions in a variety of musical cultures [1]–[3]. In Western music, this phenomenon is evident in the association of the major mode with positive or excited emotion, and the minor mode with negative or subdued emotion [4]–[7]. It is unclear whether the basis for these associations is intrinsic to tonal perception or the result of exposure to Western music. One way to approach this question is to compare how tones and tone-relationships are used to express emotion in different musical traditions. The classical music of South India (called Carnatic music) is especially useful for such comparison because, as in Western music, the emotions associated with various tone collections (called ragas) are well documented [8]–[11]. Accordingly we asked how the tones in Carnatic melodies composed in ragas associated with positive/excited and negative/subdued emotions compare with the use of tones in Western melodies composed in the major and minor modes. In particular, we wished to know whether there is a common denominator in the way musical tonality is used to convey emotion across cultures, and, if so, its probable basis.
The tonality of a Carnatic melody is organized by the raga in which it is composed. Ragas are analogous to modes in Western music in that they specify a particular collection of tones and tone-relationships. Ragas, however, are more complex because they also prescribe specific patterns of ascending and descending movement and gamakam (ornamentations in Indian music similar to vibrato, portamento and accent in Western music) [12], [13]. Many ragas are associated with specific emotional themes called rasas [8]–[11]. Traditional Indian aesthetics defines 9 rasas, two of which parallel the emotions associated with the major and minor modes and are thus suited to a cross-cultural comparison. These are Haasya, the rasa of joy, happiness, and mirth; and Karunaa, the rasa of sadness, grief, and pity [8]–[11], [14]. The ragas selected for comparison are associated with these rasas.
The basic set of tones and tone-relationships in Carnatic music from which ragas are derived is the 12-tone octave division shown in Figure 1A [12]. An additional 10 tones (microtonal variants of 10 tones in the basic set) are used in gamakams (see above) [12]. These were not included in the present analysis because the variations they introduce are small relative to the principal tones and tone-relationships in a melody, and because they are not always notated. The chromatic scale, which defines the basic set of tones and tone-relationships in Western music from which the major and minor modes are derived, is shown is Figure 1B for comparison [7], [12].
Methods
Music data
For the Haasya rasa, the selected ragas were Bilahari and Mohanam. For the Karunaa rasa, the selected ragas were Naadanaamakriya, Punnaagavaraali, and Varaali (Figure 2A) [8], [9], [11]. Several of these ragas are also associated with other rasas, but in each case a similarly positive/excited or negative/subdued emotional theme is implied. For example, Bilahari can also be associated with Viram [8], [11], the rasa of courage, pride, and confidence; and Varaali can also be associated with Bibhatsam [8], the rasa of disgust, depression, and self-pity. Melodies composed in these ragas were collected from Carnatic music study books [15]–[18]. Carnatic compositions are typically divided into three sections [12]. Each section was treated as a separate melody in the database, and comprised between 21 and 568 notes (M = 97.7, SD = 71.5). All melodies were traditionally notated using the intervals listed in Figure 1A. In total, 194 Carnatic melodies were assembled, 93 associated with positive/excited emotion, and 101 associated with negative/subdued emotion. The characteristics of these melodies were compared to a previous analysis of tonality in classical Western melodies (Figure 2B) [19].
Music Analysis
The tonal structure of the Carnatic melodies was assessed by the intervals between notes. For each note, interval size was determined in relation to the previous melody note (melodic intervals) and the tonic note of the raga (tonic intervals). In both cases interval size was calculated as the difference in cents between the fundamental frequencies (F0s) of the two notes in question. Melodic intervals thus describe variations in F0 over time, whereas tonic intervals describe the place of each F0 within the tonal context of the raga. Interval size data was tallied in 100-cent bins from −1200 to 1200 for melodic intervals, and from 0 to 1200 for tonic intervals. Bin counts were normalized as percentages of the total number of melodic or tonic intervals in a melody and averaged across melodies in one of the emotional conditions examined to obtain average distributions of interval size. Differences between emotional conditions were evaluated for statistical significance using Mann-Whitney U-tests. For melodic intervals comparisons were between the proportion of intervals smaller and larger than a major second (chosen as a cutoff because it separates the largest differences exhibited by positive/excited and negative/subdued melodies; see Figure 4). For tonic intervals, comparisons were between proportions of intervals in each 100-cent bin.
Speech data
Speech was recorded from 20 native speakers of Indian Tamil (10 male) and 20 native speakers of American English (10 male); the speakers ranged in age from 19–72 and had no significant speech or hearing pathology. To mitigate the possible influence of foreign intonation patterns, speakers were required to have lived in the country of their native language for at least 10 years. All subjects gave written informed consent as required by the Institutional Review Boards of Duke University and the National University of Singapore. Speakers read monologues and bi-syllabic expressions in their native language. The monologues consisted of 10 paragraphs, 5 with emotionally positive/excited content and 5 with negative/subdued content (see [19] for examples). The bi-syllabic expressions consisted of 20 emotionally neutral expressions such as “Okay” and “Let's go” in English or “Kodu” and “Nijam” in Tamil. In both cases the speakers were instructed to use the quality of their voice to convey either positive/excited or negative/subdued affect. The recordings were made in a sound-attenuating chamber with an omni-directional capacitor microphone coupled to a solid-state digital recorder. Data were saved on flash memory cards in .wav format (sampling rate = 44.1 kHz; bit depth = 16 bit).
The emotion expressed in the recordings was evaluated by a different group of 6 subject raters (3 native Tamil speakers [1 male] and 3 native English speakers [1 male]). Given the large number of recordings (2000), a subset comprising 240 recordings (one monologue and two bi-syllabic expressions for each speaker in each emotional condition) was randomly selected for evaluation. Raters listened to each recording, categorizing it as expressing joy or sadness and scoring the quality on a five point scale (e.g., from not very joyful to very joyful). The accuracy of emotion recognition and average quality of expression were assessed in terms of speaker language and rater language. The intended emotions were accurately recognized and comparable in terms of quality of expression in both languages (Figure S1).
Speech analysis
For each recording, the fundamental frequency and relative intensity of each 10 ms time-step was calculated using Praat's [20] built in “To Pitch” and “To Intensity” algorithms (pitch floor = 60 Hz, pitch ceiling = 600 Hz; time points with autocorrelation values below the default voicing threshold [0.45] were defined as “unvoiced”). These values were used to find time-points with sufficient periodic energy to be classified as voiced that were also local peaks in the intensity contour. These voiced intensity maxima, which correspond mostly to syllable nuclei and voiced consonants, were used to assess speech prosody (see below). Voiced speech segments, defined as 50 ms windows centered on voiced intensity maxima, were extracted for use in the spectral analysis (see below) (Figure 3). The relatively short window size (50 ms) represents a compromise between maximizing frequency resolution and minimizing temporal changes in fundamental frequency.
Prosodic analysis
For each recording, prosodic intervals were calculated as the frequency difference in cents between adjacent voiced intensity maxima. Like melodic intervals in music, prosodic intervals thus describe changes in frequency over time. Interval size data was tallied in 25-cent bins from −1200 to 1200. Bin counts were normalized as percentages of the total number of prosodic intervals in a recording and averaged across recordings in an emotional condition to obtain representative distributions of interval size. In keeping with the analysis of melodic intervals in music, differences between emotional conditions were evaluated for statistical significance using Mann-Whitney U-tests comparing the proportion of intervals smaller and larger than a major second (see above).
Spectral analysis
For each recording, normalized spectra were calculated for each voiced speech segment. Each segment's spectrum was calculated by applying a Hamming window and a fast Fourier transform (Matlab [21] algorithms ‘hamming.m’ and ‘fft.m’), and normalized in terms of frequency and amplitude. Frequency was normalized by dividing spectral frequency values by the fundamental; amplitude was normalized by dividing spectral amplitude values by the amplitude maximum. Normalized spectra were averaged across recordings in an emotional condition to obtain representative values. Differences between emotional conditions were evaluated for statistical significance by comparing the normalized power at peaks above the tenth harmonic as a percentage of the normalized power at the first 30 harmonic peaks, using Mann-Whitney U-tests (the tenth harmonic was chosen as the cutoff because all the intervals between adjacent harmonics after the tenth are smaller than a major second).
Results
Music
The average distributions of melodic interval sizes in Carnatic melodies composed in ragas associated with positive/excited and negative/subdued emotions are shown in Figure 4A. The most obvious difference between emotional conditions concerns the relative prevalence of interval sizes. Positive/excited raga melodies comprise more intervals equal to or larger than a major second (200 cents), whereas negative/subdued raga melodies comprise more intervals smaller than a major second (see Figure 4A inset). The only exception to this pattern occurs in melodic major thirds, which are slightly more prevalent in negative/subdued raga melodies than positive/excited raga melodies. In comparison to the overall pattern, however, the discrepancy in melodic major thirds is small. The average prevalence of different tonic interval sizes is shown in Figure 4B. Positive/excited raga melodies are characterized by more major intervals, including major seconds, major thirds, and major sixths, whereas negative/subdued raga melodies are characterized by more minor intervals, including minor seconds, minor thirds, tritones, and minor sixths. This pattern suggests that the differences in melodic interval sizes between emotional conditions arise from differences in major and minor tonic intervals.
The average distributions of melodic interval sizes in Western melodies composed in the major and minor modes are shown in Figure 4C. Comparison with Figure 4A shows that Carnatic and Western melodies use melodic intervals in much the same way to convey emotion. Melodies composed in tonalities associated with positive/excited emotion (i.e. major melodies) comprise more intervals equal to or larger than a major second, whereas melodies composed in tonalities associated with negative/subdued emotion (i.e. minor melodies) comprise more intervals smaller than a major second (see Figure 4C inset). The average prevalence of different tonic intervals sizes in Western melodies is shown in Figure 4D. Tonic intervals in Western and Carnatic melodies exhibit both similarities and differences. In both traditions major intervals are more prevalent in positive/excited tonalities and minor intervals are more prevalent in negative/subdued tonalities. There are, however, differences in the way these intervals are used. For example, sixths are more important than thirds in making these distinctions in Carnatic music, whereas the reverse is true in Western music (cf. Figure 4B and D). Further, a greater variety of tonic intervals are used to distinguish the emotional character of Carnatic melodies than Western melodies (10 tonic intervals show significant differences in positive/excited and negative/subdued raga melodies, whereas only 6 intervals differ significantly in major and minor melodies).
Speech prosody
The average distributions of prosodic interval sizes in positive/excited and negative/subdued Tamil and English speech are shown in Figure 5. In both languages, prosodic intervals in positive/excited speech are larger on average than prosodic intervals in negative/subdued speech. For the monologue recordings, only 49% of the intervals in positive/excited speech were smaller than a major second compared to 75% in negative/subdued speech. Likewise, for the bi-syllabic recordings, only 38% of the intervals in positive/excited speech were smaller than a major second compared to 75% in negative/subdued speech (see Figure 5 insets).
Speech spectra
The average normalized spectra of voiced segments in positive/excited and negative/subdued Tamil and English speech are shown in Figure 6. Whereas speech prosody varies in much the same way in Tamil and English speech as a function of emotional state, variations as a function of emotion in speech spectra are less consistent. In English there is relatively more energy in higher harmonics in negative/subdued compared to positive/excited speech in both monologues and bi-syllabic expressions (Figure 6B insets). This pattern tends to emphasize smaller intervals in negative/subdued English speech because the ratios relating higher numbered harmonics are necessarily smaller than those relating lower numbered harmonics (e.g., 16/15 = 1.07, whereas 3/2 = 1.5). The same tendency is apparent in bi-syllabic Tamil speech, but in not the Tamil monologues (cf. Figure 6A and B).
Discussion
Cross-cultural patterns in music
This analysis of Carnatic and Western music shows that melodic intervals in both traditions are generally larger in melodies associated with positive/excited emotion, and smaller in melodies associated with negative/subdued emotion. Similarities are also apparent with respect to major and minor tonic intervals, being more prevalent in melodies associated positive/excited and negative/subdued emotions respectively in both traditions. There are however, differences in the particular tonic intervals emphasized most in making these emotional distinctions (see Figure 4B and D). The differences indicate that the use of a particular tonic interval(s) is not critical for the expression of emotion in tonality; what matters is that the set of tonic intervals provide the right opportunities for larger/smaller melodic intervals to occur. Thus, in Western music, emphasis on tonic thirds to differentiate emotions may not reflect an inherent emotional quality of thirds (as sometimes suggested [22], [23], [24]), but other considerations, such as promoting harmony. Taken together, the results suggest that an important cross-cultural aspect of musical tonality in conveying emotion in melody is the relative size of melodic intervals.
Specific musical intervals in speech prosody
In contrast to at least one other report [22], the present analysis of speech prosody shows that variations in fundamental frequency do not emphasize discrete intervals. Inclusion of more (and more varied) speech samples in addition to examination of interval sizes at higher resolution clearly shows that prosodic interval size varies continuously (see Figure 5). The similarities between speech and music based on prosody thus reflect relative rather than absolute interval size.
The role of speech spectra
Spectral features differentiate excited and subdued speech less well than prosodic differences. In English, speech spectra show a greater proportion of energy in the higher harmonics of negative/subdued relative to positive/excited speech, thus emphasizing smaller intervals. Tamil speech, however, does not exhibit this pattern so clearly (see Figure 6). Given the differences in formant frequencies in Tamil and English speech (see Table S1 and Text S1), these spectral differences probably arise from different vowel phone usage in the two languages.
In previous work on the affective impact of the major and minor modes we focused on spectra, reporting parallel differences in music and American English speech [19]. As we show here, however, similarities between speech prosody and the musical tone collections associated with emotion are more robust than spectral ones (see Text S2 for further discussion). Thus speech prosody is likely to be more influential in driving associations between speech and musical tonality than spectral information.
Possible explanations
One way of explaining the common cross-cultural pattern in the use of melodic intervals to convey emotion is mimicry of the universal tonal characteristics of the human voice in different emotional states [19], [25]–[27]. In both Tamil and English speech negative/subdued affect is characterized by relatively small prosodic intervals, whereas positive/excited affect is characterized by relatively large prosodic intervals. Differences in prosodic interval size reflect differences in fundamental frequency variability, an acoustic parameter well known to be elevated in positive/excited vocal expressions and depressed in negative/subdued vocal expressions in a variety of languages [27]. Thus the pattern of interval sizes in speech and music as a function of emotion are similar.
This similarity, however, does not demand a causal relationship between these two domains. It remains possible, for example, that the association between melodic interval size and emotion in music is entirely the product of exposure to arbitrary cultural conventions that just happen to accord with speech. This possibility notwithstanding, such an interpretation is at odds with considerable evidence. In addition to the present results examining tonality, correlations between vocal and musical emotional expression have been documented with respect to tempo, intensity, timbre, and F0 in a variety of languages and musical traditions (reviewed in [27]). It seems unlikely that these correlations are all coincidental.
Accepting that similarities of emotional expression in speech and music reflect something deeper than coincidence, the question remains whether the voice mimics music, or whether music mimics the voice. The latter seems more likely. First, associations between emotion and the voice are physiologically constrained in ways that associations between music and the voice are not. For example, the increased muscular tension associated with high levels of arousal increases the fundamental frequency of the voice, but has no bearing on the octave a guitarist chooses to play a melody in (other than by vocal association). Second, infants spontaneously produce vocalizations with acoustic properties related to their affective status long before the development of musical abilities [28], [29]. Finally, music appears to be uniquely human, whereas affective vocalization is not [30], [31]. These further facts argue that vocal expression of emotion underlies its expression in music.
The use of musical tones to convey emotion is similar in Western and Carnatic Indian melodies. Melodic intervals are smaller in melodies composed in tonalities associated with negative/subdued emotion and larger in melodies composed in tonalities associated with positive/excited emotions. Melodic interval size as a function of emotion in music parallels the variation of interval size with emotion in speech prosody. Although musical interval size is also related to speech spectra, this association is less robust across cultures. These observations, together with physiological, developmental, and phylogenetic evidence, suggest that the affective impact of musical tonality is generated by association with the characteristics of voiced speech in corresponding emotional states.
Supporting Information
Acknowledgments
We thank Dr. R. Banumathy and Mr. V. Raghuraman of the Alapana Arts School in Singapore for helpful discussions and assistance in obtaining Carnatic song books; Carmen Lai for assistance in data collection/processing; and Mike Hogan for helpful comments on the manuscript.
Footnotes
Competing Interests: The authors have declared that no competing interests exist.
Funding: This work was supported by a grant from the National Science Foundation [BCS-0924181] and a block grant to DP from Duke-National University of Singapore Graduate Medical School. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Touma HH. The Music of the Arabs. Portland, OR: Amadeus Press; 1996. 43 [Google Scholar]
- 2.Patel AD. Music, Language, and the Brain. New York, NY: Oxford University Press; 2008. pp. 313–315. [Google Scholar]
- 3.Miller TE, Williams S. The Garland handbook of Southeast Asian Music. 256, 384. New York, NY: Routledge; 2008. [Google Scholar]
- 4.Zarlino G. Bk. 3 of Le Institutioni hamoniche, translated by Marco G and Palisca C (1968) as: The Art of Counterpoint. New Haven, CT: Yale University Press; 1558. pp. 21–23. [Google Scholar]
- 5.Hevner K. The affective character of the major and minor modes in music. American Journal of Psychology. 1935;47:103–118. [Google Scholar]
- 6.Crowder RG. Perception of the Major/Minor Distinction: I. Historical and Theoretical Foundations. Psychomusicology. 1984;4:3–12. [Google Scholar]
- 7.Aldwell E, Schachter C. Harmony & Voice Leading, 3rd Edition. Belmont, CA: Wadsworth Group/Thomson Learning; 2003. pp. 19–20. [Google Scholar]
- 8.Sambamurthy P. South Indian Music, Book V, 7th Edition. Chennai, India: Indian Music Publishing House; 1999. pp. 160–182. [Google Scholar]
- 9.Balkwill L, Thompson WF. A Cross-Cultural Investigation of the Perception of Emotion in Music: Psychophysical and Cultural Cues. Music Perception. 1999;17:43–64. [Google Scholar]
- 10.Becker J. Anthropological perspectives on music and emotion. In: Juslin PN, Slobada JA, editors. Music and Emotion: Theory and Research. Oxford, UK: Oxford University Press; 2001. pp. 135–160. [Google Scholar]
- 11.Chelladurai PT. The Splendour or South Indian Music. Dindigul, India: Vaigarai Publishers; 2010. pp. 31–32, 55–56, 58–59, 101–102, 115–116, 129–130, 177. [Google Scholar]
- 12.Capwell C. Encyclopedia entries on Indian Music. In: Randel D, editor. The New Harvard Dictionary of Music. Cambridge, MA: Belknap Press; 1986. 781, 333, 680 p. [Google Scholar]
- 13.Swift Ornamentation in South Indian Music and the Violin. Journal of the Society for Asian Music. 1990;21–2:71–89. [Google Scholar]
- 14.Rowell . Music and Musical Thought in Early India. Chicago, IL: The University of Chicago Press; 1992. pp. 327–334. [Google Scholar]
- 15.Ramanathan S. Shyama Sastrigal in Ariya Uruppadigal. Chennai, India: Kalaimagal Isai Kalluri; 1985. [Google Scholar]
- 16.Ramanathan S. Divya Nama Kirtanas of Sri Tyagaraja, Parts I&II. Chennai, India: Kalaimagal Isai Kalluri; 1988. [Google Scholar]
- 17.Banumathy R. Sruti Manjari, Parts 1&2. Singapore: Alapana Centre for Indian Music and Dance; 2002. [Google Scholar]
- 18.Varadan P. Krtimanimalai: Sri Tyagaraja Svami's Compositions, Volumes I&II, English adaptation. Mumbai, India: Vipanci Charitable Trust; 2005a. [Google Scholar]
- 19.Bowling DL, Gill K, Choi J, Prinz J, Purves D. Major and minor music compared to excited and subdued speech. Journal of the Acoustical Society of America. 2010;127:491–503. doi: 10.1121/1.3268504. [DOI] [PubMed] [Google Scholar]
- 20.Boersma P, Weenik D. Praat: doing phonetics by computer (Version 5.2.26) [computer program]. 2009. Available: http://www.praat.org/ Accessed: 2011 June 1.
- 21.MathWorks Inc. 2009. Matlab (Version R2009a) [computer program] (Natick, MA: The MathWorks Inc.)
- 22.Curtis ME, Bharucha JJ. The Minor Third Communicates Sadness in Speech, Mirroring Its Use in Music. Emotion. 2010;30-3:335–348. doi: 10.1037/a0017928. [DOI] [PubMed] [Google Scholar]
- 23.Cooke D. The Language of Music. Oxford, UK: Oxford University Press; 1959. pp. 89–90. [Google Scholar]
- 24.Danielou . The Ragas of Northern Indian Music. New Delhi, India: Munshiram Manoharlal Publishers; 1980. pp. 40–42. [Google Scholar]
- 25.Spencer H. The origin and function of music. Fraser's Magazine. 1857;56:396–408. [Google Scholar]
- 26.Scherer KR. Expression of Emotion in Voice and Music. Journal of Voice. 1995;9-3:235–248. doi: 10.1016/s0892-1997(05)80231-0. [DOI] [PubMed] [Google Scholar]
- 27.Juslin PN, Laukka P. Communication of Emotion in Vocal Expression and Music Performance: Different Channels, Same Code? Psychological Bulletin. 2003;129-5:770–814. doi: 10.1037/0033-2909.129.5.770. [DOI] [PubMed] [Google Scholar]
- 28.Porter FL, Miller RH, Marshall RE. Neonatal Pain Cries: Effect of Circumcision on Acoustic Features and Perceived Urgency. Child Development. 1986;57-3:790–802. [PubMed] [Google Scholar]
- 29.Papoušek M. Early ontogeny of vocal communication in parent-infant interactions. In: Papoušek H, Jürgens U, Papoušek M, editors. Nonverbal Vocal Communication: Comparative & Developmental Approaches. Cambridge, UK: Cambridge University Press; 1992. pp. 230–261. [Google Scholar]
- 30.Darwin C. The Expression of the Emotions in Man and Animals (4th Edition) New York, NY: Oxford University Press; 2009. pp. 88–91. (Original work published 1872) [Google Scholar]
- 31.Jürgens U. Vocalization as an emotional indicator: A Neuroethological Study in the Squirrel Monkey. Behaviour. 1979;69:88–117. doi: 10.1163/156853979x00412. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.