Editor’s Note
While the human brain is hardwired to feel pleasure for basic survival necessities, such as eating and sex, music—although obviously pleasurable—doesn’t offer the same evolutionary advantages. So why do we respond to patterns of sounds that disappear in an instant? Why do we belt music from the top of our lungs, learn to play instruments, and empty our bank accounts to see Bruce Springsteen on Broadway? Our author offers some valuable insights.
Human beings seem to have innate musicality. That is, the capacity to understand and derive pleasure from complex musical patterns appears to be culturally universal.1 Musicality is expressed very early in development.2 In this sense, music may be compared to speech—the other cognitively interesting way that we use sound. But whereas speech is most obviously important for communicating propositions or concepts, obtaining such knowledge, this is not the primary function of music. Rather, it is music’s power to communicate emotions, moods, or affective mental states that seems beneficial to our quality of life.
Which brings us to the question that forms the title of this article: why do we love music? On its face, there is no apparent reason why a sequence or pattern of sounds that has no specific propositional meaning should elicit any kind of pleasurable response. Yet music is widely considered amongst our greatest joys.3 Where does this phenomenon come from?
There are several approaches to this question. A musicologist might have a very different answer than a social scientist. Since I’m a neuroscientist, I would like to address it from that perspective—recognizing that other perspectives may also offer valuable insights. An advantage of neuroscience is that we can relate our answer to established empirical findings and draw from two especially relevant domains: the neuroscience of auditory perception and of the reward system. To give away the punch line of my article, I believe that music derives its power from an interaction between these two systems, the first of which allows us to analyze sound patterns and make predictions about them, and the second of which evaluates the outcomes of these predictions and generates positive (or negative) emotions depending on whether the expectation was met, not met, or exceeded.
The Auditory Perception System
It’s remarkable to think that all sound—a baby crying, thunder, the strains of a waltz—is carried by nothing more than vibrations of molecules in the air. Our rich phenomenological experience of these sounds is the product of a sophisticated perceptual system that takes these vibrations and transforms them into what psychologists call internal representations (perception, thoughts, memories, emotions, etc.), which can be related to our memories of other sounds and knowledge of the world in general. Part of the process has to do with extracting relevant acoustical features from the sounds and encoding them in the pattern of nerve firings.
This process is accomplished by operations happening in three separate brain areas: the brainstem, thalamus, and auditory cortex. A cello string when plucked, for example, will vibrate at a characteristic frequency based on the physics of its materials and its tension; if it is the first string of a conventionally tuned cello, for example, the entire length of the string would vibrate about 65 times in one second, corresponding to the musical note C. Neurons in the aforementioned nuclei and the cortex will respond in a synchronized manner with a corresponding neuronal oscillation4 of 65Hz, thus transforming physical energy to a pattern of neural activity representing sound frequency.
A great deal of research suggests that neurons in the auditory cortex, especially in the right cerebral hemisphere, are important for distinguishing fine gradations of frequency, creating the psychological sensation of pitch.5 Pitch is fundamental to most music, but it is not sufficient merely to detect that a pitch has changed; it is essential to determine the relationships between pitches within a musical system.
An introductory class in music theory would, accordingly, include a description of musical intervals, the ratio between the frequencies of two tones, which determine the patterns that form melodies (when the tones are sequential) and harmonies (when the tones are simultaneous). Importantly, intervals are defined by the relations between pitches independently of the pitch values themselves. That is, a minor third is defined (roughly) as the ratio six to five, so any frequencies in that relation will be perceived as a minor third.
This property, known as transposition, is what allows us to recognize the same song when sung in different keys (if we did not have this capacity, covers of familiar songs would not work). Several studies have indicated that the brain pathways for this kind of computation lie outside the auditory cortex proper, in regions connected to it that are also involved in other kinds of sensory transformations.6
A further complication is that sounds disappear instantly from the environment—unlike, say, objects in a visual scene. Because sounds are evanescent, the brain also needs a mechanism to hold them temporarily in mind, in order to calculate pitch relationships, and other properties. (This is equally important for speech, where a sentence could not otherwise be understood since each word disappears the instant it is spoken.) This capacity depends on the faculty called working memory: roughly, the ability to retain and process information over short time periods.
Several brain circuits emanating from the auditory cortex, mainly the dorsolateral frontal cortex and posterior areas in the parietal lobe, are important for this ability, and hence indispensable for musical perception.7,8 People with congenital amusia (sometimes called tone-deafness)—the inability to comprehend musical relationships and hence to perceive melodies or other musical structures—have been found to have reduced connections between auditory areas and frontal regions, and therefore struggle to figure out the relationships between sounds.8
The Prediction System
The foregoing description gives a brief and highly simplified glimpse of some of the machinery that allows us to perceive tones and determine relationships between them. But, of course, that barely scratches the surface of what’s involved in responding to musical sounds. One of the most important aspects of perception, and one that is critical for music, is the ability to anticipate future events based on past experience.9
This is an essential ability for survival, because an organism can more effectively prepare an appropriate response to an event if that event can be predicted. In the case of music—and, it is thought, of language—there is a rich statistical relationship between patterns of sounds. Every musical system, like every language, has a syntax, that is, a set of rules concerning which sounds follow other sounds. The auditory brain is exquisitely sensitive to such regularities and can learn statistical relationships quickly and efficiently, even early in life, via exposure to exemplars of the system in question (melodies, rhythms, words, and sentences). This is how babies learn about sound patterns in their environments.10
To test the neural substrates of this ability, researchers have devised procedures presenting a set of sounds that follows standard, expected rules (e.g., a sequence of chords), and then introducing a new item that either should or should not follow, based on the context (e.g., an out-of-key chord). In this situation, violations of expectancy yield a characteristic brain response that originates in auditory areas and frontal regions.11
Such results reveal that when we listen to music we not only encode sound properties and their relationships, but also make predictions about what’s coming up (otherwise we would not find the out of key chord jarring). Such predictions are based not only on what has just been experienced in the moment, but also on a knowledge of sound patterns in general drawn from our entire listening history. If one lacks sufficient exposure to the rule system of a different culture, appropriate predictions are often difficult, and that culture’s music may be hard to understand. The same principle would apply to another culture’s language.
What about Pleasure?
The brain mechanisms sketched very roughly above provide the substrates for a number of perceptual and cognitive skills without which, I argue, music would not be possible. If we could not extract pitch information, or hold it in memory, or understand pitch relationships, or make predictions, we could not have what we call music. But none of this explains why we like music so much. For insights into that question, we need to consider a totally different set of brain structures: the reward system.
Scientists have accumulated a great deal of evidence, from both animal models and human studies, to identify the system that signals the presence of a stimulus that has value for the organism. An obvious example would be a hungry rat that is trained to press a lever in response to a cue (such as a light coming on) to receive food. Early studies showed that in this situation certain neurons situated deep in the subcortex, in a structure called the striatum, responded with bursts of dopamine release when the food was delivered.12
But it soon became apparent that these responses were doing much more than merely signaling the presence of food, because after a time, these neurons stopped responding if the amount of food was constant. That is, when the food was expected the neural responses decreased; but if the amount suddenly became larger, a vigorous dopamine response would return; and if less or no food was delivered, the response would actually be inhibited below baseline level.12 Thus, this reward system was encoding the difference between what was expected and what was actually obtained, a concept that became known as reward prediction error (where positive reward prediction error corresponds to a better outcome than was anticipated).
The reward system has been shown to be responsive to a wide range of complex stimuli in both humans and animals. Human neuroimaging studies consistently show activity in the striatum and other components of the reward system when people are shown images of food, or allowed to win money in gambling, or by playing video games, or are shown erotic stimuli.13 Thus, the reward system is thought to underlie the response to many different kinds of inputs that are, globally speaking, beneficial to the organism’s survival or well-being. Food and sex are, of course, biologically necessary for survival (of the individual and the species, respectively); and money may be thought of as having value based on the fact that one may exchange it for other desired items. Imaging studies have also shown reward-system activity for various drugs, including cocaine and amphetamines.14
Music and the Reward System
So what does music have to do with rats pressing levers or people taking drugs? When our group first started to research music-induced pleasure, we did not know whether the same reward system that reacts to biologically relevant stimuli would also be engaged by an entirely abstract stimulus such as music. After all, music is not necessary for survival, nor is it a medium of exchange like money, nor a chemical substance like a drug that can trigger direct neuronal responses.
Our team set out to explore this question using brain imaging techniques that would allow us to measure activity in the striatum during the experience of high pleasure from music. But we immediately ran into a methodological problem: how to measure a subjective response, such as pleasure, in a rigorous, objective, scientifically viable manner? The study of something as complex and potentially uncontrolled as musical emotion represented a particular hurdle. In our first approach to this question, we came up with the idea of studying “chills,” the pleasurable physical response that many people experience while listening to certain musical passages.
The advantage of this approach was that chills are accompanied by physiological changes (increased heart rate, respiration, skin conductance, and so forth), from which we could derive an objective index of the timing and intensity of maximal pleasure. To implement this idea, we asked each participating individual to select their own favorite music, guaranteed to elicit maximal pleasure. Thus armed, we were able to demonstrate in a series of studies that both dorsal and ventral striatum does indeed respond to moments of peak pleasure induced by music15 and, using a neurochemically specific radioligand (a radioactive biochemical substance that binds to a relevant molecule), that dopamine release occurred in the striatum during these moments.16
These studies transformed our understanding of the neurobiology of musical pleasure but left unanswered precisely how or why the reward system is thus engaged. A clue to this question was our observation that there were two phases to the dopamine response: an anticipatory phase, occurring a few seconds prior to peak pleasure in one sub-portion of the striatum, and a second response in a different sub-region at the actual point of pleasure.16 This finding indicates that expectations are as important a source of pleasure as resolutions. Interestingly, music theorists have posited something similar for many years: that emotional arousal and pleasure in music arise from creating tension and then leading the listener to expect its resolution, which resolution is sometimes delayed or manipulated to increase the expectation even further.9
Using the chills response proved very useful; but one could ask whether the engagement of the reward system is limited to this experience; since not everyone gets chills, and since music can be very pleasurable even without any chills, it seemed important to test musical pleasure without any chills being involved. To do so we used a paradigm adapted from neuroeconomics, in which people listen to music excerpts and decide how much money they would be willing to spend to buy a recording of it. The monetary amount is then a proxy for value, and indirectly, for pleasure. With this approach we also found that the ventral striatum showed increased activity as value increased. 17
But a second clue emerged from this study because we also found that as value increased, and the response in the striatum increased, the higher was its coupling (measured in terms of correlated brain activity) to the auditory cortex and its associated network: the more listeners liked a given musical piece (indexed by their willingness to spend more money), the greater the cross-talk between striatum and auditory system.17 This finding is important because it links the activity of the perceptual system, as reviewed above, with that of the reward system. Thus, we propose that the two systems have different functions: the perceptual mechanism computes the relationships between sounds and generates expectancies based on those patterns {“I just heard this sound, followed by that sound, therefore the next one should be X”); the outcome of the prediction (sound X compared to the actual sound perceived) is then evaluated by the reward system (“X is not as good as I expected, therefore it is not pleasurable, or X is surprising and better than expected therefore it is highly pleasurable”).
And just as one might expect from our reward prediction model, the reward response is greatest neither when the outcome is exactly as expected (which is boring), nor when the outcome is completely unpredictable (confusing), but when it hits the “sweet spot” of being somehow better than expected.18 This concept, though still lacking full definition, is one that musicians find intuitive: the best music, typically, neither formulaically follows conventions nor is too complex to follow, but has the virtue of moderation in its ability to surprise the listener with novelty within a predictable framework.5
If the account of musical pleasure presented in the preceding paragraphs is roughly correct, it leads to some testable predictions. First, we reasoned that if musical pleasure arises from interactions between auditory networks and the reward system, then such interactions should be disrupted in persons who are unable to experience musical pleasure. To assess this idea, we sought out such individuals, and discovered that three to four percent of the general population exhibits what we labelled “specific musical anhedonia.” These people have reasonably intact overall hedonic capacity (they enjoy food, sex, social activities, money, even visual art), nor do they have a perceptual disorder such as amusia (tone deafness); they just don’t enjoy or appreciate music, as shown by their lack of physiological responses to it.19
When we scanned their brains, we discovered that their reward system responded normally to a gambling game, but not to music; and the coupling between auditory and reward systems was essentially absent during music listening.20 Thus, as predicted by our model, musical anhedonia emerges in the absence of the typical interaction between the two systems.
One might say that musical anhedonia represents a chicken and egg problem: perhaps it is the lack of musical pleasure that leads to decreased connectivity between auditory and reward systems, and not vice versa. To exclude such a possibility, it is critical to test a second prediction arising from our model: if activity in the reward system really underpins musical pleasure, then we should be able to modulate that pleasure by manipulating activity within that system in the normal brain.
Previous work had shown it possible to excite or inhibit the reward system, by changing dopamine activity in the striatum with a noninvasive brain stimulation technique known as transcranial magnetic stimulation.21 We recently implemented this technique while people listened to music (their own favorites and some choses by us) and found that, just as we predicted, listeners reported more pleasure and showed greater physiological responses (skin conductance) to music in the context of excitatory stimulation, and reported less pleasure, even to their own selected music, and showed diminished physiological responses during the inhibitory stimulation.22 This finding provides causal evidence that musical pleasure is directly linked to reward system activity.
I am very pleased to see that music neuroscience has shifted over the past decades from a fringe area to a solid research domain, with labs in many countries making important contributions and substantial progress reported in respected journals. What not long ago seemed like an intractable problem—how music can result in strong affective and pleasurable responses—is now a topic that we understand well enough to have significant insights into and testable hypotheses about. It is an exciting time to be working in this domain; we look forward to future developments which, based on the science discussed in this piece, we hope will include applications to clinical, educational, and even artistic domains.
Footnotes
Robert Zatorre, Ph.D., is a cognitive neuroscientist at the Montreal Neurological Institute of McGill University. His laboratory studies the neural substrate for auditory cognition, with special emphasis on two characteristically human abilities: speech and music. He and his collaborators have published about 300 scientific papers on topics including pitch perception, auditory imagery, brain plasticity, and musical pleasure. In 2006 he became the founding co-director of the international laboratory for Brain, Music, and Sound research (BRAMS), a unique multi-university consortium dedicated to the cognitive neuroscience of music. He tries to keep up his baroque repertoire on the organ whenever he gets a chance.
References
- 1.Savage PE, Brown S, Sakai E, Currie TE. Statistical universals reveal the structures and functions of human music. Proceedings of the National Academy of Sciences. 2015;112(29):8987–8992. doi: 10.1073/pnas.1414495112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Honing H. The origins of musicality. MIT Press; 2018. [Google Scholar]
- 3.Brattico E, Jacobsen T. Subjective Appraisal of Music. Annals of the New York Academy of Sciences. 2009;1169(1):308–317. doi: 10.1111/j.1749-6632.2009.04843.x. [DOI] [PubMed] [Google Scholar]
- 4.Kraus N, Anderson S, White-Schwoch T, Fay RR, Popper RN. The Frequency-Following Response. 2017. (Springer Handbook of Auditory Research). [Google Scholar]
- 5.Zatorre RJ, Belin P, Penhune VB. Structure and function of auditory cortex: music and speech. Trends in cognitive sciences. 2002;6(1):37–46. doi: 10.1016/s1364-6613(00)01816-7. [DOI] [PubMed] [Google Scholar]
- 6.Klein ME, Zatorre RJ. Representations of invariant musical categories are decodable by pattern analysis of locally distributed BOLD responses in superior temporal and intraparietal sulci. Cerebral Cortex. 2014;25(7):1947–1957. doi: 10.1093/cercor/bhu003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Kumar S, Joseph S, Gander PE, Barascud N, Halpern AR, Griffiths TD. A brain system for auditory working memory. Journal of Neuroscience. 2016;36(16):4492–4505. doi: 10.1523/JNEUROSCI.4341-14.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Albouy P, Mattout J, Bouet R, Maby E, Sanchez G, Aguera PE, Tillmann B. Impaired pitch perception and memory in congenital amusia: the deficit starts in the auditory cortex. Brain. 2013;136(5):1639–1661. doi: 10.1093/brain/awt082. [DOI] [PubMed] [Google Scholar]
- 9.Huron DB. Sweet anticipation: Music and the psychology of expectation. MIT press; 2006. [Google Scholar]
- 10.Saffran JR, Johnson EK, Aslin RN, Newport EL. Statistical learning of tone sequences by human infants and adults. Cognition. 1999;70(1):27–52. doi: 10.1016/s0010-0277(98)00075-4. [DOI] [PubMed] [Google Scholar]
- 11.Maess B, Koelsch S, Gunter TC, Friederici AD. Musical syntax is processed in Broca’s area: an MEG study. Nature neuroscience. 2001;4(5):540. doi: 10.1038/87502. [DOI] [PubMed] [Google Scholar]
- 12.Schultz W. Getting formal with dopamine and reward. Neuron. 2002;36(2):241–263. doi: 10.1016/s0896-6273(02)00967-4. [DOI] [PubMed] [Google Scholar]
- 13.Sescousse G, Caldú X, Segura B, Dreher JC. Processing of primary and secondary rewards: a quantitative meta-analysis and review of human functional neuroimaging studies. Neuroscience & Biobehavioral Reviews. 2013;37(4):681–696. doi: 10.1016/j.neubiorev.2013.02.002. [DOI] [PubMed] [Google Scholar]
- 14.Drevets WC, Gautier C, Price JC, Kupfer DJ, Kinahan PE, Grace AA, Mathis CA. Amphetamine-induced dopamine release in human ventral striatum correlates with euphoria. Biological psychiatry. 2001;49(2):81–96. doi: 10.1016/s0006-3223(00)01038-6. [DOI] [PubMed] [Google Scholar]
- 15.Blood AJ, Zatorre RJ. Intensely pleasurable responses to music correlate with activity in brain regions implicated in reward and emotion. Proceedings of the National Academy of Sciences. 2001;98(20):11818–11823. doi: 10.1073/pnas.191355898. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Salimpoor VN, Benovoy M, Larcher K, Dagher A, Zatorre RJ. Anatomically distinct dopamine release during anticipation and experience of peak emotion to music. Nature neuroscience. 2011;14(2):257. doi: 10.1038/nn.2726. [DOI] [PubMed] [Google Scholar]
- 17.Salimpoor VN, van den Bosch I, Kovacevic N, McIntosh AR, Dagher A, Zatorre RJ. Interactions between the nucleus accumbens and auditory cortices predict music reward value. Science. 2013;340(6129):216–219. doi: 10.1126/science.1231059. [DOI] [PubMed] [Google Scholar]
- 18.Chmiel A, Schubert E. Back to the inverted-U for music preference: A review of the literature. Psychology of Music. 2017;45(6):886–909. [Google Scholar]
- 19.Mas-Herrero E, Zatorre RJ, Rodriguez-Fornells A, Marco-Pallarés J. Dissociation between musical and monetary reward responses in specific musical anhedonia. Current Biology. 2014;24(6):699–704. doi: 10.1016/j.cub.2014.01.068. [DOI] [PubMed] [Google Scholar]
- 20.Martínez-Molina N, Mas-Herrero E, Rodríguez-Fornells A, Zatorre RJ, Marco-Pallarés J. Neural correlates of specific musical anhedonia. Proceedings of the National Academy of Sciences. 2016;113(46):E7337–E7345. doi: 10.1073/pnas.1611211113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Strafella AP, Paus T, Barrett J, Dagher A. Repetitive transcranial magnetic stimulation of the human prefrontal cortex induces dopamine release in the caudate nucleus. Journal of Neuroscience. 2001;21(15):RC157–RC157. doi: 10.1523/JNEUROSCI.21-15-j0003.2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Mas-Herrero E, Dagher A, Zatorre RJ. Modulating musical reward sensitivity up and down with transcranial magnetic stimulation. Nature Human Behaviour. 2018;2(1):27. doi: 10.1038/s41562-017-0241-z. [DOI] [PubMed] [Google Scholar]