Abstract
Hemispheric asymmetries in auditory cognition have been recognized for a long time, but their neural basis is still debated. Here I focus on specialization for processing of speech and music, the two most important auditory communication systems that humans possess. A great deal of evidence from lesion studies and functional imaging suggests that aspects of music linked to the processing of pitch patterns depend more on right than left auditory networks. A complementary specialization for temporal resolution has been suggested for left auditory networks. These diverse findings can be integrated within the context of the spectrotemporal modulation framework, which has been developed as a way to characterize efficient neuronal encoding of complex sounds. Recent studies show that degradation of spectral modulation impairs melody perception but not speech content, whereas degradation of temporal modulation has the opposite effect. Neural responses in the right and left auditory cortex in those studies are linked to processing of spectral and temporal modulations, respectively. These findings provide a unifying model to understand asymmetries in terms of sensitivity to acoustical features of communication sounds in humans. However, this explanation does not account for evidence that asymmetries can shift as a function of learning, attention, or other top-down factors. Therefore, it seems likely that asymmetries arise both from bottom-up specialization for acoustical modulations and top-down influences coming from hierarchically higher components of the system. Such interactions can be understood in terms of predictive coding mechanisms for perception.
Keywords: lateralization, music, speech, neuroimaging, spectrotemporal modulation
Introduction
We have known since observations in the mid-19th century about aphasia that the two cerebral hemispheres of the human brain do not have identical functions (Manning and Thomas-Antérion, 2011). Yet, debate continues to this day on the underlying principles that govern these differences. Asymmetries have been described in many domains, including visuospatial, motor, and affective functions. But here I will focus on asymmetries related to auditory processes. A great deal of work has been carried out on the linguistic functions of the left hemisphere, in part because those earliest observations showed such salient effects of left-hemisphere lesions on language in general and speech in particular. But it is instructive to compare speech to that other auditory-motor communication system that we humans possess: music.
Comparisons between music and language are extremely valuable for many reasons (Patel, 2010), and can be carried out at many different levels of analysis. In this mini-review I will focus on certain acoustical features that I argue are critical for important aspects of musical processing, and contrast them with those most relevant for speech, to show that auditory networks within each hemisphere are specialized in terms of sensitivity to those features. However, one of the main points I wish to make is that those input-driven specializations interact with top-down mechanisms to yield a complex interplay between the two hemispheres.
Specialization for spectral features
A great deal of evidence supports the idea that certain aspects of musical perceptual functions depend to a greater extent on auditory networks in the right hemisphere than the left. This conclusion is supported by a recent meta-analysis of the effects of vascular lesions on musical perceptual skills (Sihvonen et al., 2019), as well as by early experimental studies of the consequences of temporal-lobe excisions (Milner, 1962; Samson and Zatorre, 1988; Zatorre, 1988; Liégeois-Chauvel et al., 1998). Apart from these effects of acquired lesions, deficits in congenital amusia (also termed tone-deafness) also seem to be linked to a disruption in the organization of connections between right auditory cortex and right inferior frontal regions. Evidence for this conclusion comes from studies of functional activation (Albouy et al., 2013) and functional connectivity (Hyde et al., 2011; Albouy et al., 2015), as well as anatomical measures of cortical thickness (Hyde et al., 2007) and of white-matter fiber connections (Loui et al., 2009; Albouy et al., 2013).
These findings are compelling, but what particular aspects of perception are most relevant in eliciting these asymmetries? A hint comes from the amusia literature, where several authors have found that the ability to process fine pitch differences seems to be particularly impaired (Hyde and Peretz, 2004; Tillmann et al., 2016). Those results are echoed in surgical lesion studies showing that damage to an area adjacent to right primary auditory cortex specifically leads to elevated pitch-direction discrimination thresholds compared to equivalent lesions on the left side (Johnsrude et al., 2000). Fine pitch resolution is important for processing musical features such as melody and harmony (Zatorre and Baum, 2012), which is why if that function is impaired, amusia typically follows (Peretz et al., 2002; Hyde and Peretz, 2004).
Many neuroimaging studies also align well with the idea that the right auditory cortical system is specialized for fine pitch processing. Several experiments have found that functional MRI responses in right auditory cortex scale more strongly than those on the left as pitch distance is manipulated from smaller to larger in a tone pattern; that is, the right side is more sensitive to variation of this parameter (Zatorre and Belin, 2001; Jamison et al., 2006; Hyde et al., 2008; Zatorre et al., 2012). Supportive findings also come from an MEG experiment examining spectral and temporal deviant detection (Okamoto and Kakigi, 2015). Importantly, the asymmetry of response seems to be linked to individual differences in pitch perception skill, thus showing a direct brain-behavior link. For example, functional MRI activity in the right (but not the left) auditory cortex of a group of musicians was correlated with their individual pitch discrimination thresholds (Bianchi et al., 2017). A correlation between individual pitch discrimination thresholds and the amplitude of the frequency-following response measured from the right (but not the left) auditory cortex was also observed using MEG (Coffey et al., 2016).
If spectral resolution on the right is better than on the left, what could be the physiological mechanism behind it? One possible answer was provided by an analysis of local functional connectivity patterns in relation to frequency tuning (Cha et al., 2016). This study found that the interconnectivity between voxels in auditory cortex is greater for those whose frequency tuning is more similar than for voxels which are tuned to more distant frequencies. But of greater relevance is that this pattern was more marked within right than left core auditory regions. In other words, frequency selectivity played a greater role on the right than the left, which would then lead to sharper tuning on the right, since there would be summation of activity from neurons with similar response properties. This conclusion is in line with electrophysiological recordings indicating that sharp tuning of neurons to frequency in early auditory cortex depends on excitatory intracortical inputs, rather than thalamic inputs (Liu et al., 2007).
Specialization for temporal features
The evidence favoring a relative enhancement of frequency resolution in the right auditory networks is paralleled by evidence favoring a relative enhancement of temporal resolution in the left hemisphere. Several functional neuroimaging studies have shown that parametric variation of temporal features of stimuli is better tracked by responses coming from the left auditory cortex and adjacent regions compared to the right (Zatorre and Belin, 2001; Schönwiesner et al., 2005; Jamison et al., 2006; Obleser et al., 2008). Causal evidence in favor of this concept was also provided by a brain stimulation experiment showing increased thresholds for gap detection, after left, but not right auditory cortex disruption (Heimrath et al., 2014).
Spectrotemporal modulations
A theoretically powerful way to integrate these findings is by considering how these patterns fit with models of spectrotemporal modulation. Many neurophysiological studies exist showing that the response properties of auditory cortical neurons across species are well described in terms of joint sensitivity to spectral and temporal modulations found in the stimulus (Shamma, 2001). This mechanism is thought to enable efficient encoding of complex real-world sounds (Singh and Theunissen, 2003), especially those that are an important part of the animal’s communicative repertoire (Gehr et al., 2000; Woolley et al., 2005). Sensitivity to spectrotemporal modulations in auditory cortex has also been described using both neuroimaging (Schonwiesner and Zatorre, 2009; Santoro et al., 2014; Venezia et al., 2019) and intracortical recordings in humans (Mesgarani et al., 2014; Hullett et al., 2016).
Two recent studies have brought together the research questions surrounding hemispheric differences with the spectrotemporal modulation hypothesis to yield evidence that functional asymmetries map well onto this theoretical framework. One study (Flinker et al., 2019) used MEG to measure brain activity associated either with the verbal content, or the timbre (male vs. female voice, which is largely based on spectral cues) of spoken sentences. Behaviorally, they reported that when temporal modulations were filtered out, speech comprehension was affected but vocal timbre was not, and vice-versa for filtering of spectral modulations. The imaging data showed greater left auditory cortex response for the temporal cues in speech, and a right, albeit weaker lateralization effect for the spectral cues. The second study (Albouy et al., 2020) used sung sentences whose speech and melodic content had been fully orthogonalized ensuring independence of the two types of cues (Figure 1). Behavioral data showed a double dissociation such that degradation of temporal cues affected comprehension of the words to the song but not the melody, whereas degradation of spectral cues affected discrimination of the melodies but had no effect on the speech component. The functional imaging data reflected the behavioral data in that speech content could only be decoded from left auditory cortex, but was abolished by temporal degradation, whereas melodic content could only be decoded from right auditory cortex, but was abolished by spectral degradation.
These converging findings from experiments using different techniques strongly support the idea that left and right auditory cortices are linked to heightened resolution in temporal and spectral modulation, respectively. This explanation fits with a broader idea that the nervous system optimizes its representations according to the properties of the physical environment that are most relevant, as has been proposed for vision (Simoncelli and Olshausen, 2001), and for speech (Gervain and Geffen, 2019). I suggest expanding this concept to encompass hemispheric asymmetries on the grounds that humans have two main auditory communication systems, speech and music (Zatorre et al., 2002; Mehr et al., 2021), and that they each exploit, to some extent at least, opposite ends of the temporal-spectral continuum; so the best way to accommodate the competing requirements of the two types of signals is by segregating the necessary specializations within each hemisphere. Thus, rather than think in terms of specializations at the cognitive domain level (speech vs. music), we can reconceptualize it in terms of specialization at the acoustical feature level.
This interpretation predicts that the two domains are lateralized only to the extent that they make greater use of one or another of those cues. We need to keep in mind that both music and speech utilize both temporal and spectral modulations. In the case of speech, spectral modulations are important in carrying prosodic information, and, in tonal languages, lexical information. Interestingly, a good amount of evidence suggests that prosodic processing depends more on right-hemisphere structures, in accord with our model (Sammler et al., 2015). These kinds of spectral cues are important for some aspects of communication, but they do not seem to be quite as important for speech comprehension as those driven by temporal modulations, based on the fact that degradation of temporal but not spectral cues abolishes speech comprehension in the two studies mentioned earlier (Flinker et al., 2019; Albouy et al., 2020). That conclusion was already known from an early influential study (Shannon et al., 1995) that demonstrated that comprehension was well-preserved when normal speech was replaced by amplitude-modulated noise passed through as few as three or four filter banks centered at different frequencies. This procedure degraded the spectral content but preserved most of the temporal modulations. Indeed, this property is what enables cochlear implants to transmit comprehensible speech despite poor representation of spectral modulations due to the limited number of channels available.
Music also contains both spectral and temporal modulations. The latter are obviously critical for transmitting information that can be used to perceive rhythm and metrical organization, and hence the importance of temporal modulations may vary depending on the nature of the music and of the instruments used to generate it (e.g., percussion vs. song). Moreover, pitch information and temporal information interact in interesting, complex ways in music cognition (Jones, 2014). So it is simplistic to think of the two dimensions are entirely independent of one another. However, the fact remains that, as mentioned above, the poor spectral resolution that can be observed with congenital amusia seems to lead to a more global inability to learn the relevant rules of music, and results in a fairly global deficit. So this observation would argue that even if both types of cues are present and important for music, spectral cues seem to play a more prominent role.
Top-down effects
One might conclude from all the foregoing that hemispheric differences are driven exclusively by low-level acoustical features. But that does not seem to be the whole story. There are in fact numerous experiments showing that even when acoustics are held constant, hemispheric responses can be modulated. A good example is provided by studies showing that sine-wave speech analogs elicit left auditory cortex responses only after training that led to them being perceived as speech, and not in the naive state when they were perceived as just weird sounds (Liebenthal et al., 2003; Dehaene-Lambertz et al., 2005; Möttönen et al., 2006). A complementary phenomenon can be seen with speech sounds that when looped repeatedly begin to sound like music (Deutsch et al., 2011; Falk et al., 2014). Once the stimulus was perceived as music, more brain activity was seen in some right-hemisphere regions that were not detected before the perceptual transformation (Tierney et al., 2013). Tracking of pitch contours in speech can also shift from right to bilateral auditory regions as a function of selective attention (Brodbeck and Simon, 2022).
These kinds of results have sometimes been interpreted as evidence in favor of domain-specific models, on the grounds that bottom-up mechanisms cannot explain the results since the inputs are held constant in those studies. However, given the strength of the findings reviewed above that spectrotemporal tuning is asymmetric, another way to interpret these effects is that they represent interactions between feedforward and feedback systems that interconnect auditory areas with higher-order processing regions, especially in the frontal cortex. Although this idea remains to be worked out in any detail, it would be compatible with known control functions of the frontal cortex, which is reciprocally connected with auditory cortical processing streams.
The idea that interactions occur between ascending, stimulus-driven responses, and descending, more cognitive influences can also be thought of in the context of predictive coding models (Friston, 2010). A great deal of work has recently been devoted to this framework, which essentially proposes that perception is enabled by the interface between predictions generated at higher levels of the hierarchy that influence stimulus-driven encoding processes at lower levels of the hierarchy. When the latter signals do not match the prediction, an error signal is generated, which can be used for updating of the internal model (that is, learning). These models have gained prominence because they can explain many phenomena not easily accounted for by more traditional bottom-up driven models of perception, even if they also raise questions that are not yet fully answered (Heilbron and Chait, 2018).
As applied to the question at hand, the idea would be that as a complex stimulus like speech or music is being processed, continuous predictions and confirmations/errors would be generated at different levels of the system. Depending on the spectrotemporal content of the signal, neuronal networks in the left or right auditory cortex would predominate in the initial processing; but as top-down predictions are generated that are based on higher-order features, then the activity could shift from one side to another. So, in the case of sine-wave speech for instance, initial, naïve processing would presumably involve right auditory cortex since the stimulus contains a great deal of spectral modulation. But once the listener is able to apply top-down control to disambiguate how those sounds could fit into a linguistic pattern, then more language-relevant predictions would be generated that could inhibit spectral-based processing in favor of temporal-based processing. By the same token, hemispheric differences could be amplified by these interactions even if initial processing differences in early parts of the auditory system are only slightly asymmetric. This scenario remains largely speculative at the moment, but at least sets up some testable hypotheses for future research.
Author contributions
RZ wrote the manuscript and approved the submitted version.
Conflict of interest
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
- Albouy P., Benjamin L., Morillon B., Zatorre R. J. (2020). Distinct sensitivity to spectrotemporal modulation supports brain asymmetry for speech and melody. Science 367:1043. 10.1126/science.aaz3468 [DOI] [PubMed] [Google Scholar]
- Albouy P., Mattout J., Bouet R., Maby E., Sanchez G., Aguera P.-E., et al. (2013). Impaired pitch perception and memory in congenital amusia: The deficit starts in the auditory cortex. Brain 136 1639–1661. 10.1093/brain/awt082 [DOI] [PubMed] [Google Scholar]
- Albouy P., Mattout J., Sanchez G., Tillmann B., Caclin A. (2015). Altered retrieval of melodic information in congenital amusia: Insights from dynamic causal modeling of MEG data. Front. Hum. Neurosci. 9:20. 10.3389/fnhum.2015.00020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bianchi F., Hjortkjaer J., Santurette S., Zatorre R. J., Siebner H. R., Dau T. (2017). Subcortical and cortical correlates of pitch discrimination: Evidence for two levels of neuroplasticity in musicians. Neuroimage 163 398–412. 10.1016/j.neuroimage.2017.07.057 [DOI] [PubMed] [Google Scholar]
- Brodbeck C., Simon J. Z. (2022). Cortical tracking of voice pitch in the presence of multiple speakers depends on selective attention. Front. Neurosci. 16:828546. 10.3389/fnins.2022.828546 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cha K., Zatorre R. J., Schonwiesner M. (2016). Frequency selectivity of voxel-by-voxel functional connectivity in human auditory cortex. Cereb. Cortex 26 211–224. 10.1093/cercor/bhu193 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Coffey E. B., Herholz S. C., Chepesiuk A. M., Baillet S., Zatorre R. J. (2016). Cortical contributions to the auditory frequency-following response revealed by MEG. Nat. Commun. 7:11070. 10.1038/ncomms11070 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dehaene-Lambertz G., Pallier C., Serniclaes W., Sprenger-Charolles L., Jobert A., Dehaene S. (2005). Neural correlates of switching from auditory to speech perception. Neuroimage 24 21–33. 10.1016/j.neuroimage.2004.09.039 [DOI] [PubMed] [Google Scholar]
- Deutsch D., Henthorn T., Lapidis R. (2011). Illusory transformation from speech to song. J. Acoust. Soc. Am. 129 2245–2252. 10.1121/1.3562174 [DOI] [PubMed] [Google Scholar]
- Falk S., Rathcke T., Dalla Bella S. (2014). When speech sounds like music. J. Exp. Psychol. Hum. Percept. Perform. 40 1491–1506. 10.1037/a0036858 [DOI] [PubMed] [Google Scholar]
- Flinker A., Doyle W. K., Mehta A. D., Devinsky O., Poeppel D. (2019). Spectrotemporal modulation provides a unifying framework for auditory cortical asymmetries. Nat. Hum. Behav. 3 393–405. 10.1038/s41562-019-0548-z [DOI] [PMC free article] [PubMed] [Google Scholar]
- Friston K. (2010). The free-energy principle: A unified brain theory? Nat. Rev. Neurosci. 11 127–138. 10.1038/nrn2787 [DOI] [PubMed] [Google Scholar]
- Gehr D. D., Komiya H., Eggermont J. J. (2000). Neuronal responses in cat primary auditory cortex to natural and altered species-specific calls. Hear. Res. 150 27–42. 10.1016/s0378-5955(00)00170-2 [DOI] [PubMed] [Google Scholar]
- Gervain J., Geffen M. N. (2019). Efficient neural coding in auditory and speech perception. Trends Neurosci. 42 56–65. 10.1016/j.tins.2018.09.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Heilbron M., Chait M. (2018). Great expectations: Is there evidence for predictive coding in auditory cortex? Neuroscience 389 54–73. 10.1016/j.neuroscience.2017.07.061 [DOI] [PubMed] [Google Scholar]
- Heimrath K., Kuehne M., Heinze H. J., Zaehle T. (2014). Transcranial direct current stimulation (tDCS) traces the predominance of the left auditory cortex for processing of rapidly changing acoustic information. Neuroscience 261 68–73. 10.1016/j.neuroscience.2013.12.031 [DOI] [PubMed] [Google Scholar]
- Hullett P. W., Hamilton L. S., Mesgarani N., Schreiner C. E., Chang E. F. (2016). Human superior temporal gyrus organization of spectrotemporal modulation tuning derived from speech stimuli. J. Neurosci. 36 2014–2026. 10.1523/JNEUROSCI.1779-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hyde K. L., Lerch J. P., Zatorre R. J., Griffiths T. D., Evans A. C., Peretz I. (2007). Cortical thickness in congenital amusia: When less is better than more. J. Neurosci. 27 13028–13032. 10.1523/JNEUROSCI.3039-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hyde K. L., Peretz I. (2004). Brains that are out of tune but in time. Psychol. Sci. 15 356–360. 10.1111/j.0956-7976.2004.00683.x [DOI] [PubMed] [Google Scholar]
- Hyde K. L., Peretz I., Zatorre R. J. (2008). Evidence for the role of the right auditory cortex in fine pitch resolution. Neuropsychologia 46 632–639. 10.1016/j.neuropsychologia.2007.09.004 [DOI] [PubMed] [Google Scholar]
- Hyde K. L., Zatorre R. J., Peretz I. (2011). Functional MRI evidence of an abnormal neural network for pitch processing in congenital amusia. Cereb. Cortex 21 292–299. 10.1093/cercor/bhq094 [DOI] [PubMed] [Google Scholar]
- Jamison H. L., Watkins K. E., Bishop D. V., Matthews P. M. (2006). Hemispheric specialization for processing auditory nonspeech stimuli. Cereb. Cortex 16 1266–1275. 10.1093/cercor/bhj068 [DOI] [PubMed] [Google Scholar]
- Johnsrude I. S., Penhune V. B., Zatorre R. J. (2000). Functional specificity in the right human auditory cortex for perceiving pitch direction. Brain 123(Pt 1) 155–163. 10.1093/brain/123.1.155 [DOI] [PubMed] [Google Scholar]
- Jones M. R. (2014). “Dynamics of musical patterns: How do melody and rhythm fit together?,” in Psychology and music, eds Tighe T. J., Dowling W. J. (London: Psychology Press; ), 67–92. [Google Scholar]
- Liebenthal E., Binder J. R., Piorkowski R. L., Remez R. E. (2003). Short-term reorganization of auditory analysis induced by phonetic experience. J. Cogn. Neurosci. 15 549–558. 10.1162/089892903321662930 [DOI] [PubMed] [Google Scholar]
- Liégeois-Chauvel C., Peretz I., Babaï M., Laguitton V., Chauvel P. (1998). Contribution of different cortical areas in the temporal lobes to music processing. Brain 121 1853–1867. 10.1093/brain/121.10.1853 [DOI] [PubMed] [Google Scholar]
- Liu B.-H., Wu G. K., Arbuckle R., Tao H. W., Zhang L. I. (2007). Defining cortical frequency tuning with recurrent excitatory circuitry. Nat. Neurosci. 10 1594–1600. 10.1038/nn2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Loui P., Alsop D., Schlaug G. (2009). Tone deafness: A new disconnection syndrome? J. Neurosci. 29 10215–10220. 10.1523/JNEUROSCI.1701-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Manning L., Thomas-Antérion C. (2011). Marc Dax and the discovery of the lateralisation of language in the left cerebral hemisphere. Rev. Neurol. 167 868–872. 10.1016/j.neurol.2010.10.017 [DOI] [PubMed] [Google Scholar]
- Mehr S. A., Krasnow M. M., Bryant G. A., Hagen E. H. (2021). Origins of music in credible signaling. Behav. Brain Sci. 44:e60. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mesgarani N., Cheung C., Johnson K., Chang E. F. (2014). Phonetic feature encoding in human superior temporal gyrus. Science 343 1006–1010. 10.1126/science.1245994 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Milner B. (1962). “Laterality effects in audition,” in Interhemispheric relations and cerebral dominance, ed. Mountcastle V. B. (Baltimore, MD: The Johns Hopkins Press; ), 177–195. [Google Scholar]
- Möttönen R., Calvert G. A., Jääskeläinen I. P., Matthews P. M., Thesen T., Tuomainen J., et al. (2006). Perceiving identical sounds as speech or non-speech modulates activity in the left posterior superior temporal sulcus. Neuroimage 30 563–569. 10.1016/j.neuroimage.2005.10.002 [DOI] [PubMed] [Google Scholar]
- Obleser J., Eisner F., Kotz S. A. (2008). Bilateral speech comprehension reflects differential sensitivity to spectral and temporal features. J. Neurosci. 28 8116–8123. 10.1523/JNEUROSCI.1290-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Okamoto H., Kakigi R. (2015). Hemispheric asymmetry of auditory mismatch negativity elicited by spectral and temporal deviants: A magnetoencephalographic study. Brain Topogr. 28 471–478. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patel A. D. (2010). Music, language, and the brain. New York, NY: Oxford university press. [Google Scholar]
- Peretz I., Ayotte J., Zatorre R. J., Mehler J., Ahad P., Penhune V. B., et al. (2002). Congenital amusia: A disorder of fine-grained pitch discrimination. Neuron 33 185–191. 10.1016/S0896-6273(01)00580-3 [DOI] [PubMed] [Google Scholar]
- Sammler D., Grosbras M.-H., Anwander A., Bestelmeyer P. E. G., Belin P. (2015). Dorsal and ventral pathways for prosody. Curr. Biol. 25 3079–3085. 10.1016/j.cub.2015.10.009 [DOI] [PubMed] [Google Scholar]
- Samson S., Zatorre R. J. (1988). Melodic and harmonic discrimination following unilateral cerebral excision. Brain Cogn. 7 348–360. 10.1016/0278-2626(88)90008-5 [DOI] [PubMed] [Google Scholar]
- Santoro R., Moerel M., De Martino F., Goebel R., Ugurbil K., Yacoub E., et al. (2014). Encoding of natural sounds at multiple spectral and temporal resolutions in the human auditory cortex. PLoS Comput. Biol. 10:e1003412. 10.1371/journal.pcbi.1003412 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schönwiesner M., Rübsamen R., Von Cramon D. Y. (2005). Hemispheric asymmetry for spectral and temporal processing in the human antero-lateral auditory belt cortex. Eur. J. Neurosci. 22 1521–1528. 10.1111/j.1460-9568.2005.04315.x [DOI] [PubMed] [Google Scholar]
- Schonwiesner M., Zatorre R. J. (2009). Spectro-temporal modulation transfer function of single voxels in the human auditory cortex measured with high-resolution fMRI. Proc. Natl. Acad. Sci. U.S.A. 106 14611–14616. 10.1073/pnas.0907682106 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shamma S. (2001). On the role of space and time in auditory processing. Trends Cogn. Sci. 5 340–348. 10.1016/S1364-6613(00)01704-6 [DOI] [PubMed] [Google Scholar]
- Shannon R. V., Zeng F.-G., Kamath V., Wygonski J., Ekelid M. (1995). Speech recognition with primarily temporal cues. Science 270 303–304. 10.1126/science.270.5234.303 [DOI] [PubMed] [Google Scholar]
- Sihvonen A. J., Särkämö T., Rodríguez-Fornells A., Ripollés P., Münte T. F., Soinila S. (2019). Neural architectures of music–insights from acquired amusia. Neurosci. Biobehav. Rev. 107 104–114. [DOI] [PubMed] [Google Scholar]
- Simoncelli E. P., Olshausen B. A. (2001). Natural image statistics and neural representation. Annu. Rev. Neurosci. 24 1193–1216. [DOI] [PubMed] [Google Scholar]
- Singh N. C., Theunissen F. E. (2003). Modulation spectra of natural sounds and ethological theories of auditory processing. J. Acoust. Soc. Am. 114 3394–3411. 10.1121/1.1624067 [DOI] [PubMed] [Google Scholar]
- Tierney A., Dick F., Deutsch D., Sereno M. (2013). Speech versus song: Multiple pitch-sensitive areas revealed by a naturally occurring musical illusion. Cereb. Cortex 23 249–254. 10.1093/cercor/bhs003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tillmann B., Lévêque Y., Fornoni L., Albouy P., Caclin A. (2016). Impaired short-term memory for pitch in congenital amusia. Brain Res. 1640 251–263. 10.1016/j.brainres.2015.10.035 [DOI] [PubMed] [Google Scholar]
- Venezia J. H., Thurman S. M., Richards V. M., Hickok G. (2019). Hierarchy of speech-driven spectrotemporal receptive fields in human auditory cortex. Neuroimage 186 647–666. 10.1016/j.neuroimage.2018.11.049 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Woolley S. M. N., Fremouw T. E., Hsu A., Theunissen F. E. (2005). Tuning for spectro-temporal modulations as a mechanism for auditory discrimination of natural sounds. Nat. Neurosci. 8 1371–1379. 10.1038/nn1536 [DOI] [PubMed] [Google Scholar]
- Zatorre R. J. (1988). Pitch perception of complex tones and human temporal-lobe function. J. Acoust. Soc. Am. 84 566–572. 10.1121/1.396834 [DOI] [PubMed] [Google Scholar]
- Zatorre R. J., Baum S. R. (2012). Musical melody and speech intonation: Singing a different tune. PLoS Biol. 10:e1001372. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zatorre R. J., Belin P. (2001). Spectral and temporal processing in human auditory cortex. Cereb. Cortex 11 946–953. 10.1093/cercor/11.10.946 [DOI] [PubMed] [Google Scholar]
- Zatorre R. J., Belin P., Penhune V. B. (2002). Structure and function of auditory cortex: Music and speech. Trends Cogn. Sci. 6 37–46. [DOI] [PubMed] [Google Scholar]
- Zatorre R. J., Delhommeau K., Zarate J. M. (2012). Modulation of auditory cortex response to pitch variation following training with microtonal melodies. Front. Psychol. 3:544. 10.3389/fpsyg.2012.00544 [DOI] [PMC free article] [PubMed] [Google Scholar]