Skip to main content
The Journal of Neuroscience logoLink to The Journal of Neuroscience
. 2018 Jul 18;38(29):6442–6444. doi: 10.1523/JNEUROSCI.0988-18.2018

Delving Deep into Crossmodal Integration

N Alex Cayco-Gajic 1,, Yann Sweeney 2
PMCID: PMC6052241  PMID: 30021764

“Don't you wonder sometimes

'Bout sound and vision?”

—David Bowie

To create a coherent representation of the world, the brain must consolidate information across different sensory modalities. This process, called multisensory integration, is key for the meaningful perception of objects and experiences (Maddox et al., 2015). Consider, for instance, how disconcerting it is to watch a film in which the audio and video are slightly out of sync. Traditionally, it was believed that information from multiple senses was integrated in higher cortical areas, after being processed independently in primary sensory areas. This view has recently been challenged by anatomical and functional evidence of crossmodal integration at even the earliest stages of cortical processing (Ghazanfar and Schroeder, 2006).

What is the computational advantage of multisensory enhancement in primary sensory cortex? Recent imaging studies in mouse visual cortex have shown that concurrent auditory stimuli can enhance visual coding by sharpening tuning and modulating firing rates (Ibrahim et al., 2016; Meijer et al., 2017). Moreover, activating auditory and somatosensory cortices elicit similar responses in visual cortex, indicating that the mechanism behind crossmodal integration may be broadly similar across non-primary modalities (Iurilli et al., 2012). There is also considerable evidence of visual and somaesthetic modulation of activity in auditory cortex (for review, see Ghazanfar and Schroeder, 2006). In this case, however, several basic questions remain unanswered, including: what nonauditory features are represented in auditory cortical neurons, how is that information integrated into local cortical circuits, and what effect does this have on auditory processing? To address these questions, and to understand the functional role of crossmodal integration more generally, further interrogation of the circuit mechanism is needed. In a recent article in The Journal of Neuroscience, Morrill and Hasenstaub (2018) took a step toward answering these questions by probing the laminar dependence of visual responsiveness in auditory cortex.

Morrill and Hasenstaub (2018) recorded extracellularly from the auditory cortex of awake mice while presenting either auditory (tone) or visual (flash) stimuli. They observed visually-evoked increases of firing rate in 58% of recordings, in both primary and secondary cortical areas, as judged by frequency tuning and auditory response latencies. The use of laminar probes allowed the authors to isolate the effect in different layers, revealing that the significant majority of visual responses occurred in infragranular layers, with minimal responses in L1–L4.

These findings are timely, as they allow direct comparison with several recent experiments that, by contrast, investigate auditory responses in visual cortex. This comparison reveals a functional asymmetry in audiovisual integration in visual and auditory areas. In mouse primary visual cortex, tones and bursts of noise elicit responses in supragranular as well as infragranular layers (but not in L4; Iurilli et al., 2012). In particular, up to 10% of L2/3 neurons respond to tone presentation alone (Meijer et al., 2017). Conversely, Morrill and Hasenstaub (2018) report that <1% of multiunits in L2/3 of auditory cortex were visually responsive.

The strong functional asymmetry between visual and auditory cortex likely stems from a difference in crossmodal input. To understand the source of this difference, it is necessary to identify the main pathways of visual information in auditory cortex. There are three possible pathways: top-down connections from higher-order multisensory areas, connections from thalamus (either visual or multisensory regions), and lateral connections from visual cortex. In rodents, anatomical connections have been observed from all three of these candidates (Ghazanfar and Schroeder, 2006; Banks et al., 2011; Tyll et al., 2011). In the case of lateral connections, there is a striking imbalance between auditory and visual corticocortical projections. For example, a recent tracing study in mice found that, despite significant projections from primary auditory cortex to primary visual cortex, projections in the reverse direction were absent (Ibrahim et al., 2016). Instead, auditory cortex receives input from secondary visual cortex (Banks et al., 2011). Even from these areas, however, an overall asymmetry is apparent. A quick calculation from the Allen Mouse Brain Connectivity Atlas reveals that auditory cortical regions send a greater fraction (by an order of magnitude) of their outgoing projections to visual regions than the converse (Oh et al., 2014). Moreover, the timing of visual responses in auditory cortex is not fast enough to implicate direct projections from early visual cortex as the predominant channel for visual information. For example, Iurilli et al. (2012) reported that activation of auditory cortex elicited responses in visual cortex with a latency of 6 ms. In contrast, Morrill and Hasenstaub (2018) measured visually evoked response latencies of multiunits in both auditory cortex (90 ms) and visual cortex (40 ms). This delay is considerably longer than expected for a monosynaptic connection, suggesting that visual information may be coming primarily from multisensory corticothalamic or higher cortical inputs, at least in mice.

To determine what kind of visual information is integrated in auditory cortex, Morrill and Hasenstaub (2018) repeated their recordings in auditory cortex while presenting drifting gratings of varying orientation. Visually responsive single units were significantly less orientation selective than units in visual cortex, suggesting that these units primarily signaled the timing and presence of a visual stimulus, as opposed to specific visual features. This finding supports the idea that timing is particularly important for crossmodal integration. Indeed, recent studies have demonstrated that temporally congruent auditory and visual stimuli (i.e., having the same temporal frequency) preferentially modulate activity in both ferret auditory cortex (Atilgan et al., 2018) and mouse visual cortex (Meijer et al., 2017) compared with incongruent stimuli. Furthermore, it has recently been demonstrated that projections from auditory cortex to primary visual cortex are dominated by neurons that encode the abrupt onset of sounds (Deneux et al., 2018). However, these recent studies contrast with classic electrophysiological studies, which found evidence of precise frequency and spatial information about auditory stimuli in the visual cortex of cats (Spinelli et al., 1968; Fishman and Michael, 1973). One explanation for this disparity may be the fact that cats have more advanced visual processing compared with rodents. Another possibility is that visual responses in mouse auditory cortex contain information about more complex visual stimuli than the gratings tested by Morrill and Hasenstaub (2018). This may be expected considering that mouse auditory cortex receives direct projections from secondary visual cortex (Banks et al., 2011). However, what visual features these regions represent in mice is unknown (Glickfeld and Olsen, 2017).

Finally, a key result of Morrill and Hasenstaub (2018) is that visual information in auditory cortex was almost exclusively found in infragranular layers, especially in L6. This finding shines a light onto the mysterious role of deep layer neurons. In comparison with their more superficial counterparts, less is known about how L6 neurons contribute to sensory processing. This is due in part to the technical difficulty of accessing deep layers, as well as to the heterogeneous morphologies and unusual response properties of these neurons. Previous work in primary auditory cortex of rats (Sakata and Harris, 2009; Zhou et al., 2010) and cats (Atencio et al., 2009) found that L6 pyramidal cells are less feature selective than cells in superficial layers with complex receptive fields and little stimulus information. These properties have made it difficult to understand the role of L6 neurons for representing auditory stimuli. Although there are likely to be cross-species differences, the findings of Morrill and Hasenstaub (2018) may explain these results by pointing to a more complex role for L6 beyond unimodal auditory processing.

The discovery of a subpopulation of visually-responsive cells in L6 suggests that this layer may serve as a gateway for contextual information from other modalities. Two recent studies in V1 further support this hypothesis. Vélez-Fort et al. (2018) found that L6 pyramidal cells could convey head velocity signals inherited via a direct connection from retrosplenial cortex. Similarly, Leinweber et al. (2017) found that L6 received predictive signals about expected visual flow from motor cortex. Morrill and Hasenstaub (2018) complement these studies by showing that audiovisual integration also takes place in L6 of auditory cortex. Crossmodal integration in L6 could therefore be used to control auditory processing based on nonauditory contextual signals, as L6 of visual cortex has previously been shown to perform gain control on superficial populations without changing their preferred orientation (Olsen et al., 2012). More recently, it has been shown that optogenetic activation of L6 of auditory cortex modulates auditory tuning, and that this could control a tradeoff between sound detection and discrimination performance (Guo et al., 2017). Importantly, this behavioral enhancement was highly dependent on the timing between sensory stimulation and L6 spiking. Combined with the results of Morrill and Hasenstaub (2018), this suggests that visual timing information in L6 may enhance auditory processing. Intriguingly, these studies all target the same Ntsr1-Cre transgenic mouse line, in which Cre-expression is limited to L6 corticothalamic neurons (Sundberg et al., 2018). These findings together suggest the possibility that a population of L6 pyramidal cells perform a crucial role by modulating early sensory processing to generate coherent sensory representations.

The recent burst of work on multisensory enhancement in sensory cortex and on L6 pyramidal cells make this an exciting time for unraveling the circuit mechanism underlying crossmodal integration. Morrill and Hasenstaub (2018) have made a key contribution by revealing the laminar specificity of visual information in auditory cortex. Future studies could help to tease apart the circuit mechanisms of crossmodal integration even further; for example, by dissecting the role of local deep-layer inhibitory circuits. New techniques for large-scale characterization of long-range projections will also clarify how crossmodal information is transmitted between regions (Han et al. 2018). Another open question is whether crossmodal signals can be enhanced by multimodal behavioral tasks. In particular, it would be valuable to investigate whether animals trained to detect specific audiovisual combinations develop tuned visual responses in auditory cortex. Evidence from sensory deprivation experiments hints that such a substrate exists for expressing crossmodal plasticity (Bavelier and Neville, 2002). Ultimately, determining the circuit mechanisms behind crossmodal integration will lead neuroscience further toward understanding naturalistic behavior in dynamic, multisensory environments.

Footnotes

Editor's Note: These short reviews of recent JNeurosci articles, written exclusively by students or postdoctoral fellows, summarize the important findings of the paper and provide additional insight and commentary. If the authors of the highlighted article have written a response to the Journal Club, the response can be found by viewing the Journal Club at www.jneurosci.org. For more information on the format, review process, and purpose of Journal Club articles, please see http://jneurosci.org/content/preparing-manuscript#journalclub.

The authors declare no competing financial interests.

References

  1. Atencio CA, Sharpee TO, Schreiner CE (2009) Hierarchical computation in the canonical auditory cortical circuit. Proc Natl Acad Sci U S A 106:21894–21899. 10.1073/pnas.0908383106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Atilgan H, Town SM, Wood KC, Jones GP, Maddox RK, Lee AKC, Bizley JK (2018) Integration of visual information in auditory cortex promotes auditory scene analysis through multisensory binding. Neuron 97:640–655.e4. 10.1016/j.neuron.2017.12.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Banks MI, Uhlrich DJ, Smith PH, Krause BM, Manning KA (2011) Descending projections from extrastriate visual cortex modulate responses of cells in primary auditory cortex. Cereb Cortex 21:2620–2638. 10.1093/cercor/bhr048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bavelier D, Neville HJ (2002) Cross-modal plasticity: where and how? Nat Rev Neurosci 3:443–452. 10.1038/nrn848 [DOI] [PubMed] [Google Scholar]
  5. Deneux T, Kempf A, Bathellier B (2018) Context-dependent signaling of coincident auditory and visual events in primary visual cortex. BioRxiv. Advance online publication. Retrieved April 4, 2018. doi: 10.1101/258970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Fishman MC, Michael P (1973) Integration of auditory information in the cat visual cortex. Vision Res 13:1415–1419. 10.1016/0042-6989(73)90002-3 [DOI] [PubMed] [Google Scholar]
  7. Ghazanfar AA, Schroeder CE (2006) Is neocortex essentially multisensory? Trends Cogn Sci 10:278–285. 10.1016/j.tics.2006.04.008 [DOI] [PubMed] [Google Scholar]
  8. Glickfeld LL, Olsen SR (2017) Higher-order areas of the mouse visual cortex. Annu Rev Vis Sci 3:251–273. 10.1146/annurev-vision-102016-061331 [DOI] [PubMed] [Google Scholar]
  9. Guo W, Clause AR, Barth-Maron A, Polley DB (2017) A corticothalamic circuit for dynamic switching between feature detection and discrimination. Neuron 95:180–194e5. 10.1016/j.neuron.2017.05.019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Han Y, Kebschull JM, Campbell RA, Cowan D, Imhof F, Zador AM, Mrsic-Flogel TD (2018) The logic of single-cell projections form visual cortex. Nature 556:51–56. 10.1038/nature26159 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Ibrahim L, Mesik L, Ji XY, Fang Q, Li HF, Li YT, Zingg B, Zhang LI, Tao HW (2016) Cross-modality sharpening of visual cortical processing through layer-1-mediated inhibition and disinhibition. Neuron 89:1031–1045. 10.1016/j.neuron.2016.01.027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Iurilli G, Ghezzi D, Olcese U, Lassi G, Nazzaro C, Tonini R, Tucci V, Benfenati F, Medini P (2012) Sound-driven synaptic inhibition in primary visual cortex. Neuron 73:814–828. 10.1016/j.neuron.2011.12.026 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Leinweber M, Ward DR, Sobczak JM, Attinger A, Keller GB (2017) A sensorimotor circuit in mouse cortex for visual flow predictions. Neuron 96:1204. 10.1016/j.neuron.2017.11.009 [DOI] [PubMed] [Google Scholar]
  14. Maddox RK, Atligan H, Bizley JK, Lee AK (2015) Auditory selective attention is enhanced by a task-irrelevant temporally coherent visual stimulus in human listeners. eLife 4:e04995. 10.7554/eLife.04995 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Meijer GT, Montijn JS, Pennartz CM, Lansink CS (2017) Audiovisual modulation in mouse primary visual cortex depends on cross-modal stimulus configuration and congruency. J Neurosci 37:8783–8796. 10.1523/JNEUROSCI.0468-17.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Morrill RJ, Hasenstaub AR (2018) Visual information present in infragranular layers of mouse auditory cortex. J Neurosci 38:2854–2862. 10.1523/JNEUROSCI.3102-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Oh SW, Harris JA, Ng L, Winslow B, Cain N, Mihalas S, Wang Q, Lau C, Kuan L, Henry AM, Mortrud MT, Ouellette B, Nguyen TN, Sorensen SA, Slaughterbeck CR, Wakeman W, Li Y, Feng D, Ho A, Nicholas E, et al. (2014) A mesoscale connectome of the mouse brain. Nature 508:207–214. 10.1038/nature13186 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Olsen SR, Bortone DS, Adesnik H, Scanziani M (2012) Gain control by layer six in cortical circuits of vision. Nature 483:47–52. 10.1038/nature10835 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Sakata S, Harris KD (2009) Laminar structure of spontaneous and sensory-evoked population activity in auditory cortex. Neuron 64:404–418. 10.1016/j.neuron.2009.09.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Spinelli DN, Starr A, Barrett TW (1968) Auditory specificity in unit recordings from cat's visual cortex. Exp Neurol 22:75–84. 10.1016/0014-4886(68)90020-4 [DOI] [PubMed] [Google Scholar]
  21. Sundberg SC, Lindström SH, Sanchez GM, Granseth B (2018) Cre-expressing neurons in visual cortex of Ntsr1-cre GN220 mice are corticothalamic and are depolarized by acetylcholine. J Comp Neurol 526:120–132. 10.1002/cne.24323 [DOI] [PubMed] [Google Scholar]
  22. Tyll S, Budinger E, Noesselt T (2011) Thalamic influences on multisensory integration. Commun Integr Biol 4:378–381. 10.4161/cib.15222 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Vélez-Fort M, Bracey EF, Keshavarzi S, Rousseau CV, Cossell L, Lenzi SC, Strom M, Margrie TW (2018) A circuit for integration of head- and visual-motion signals in layer 6 of mouse primary visual cortex. Neuron 98:179–191.e6. 10.1016/j.neuron.2018.02.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Zhou Y, Liu BH, Wu GK, Kim YJ, Xiao Z, Tao HW, Zhang LI (2010) Preceding inhibition silences layer 6 neurons in auditory cortex. Neuron 65:706–717. 10.1016/j.neuron.2010.02.021 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from The Journal of Neuroscience are provided here courtesy of Society for Neuroscience

RESOURCES