Abstract
Sensory pathways are typically studied by starting at receptor neurons and following postsynaptic neurons into the brain. However, this leads to a bias in analyses of activity toward the earliest layers of processing. Here, we present new methods for volumetric neural imaging with precise across-brain registration to characterize auditory activity throughout the entire central brain of Drosophila and make comparisons across trials, individuals and sexes. We discover that auditory activity is present in most central brain regions and in neurons responsive to other modalities. Auditory responses are temporally diverse, but the majority of activity is tuned to courtship song features. Auditory responses are stereotyped across trials and animals in early mechanosensory regions, becoming more variable at higher layers of the putative pathway, and this variability is largely independent of ongoing movements. This study highlights the power of using an unbiased, brain-wide approach for mapping the functional organization of sensory activity.
Acentral problem in neuroscience is determining how information from the outside world is represented by the brain. Solving this problem requires knowing which brain areas and neurons encode particular aspects of the sensory world, how representations are transformed from one layer to the next and the degree to which these representations vary across presentations and individuals. Most sensory pathways have been successfully studied by starting at the periphery—the receptor neurons—and then characterizing responses along postsynaptic partners in the brain. We therefore know the most about the earliest stages of sensory processing in most model organisms. However, even for relatively simple brains, there exists a significant amount of cross-talk between brain regions1,2. Thus, neurons and brain areas may be more multimodal than appreciated. In this article, we present methods to systematically characterize sensory responses throughout the Drosophila brain, for which abundant neural circuit tools facilitate moving from activity maps to the targeting of identifiable neurons3.
We focus on auditory coding in the Drosophila brain because, despite its relevance to courtship behavior, surprisingly little is known regarding how auditory information is processed downstream of primary mechanosensory neurons. In addition, in contrast to olfaction or vision, for which the range of odors or visual patterns that flies experience in naturalistic conditions is large, courtship song is the main auditory cue to which flies respond and comprises a narrow range of species-specific features that are easily probed in experiments4. Drosophila courtship behavior unfolds over many minutes, and females must extract features from male songs to inform mating decisions5,6, whereas males listen to songs to inform their courtship decisions in group settings7. Fly song has only three modes, two types of ‘pulse song’ and one ‘sine song’8, and males both alternate between these modes to form song bouts and modulate song intensity based on sensory feedback9,10.
Flies detect sound using a feathery appendage of the antenna, the arista, which vibrates in response to near-field sounds11. Antennal displacements activate mechanosensory Johnston’s organ neurons (JONs) housed within the antenna. Three major populations of JONs (A, B and D) respond to vibratory stimuli at frequencies found in natural courtship song12,13. These neurons project to distinct areas of the antennal mechanosensory and motor center (AMMC) in the central brain. Recent studies suggest that the auditory pathway continues from the AMMC to the wedge (WED), then to the ventrolateral protocerebrum (VLP) and to the lateral protocerebral complex (LPC)4,14–19. However, our knowledge of the fly auditory pathway remains incomplete, and the functional organization of regions downstream of the AMMC and WED are largely unexplored. Moreover, nearly all studies of auditory coding in Drosophila have been performed using female brains, even though both males and females process courtship song information4.
To address these issues, we developed methods to investigate the representation of behaviorally relevant auditory signals throughout the central brain of Drosophila and to make comparisons across animals. We use two-photon microscopy to sequentially target the entirety of the Drosophila central brain in vivo, combined with fully automated segmentation of regions of interest (ROIs)20. In contrast to recent brain-wide imaging studies of Drosophila1,21, we traded off temporal speed for enhanced spatial resolution. Imaging at high spatial resolution facilitates automated ROI segmentation, with each ROI covering subneuropil structures, including cell bodies and neurites. ROIs were accurately registered into an in vivo template brain to compare activity across trials, individuals and sexes, and to build comprehensive maps of auditory activity throughout the central brain. Our results reveal that the representation of auditory signals is broadly distributed throughout 33 out of 36 major brain regions, including in regions known to process other sensory modalities, such as all levels of the olfactory pathway, or to drive various motor behaviors, such as the central complex. The representation of auditory stimuli is diverse across brain regions, but focused on conspecific features of courtship song. Auditory activity is more stereotyped (across trials and individuals) at early stages of the putative mechanosensory pathway, becoming more variable and more selective for particular aspects of the courtship song at higher stages. This variability cannot be explained by simultaneous measurements of ongoing fly behavior. Meanwhile, auditory maps are largely similar between male and female brains, despite extensive sexual dimorphisms in neuronal number and morphology. These findings provide the first brain-wide description of sensory processing and feature tuning in Drosophila.
Results
A new pipeline for mapping sensory activity throughout the central brain of Drosophila.
We developed a pipeline to volumetrically image sensory responses (via the pan-neuronal expression of GCaMP6s) and then precisely register sequentially imaged volumes into an in vivo template (Fig. 1a and Extended Data Fig. 1a). We presented three distinct auditory stimuli, with the first two representing the major modes of Drosophila melanogaster courtship song (pulse and sine). These stimuli should drive auditory neurons but not neurons sensitive to slow vibrations, such as those induced by wind or gravity12,13. For each fly, we imaged 14–17 subvolumes at 1 Hz (voxel size of 1.2 × 1.4 × 2 μm3) until the whole extent of the z axis per fly (posterior to anterior) and one-quarter of the central brain in the x–y axis were covered, what we refer to hereafter as a ‘volume’. These volumes were time-series motion-corrected and stitched along the posterior–anterior axis using a secondary signal22 (Fig. 1a, step (1)). We imaged volumes from either dorsal or ventral quadrants, providing full hemisphere coverage by imaging only two flies (Fig. 1b). We mirrored each volume such that all dorsal or ventral volumes imaged could be compared in the same reference brain coordinates (see below); this is why all activity maps appear on half of the brain (for example, see Extended Data Fig. 1b).
To automate the segmentation of ROIs throughout the brain, we used a constrained non-negative matrix factorization algorithm optimized for three-dimensional (3D) time-series data20,23 (Methods and Fig. 1a, step (2)). This method groups contiguous pixels that are similar in their temporal responses; thus, each ROI is a functional unit. We remained agnostic as to the specific relationship between ROIs and neurons, whereby a single ROI could represent activity from many neurons or a single neuron could be segmented into multiple ROIs if different neuronal compartments differ in their activity. Importantly, this method enables the effective segmentation of responses from neuropil (axons, dendrites and processes), not only somas, and this is critical for capturing sensory responses in Drosophila, for which somas may not reflect activity in processes24. We typically extracted thousands of ROIs per volume and collectively sampled activity from all central brain regions (Fig. 1c and Extended Data Fig. 1b,c). These ROIs included both spontaneous and stimulus-driven activity (Fig. 1c), whereby stimulus-driven ROIs were identified using quantitative criteria (Methods, Fig. 1a, step (3), and Extended Data Fig. 1d,e). These criteria selected ROIs with responses consistent across trials (Fig. 1d), but see below for the analysis of trial-to-trial variation even given these criteria. We found hundreds of auditory ROIs per fly ranging from 1.2% to 25.4% of all segmented ROIs per fly (Extended Data Fig. 1f).
To compare activity across individuals and sexes, we next registered functionally imaged brain segments into an in vivo intersex atlas (IVIA) using the computational morphometry (CMTK) toolkit (Methods, Fig. 1a, step (4), and Extended Data Figs. 1a and 2a). Registration of individual fly brains to our IVIA was at cellular accuracy (~4.3 μm across all tracts; Extended Data Fig. 2b). Using this approach, we successfully registered volumes from 45 out of 48 flies (see examples in Extended Data Fig. 2c).
The Insect Brain Nomenclature Whole Brain (IBNWB) Drosophila fixed brain atlas25 contains detailed brain neuropil and neurite tract segmentation information. By linking our IVIA template to the IBNWB atlas, we could map auditory activity onto known brain neuropils and processes (Fig. 1f,g and Supplementary Table 1) and thereby relate it to the entire network of D. melanogaster anatomical repositories3. We built a bridge registration between IVIA and the IBNWB atlas using a point set registration algorithm26 with high accuracy (Extended Data Fig. 2d–g).
In summary, we developed a novel open-source pipeline for monitoring, anatomically annotating and directly comparing sensory-driven activity, at high spatial resolution, across trials, individuals and sexes.
Auditory activity is widespread across the central brain.
We extracted 19,036 auditory ROIs from 33 female and male brains (Fig. 2a and Extended Data Fig. 3a). Within a single fly, auditory ROIs were broadly distributed across the anterior–posterior and dorsal–ventral axes (Fig. 2b and Supplementary Videos 1–4). Probability density maps across flies using the spatial components of the ROIs (Methods) revealed across-individual (Fig. 2c) and across-sex (Extended Data Fig. 3b) consistency. A total of 77% of all segmented ROIs were assigned to an identifiable neuropil or tract (Fig. 2d–h, Extended Data Fig. 3c,d and Supplementary Video 5). Surprisingly, 33 out of 36 central brain neuropils contained robust auditory activity (Methods). In addition, testing other song-relevant stimuli confirmed that we did not miss auditory responses that were narrowly selective to additional frequencies and patterns not present in our sine and pulse stimuli (Extended Data Fig. 3e–l).
We found auditory activity in all neuropils innervated by neurons previously described as auditory or mechanosensory4,6,14–16,18,27–29; that is, in the AMMC, the saddle (SAD), the WED, the anterior VLP (AVLP), the gnathal ganglion (GNG) and the inferior and superior clamp, which we herein refer to as ‘expected auditory’ neuropils (Fig. 2d,e). Similarly, we detected auditory ROIs in many tracts or commissures (the great commissure, the wedge commissure, and the inferior and superior saddle commissure) connecting expected auditory neuropils (Fig. 2g and Extended Data Fig. 3m,n).
Surprisingly, we found that a number of neuropils and tracts outside the expected auditory neuropils contained a large percentage of auditory responses (fraction of voxels with auditory activity (Fig. 2e)). For example, neuropils of the olfactory pathway (the antennal lobe (AL), the medial antennal lobe tract containing olfactory projection neurons (PNs), the mushroom body (MB), including the peduncle, lobes and calyx, and the lateral horn (LH)), parts of the visual pathway (the posterior VLP (PVLP) and the posterior lateral protocerebrum (PLP)), and the central complex (the ellipsoid body, the fan-shaped body, the protocerebral bridge, the bulb and the lateral accessory lobe); such neuropils are referred to as ‘new auditory’ neuropils (Fig. 2e,h). We confirmed that activity within the olfactory pathway was present within the somas of individual AL PNs and MB Kenyon cells (Extended Data Fig. 4a–d). Some of this new activity was in neuropils known to be downstream targets of expected auditory neuropils, but never previously studied functionally (for example, the inferior posterior slope and the gorget17). These results demonstrate that auditory-evoked activity is widespread throughout the brain and extends beyond expected auditory neuropils, including in the olfactory, visual and pre-motor pathways.
Widespread auditory activity originates in mechanosensory receptor neurons.
To determine how auditory information makes its way to so many neuropils, we used the new ‘hemibrain’ connectomic resource30. Mechanosensory receptor neurons (JONs) project to the AMMC/SAD, and AMMC/SAD neurons connect with the WED14,17. We determined the primary and secondary target neuropils for all AMMC/SAD and WED neurons identified in the hemibrain. This analysis revealed that neurons that originate in the AMMC/SAD or WED directly connect with neurons in 22 or 24 neuropils, and with one additional synapse, connect with neurons in 32 or 33 neuropils, respectively (Fig. 2i).
To confirm that this activity was in fact driven by mechanosensory receptor neurons, we imaged flies carrying the iav1 mutation, which renders flies deaf31,32. This manipulation resulted in a 99% reduction in activity (Extended Data Fig. 4e,f). Response magnitudes of the small number of auditory ROIs found in iav1 flies were substantially reduced compared with wild-type flies (Extended Data Fig. 4g). We therefore conclude that the overwhelming majority of the auditory activity we mapped with our stimuli are transduced through the mechanoreceptor neurons.
Auditory activity throughout the central brain is characterized by diverse temporal responses.
We next hierarchically clustered responses from auditory ROIs based on their temporal profiles (Methods); this process identified 18 distinct stimulus-locked response types (Fig. 3a). These same 18 response types were present in both male and female brains (Extended Data Fig. 5a), with only minor differences between the sexes (Extended Data Fig. 5b,c)—we therefore pooled male and female auditory ROIs for subsequent analyses. We measured the frequency of response types for each neuropil to evaluate differences in auditory representations by neuropil (Fig. 3b). Auditory activity was diverse throughout the central brain, which consisted of both inhibitory (types 1–3) and excitatory (types 5–18) responses, along with diversity in response kinetics (Fig. 3d–f). Response types also differed in their selectivity for auditory stimuli (Fig. 3g).
The diversity of response types (Fig. 3b) suggests that our method samples activity from many different neurons per neuropil. In addition, each response type had a distinct spatial distribution (Extended Data Fig. 5d,e). When we measured the entropy of response types per neuropil as a proxy for the diversity of auditory activity, responses were less diverse in the earliest mechanosensory areas and became more diverse at higher levels of the putative pathway (Figs. 2d and 3c). Response diversity was highest in new auditory areas, such as the PLP, the LH and the superior lateral protocerebrum. Clustering the data in Fig. 3b revealed neuropils with highly similar response profiles, which suggests that there is functional interconnectivity (Extended Data Fig. 6a).
Within the expected auditory neuropils (Fig. 2d), we found that the AMMC contained predominantly nonselective (type 13) and sine-preferring (type 18) response types, but that brain regions downstream of the AMMC showed changes in the distribution of response types, such as a decrease in nonselective response types and an increase in both noise-preferring and pulse-preferring response types (Fig. 3b). Prior work on individual neurons innervating the AVLP suggested that this brain area mostly encodes pulse-like stimuli6,14,16,19. In contrast, we found strong representation of both sine-preferring and pulse-preferring response types, with responses becoming more narrowly tuned in this region (for example, response types 7 and 11). In the lateral junction (the inferior and superior clamp), as expected from innervations of pulse-tuned neurons4,19, there were a number of pulse-preferring response types (types 11 and 12). However, the dominant response types were sine-preferring.
In the new auditory neuropils, we found that the olfactory system was dominated by nonselective (responding equivalently to pulse and sine stimuli) and sine-preferring response types (Fig. 3b and Extended Data Fig. 4a–d); however, the response types became more diverse in higher-order olfactory regions (Fig. 3c). In the visual neuropils (the PVLP, the PLP and the anterior optic tubercle), auditory activity was highly compartmentalized. For example, in the PVLP, activity was concentrated in the vicinity of ventromedial optic glomeruli, while in the PLP, auditory ROIs formed a single discrete compartment—minimally overlapping with optic glomeruli—that runs from the anterior WED–PLP boundary to a more posterior PLP–LH boundary (Extended Data Fig. 6b). In the central complex (the ellipsoid body, the fan-shaped body and the protocerebral bridge) we found primarily nonselective and inhibitory response types (type 1), while the lateral complex (the bulb and the lateral accessory lobe) contained both sine-tuned (types 6 and 18) and nonselective excitatory response types (types 13 and 16) (Fig. 3b).
In summary, auditory activity across the central brain comprises at least 18 distinct response types, with both expected and new auditory neuropils showing a diversity of response types and with responses becoming more stimulus-selective in deeper brain neuropils.
Widespread auditory activity is centered on features of conspecific courtship songs.
To investigate tuning for courtship song features (Fig. 4a), we focused on the neuropils with strongest auditory activity (Figs. 2e and Fig. 4b) and imaged subregions of each of these neuropils (Extended Data Fig. 7a) to sample ROIs with varying degrees of pulse or sine selectivity. Tuning curves were formed for each ROI and for each set of stimuli (Extended Data Fig. 7b). These tuning curves could be clustered into seven tuning types (Fig. 4c,d), but auditory responses were still divided into three main categories (pulse-tuned (tuning type 1), sine-tuned (tuning type 5–7) or nonselective (tuning type 2–4); compare with Fig. 3), thereby validating our initial choice of stimuli.
Sine-tuned ROIs were either low pass (tuning type 5–6: strongest responses to 100 or 150 Hz sines, matching conspecific sine song frequencies) or had a broader distribution of preferred carrier frequencies (tuning type 7) (Fig. 4e). Pulse-tuned ROIs responded best to pulse frequencies, pauses and durations present in natural courtship song (Fig. 4e,f). Sine-tuned ROIs preferred continuous stimuli, which explains their preference for pulses with short pauses (below 20 ms) and longer pulse durations. Most tuning types were sensitive to the broad range of intensities tested (0.5–5 mm s−1), with proportional increases in response to stimulus magnitude, and responded best to longer pulse train or sine tone durations (4 s or longer) (Extended Data Fig. 7d,e). Although, during natural courtship, pulse or sine trains (stretches of each type of song) have a mode of ~360 ms and rarely last longer than 4 s9, preference for unnaturally long bouts have been recently reported4. Finally, the distribution of tuning types shifted along the pathway from the AMMC to the six additional neuropils examined here (Fig. 4d).
Drosophila melanogaster pulse song is a sine tone (carrier frequency of ~250 Hz) with an amplitude modulation at ~27 Hz, whereas sine song is an unmodulated sinusoid of ~150 Hz. How do carrier frequency and envelope interact when it comes to tuning for pulse and sine stimuli? To address this question, we used a novel set of stimuli. We added an ~27-Hz envelope to pure tones of 150 or 250 Hz carrier frequency at different strengths (Methods). Responses tuned to pulses increased their response with the strength of the envelope (in other words, as a sine stimulus became more pulse-like), and the magnitude of this enhancement depended on the carrier frequency (Fig. 4g). Similarly, responses tuned to sine tones were strongly attenuated by the presence of an envelope, with different sensitivities by tuning type. These experiments suggest that there is a strong categorical boundary between pulse and sine stimuli among neural responses, with little evidence for selectivity for intermediate envelopes and frequencies.
Evaluating auditory response stereotypy across both trials and individuals.
It is not known how reliable sensory responses are across trials, individuals and sexes for most sensory modalities and brain areas. Our method of precise registration of volumetrically imaged auditory activity enabled us to systematically evaluate response variability across different neuropils and tracts. We first measured the residual of the variance explained by the mean across trials for all auditory ROIs (Methods). Our criteria for selecting auditory ROIs (Extended Data Fig. 1d–f) excluded ROIs that did not show consistent responses across trials to at least one auditory stimulus. Nonetheless, we still observed a range of across-trial variabilities for auditory ROIs (Fig. 5a), with a pattern across neuropils. Early mechanosensory neuropils (the AMMC, the SAD, the WED and the AVLP) contained the lowest across-trial response variance, while all other neuropils exhibited higher trial-to-trial variability (Fig. 5b). This variation was, to some degree, correlated with the magnitude of auditory responses across neuropils (Extended Data Fig. 8a).
We showed above that particular response types (Fig. 3a) are found in consistent spatial locations across flies (Extended Data Fig. 5e), which suggests that there is a high degree of across-individual stereotypy of auditory responses. To quantify this, we used a similarity index (Methods) to compare activity profiles across individuals (Fig. 4c,d and Extended Data Fig. 9). Some neuropils, like the GNG, contained ROIs with temporal responses highly similar across individuals (for example, all individuals have a high frequency of response type 18 in the GNG), whereas other neuropils, like the PLP, showed more variable activity across individuals. Comparing the distribution of the similarity index across all neuropils revealed that early mechanosensory neuropils also had the highest across-individual similarity in auditory responses (Fig. 4d and Extended Data Fig. 8b). Similarity decreased in downstream neuropils, and further decreased in neuropils outside the canonical mechanosensory system (for example, visual, olfactory and pre-motor neuropils). These differences in similarity across neuropils were robust against changes in how auditory responses are clustered (Extended Data Fig. 8c) and did not arise from variations in GCaMP fluorescence values across animals (Extended Data Fig. 8d).
Spontaneous movements do not account for auditory response variability.
Across-trial variability in auditory responses could arise from the time-varying behavioral state of the animal over the course of an experiment, as has been recently shown for cortical activity in mice33,34. Above, we did not track animal behavior; therefore, we could not correlate response variation with behavioral variation. To address this issue, we collected a new dataset in which we simultaneously recorded fly motion on a spherical treadmill while imaging pan-neuronally from brain regions with a range of across-trial variation from stereotyped (the AMMC and the SAD) to more variable (the AVLP, the PVLP and the LH) (Figs. 4 and 6a–e). We observed strong auditory activity, with across-trial variability similar to our previous dataset (Extended Data Fig. 10a), whereby flies showed a range of walking speeds on the ball (Extended Data Fig. 10b) that matched natural speeds during fly courtship9. Fly movements in these experiments appeared to be spontaneous because they were not reliably predicted by the auditory stimuli (Extended Data Fig. 10c). In addition, despite observing auditory activity in pre-motor neuropils (Fig. 3b), imaging from the fly ventral nerve cord (VNC) in non-walking flies did not uncover much activity (only 247 out of 39,580 ROIs recorded from the VNC were significantly correlated with any of the stimuli; Extended Data Fig. 10d–g). This suggests that auditory ROIs in our study reflect sensory responses to stimuli and not motor responses.
Because flies spontaneously moved during the presentation of auditory stimuli, this allowed us to determine what fraction of the across-trial variance in responses could be explained by the stimulus, by fly behavior or by some combination of the two. We built a linear encoding model to predict neural responses (Fig. 6f), similar to ref. 33, from a combination of the mean response to the stimulus (μ(t)) and behavior predictors (X(t)), which consisted of all tracked motor variables (Fig. 6b–e and Methods). For this analysis, we used all ROIs and not just those that passed our criteria for being auditory (Extended Data Fig. 1d). Of 63,106 ROIs segmented from the central brain, each had a different amount of explained variance from either the stimulus or behavior (Fig. 6g,h). However, the stimulus component (μ(t)) explained the majority of variance, while the behavior component (Beta × X(t)) explained a much smaller fraction (Fig. 6i); the full model (μ(t) + Beta × X(t)) was as good at explaining the variance across ROIs as the stimulus component alone (Fig. 6j). When plotting the explained variance of the stimulus component against the behavior component, we did not find a substantial trend for explained variance from the behavior predictors; that is, ROIs with little correlation to the stimulus had as much explained variance from the motor variables as ROIs with stronger correlation to the stimulus (Fig. 6k). These results suggest that the variability we observed across trials in auditory ROIs, at least for the brain regions imaged here, does not arise from fluctuations in the behavioral state of the animal.
Discussion
Sensory systems are typically studied starting from the periphery and continuing to downstream partners guided by anatomy. This has limited our understanding of sensory processing to early stages of a given sensory pathway. Here, we used a brain-wide imaging method to unbiasedly screen for auditory responses beyond the periphery and, via precise registration of recorded activity, to compare auditory representations across brain regions, individuals and sexes (Fig. 1). We found that auditory activity is widespread, extending well beyond the canonical mechanosensory pathway, and is present in brain regions and tracts known to process other sensory modalities (that is, olfaction and vision) or to drive motor behaviors (Fig. 2). The representation of auditory stimuli diversified, in terms of both temporal responses to stimuli and tuning for stimulus features, from the AMMC to later stages of the putative pathway, becoming more selective for particular aspects of courtship song (that is, sine or pulse song, and their characteristic spectrotemporal patterns) (Figs. 3 and 4). Auditory representations were more stereotypic across trials and individuals in early stages of mechanosensory processing, and more variable at later stages (Fig. 5). By recording neural activity in behaving flies, we found that fly movements accounted for only a small fraction of the variance in neural activity, which suggests that across-trial auditory response variability stems from other sources (Fig. 6). These results have important implications for how the brain processes auditory information to extract salient features and guide behavior.
Our understanding of the Drosophila auditory circuit thus far has been built up from targeted studies of neural cell types that innervate particular brain regions close to the auditory periphery14–17. Altogether, these studies have delineated a pathway that starts in the Johnston’s organ and extends from the AMMC to the WED, the VLP and the LPC. By imaging pan-neuronally, we found widespread auditory responses that spanned brain regions beyond the canonical pathway, which suggests that auditory processing is more distributed. However, for neuropil signals, it was challenging to determine the number of neurons that contribute to the ROI responses we describe. Although the diverse set of temporal and tuning types per neuropil (Fig. 3) suggests that we sampled many neurons per neuropil, restricting GCaMP to spatially restricted genetic enhancer lines (Extended Data Fig. 4c,d) will assist with linking broad functional maps with the cell types constituting them35.
Our findings of widespread auditory activity are likely not unique to audition. So far, in adult Drosophila, only taste processing has been broadly surveyed36. While that study did not map activity onto neuropils and tracts, nor did it make comparisons across individuals, it suggested that taste processing was distributed throughout the brain. Similarly, in vertebrates, widespread responses to visual and nociceptive stimuli have been observed throughout the brain37,38. Our findings are consistent with anatomical studies that found connections from the AMMC and the WED to several other brain regions17,39, and with our own analysis of the hemibrain connectome (Fig. 2). While we do not yet know what role this widespread auditory activity plays in behavior, we show that ROIs that respond to auditory stimuli do so with mostly excitatory (depolarizing) responses and that activity throughout the brain is predominantly tuned to features of the courtship song. During courtship, flies evaluate multiple sensory cues (olfactory, auditory, gustatory and visual) to inform mating decisions and to modulate their mating drive. Although integration of multiple sensory modalities has been described in higher-order brain regions40, our results suggest that song representations are integrated with olfactory and visual information at earlier stages (Figs. 2 and 3). In addition, song information may modulate the processing of non-courtship stimuli. Song representations in the MB may be useful for learning associations between song and general olfactory, gustatory or visual cues41, while diverse auditory activity throughout all regions of the LH may indicate an interaction between song processing and innate olfactory behaviors42. Finally, we found auditory activity in brain regions involved in locomotion and navigation (the central and lateral complex, and the superior and ventromedial neuropils). Auditory activity in these regions is diverse (Fig. 3), which suggests that pre-motor circuits receive information about courtship song patterns and could therefore underlie stimulus-specific locomotor responses4.
D. melanogaster songs are composed of pulses and sines that differ in their spectral and temporal properties; however, it is unclear how and where selectivity for the different song modes arises in the brain. Since neurons in the LPC are tuned for pulse song across all time scales that define that mode of song4, neurons upstream must carry the relevant information to generate such tuning. Here, we found many ROIs that are selective for either sine or pulse stimuli throughout the entire central brain (Fig. 3). A more detailed systematic examination of tuning in the neuropils that carry the most auditory activity (the AMMC, the SAD, the WED, the AVLP and PVLP, the PLP and the LH) revealed that most ROIs are tuned to either pulse or sine features, with few ROIs possessing intermediate tuning (Fig. 4). This suggests that sine and pulse information splits early in the pathway. We also found that sine-selective responses dominate throughout the brain. Although some of this selectivity may simply reflect preference for continuous versus pulsatile stimuli, our investigation of feature tuning revealed that many of these ROIs preferred frequencies that are specifically present in courtship songs (Fig. 4). Previous studies indicated that pulse song is more important for mating decisions, with sine song purported to play a role in only priming females5,43. However our results, in combination with the fact that males spend a greater proportion of time in courtship singing sine versus pulse song9, suggest a need for reevaluation of the importance and role of sine song in mating decisions. This study therefore lays the foundation for exploring how song selectivity arises in the brain.
Our results also revealed that early mechanosensory brain areas contain ROIs with less variable auditory activity across trials and animals (Fig. 5). Our results for across-animal variability have parallels to the Drosophila olfactory pathway, whereby third-order MB neurons are not stereotyped, while presynaptic neurons in the AL, the PNs, are44. Similarly, a lack of stereotypy beyond early mechanosensory brain areas may reflect stochasticity in synaptic wiring. The amount of variation we observed in some brain areas was large (Extended Data Fig. 9), and follow-up experiments with sparser driver lines will be needed to validate whether what we report here applies to variation across individual identifiable neurons.
We also observed a wide range of across-trial variability throughout neuropils with auditory activity (Fig. 5b). Imaging from a subset of brain regions in behaving flies revealed that trial-to-trial variance in auditory responses is not explained by spontaneous movements (Fig. 6), which suggests that variance is driven by internal dynamics. This result differs from recent findings in the mouse brain33,34, which showed that a large fraction of activity in sensory cortices corresponds to non-task-related or spontaneous movements. This may indicate an important difference between invertebrate and vertebrate brains and the degree to which ongoing movements shape activity across different brains. However, we should point out that while motor activity is known to affect sensory activity in flies45, this modulation is tied to movements that are informative for either optomotor responses or steering46,47. In our experiments, although flies walked abundantly, they did not produce reliable responses to auditory stimuli, although playback of the same auditory stimuli can reliably change walking speed in freely behaving flies4. Adjusting our paradigm to drive such responses48 might uncover behavioral modulation of auditory activity. Alternatively, behavioral modulation of auditory responses may occur primarily in motor areas, such as the central complex, or areas containing projections of descending neurons49. Further dissection of the sources of this variability would require the simultaneous capture of more brain activity in behaving animals while not significantly compromising spatial resolution.
Here, we provide tools for characterizing sensory activity registered in common atlas coordinates for comparisons across trials, individuals and sexes. By producing maps for additional modalities and stimulus combinations and by combining these maps with information on connectivity between and within brain regions30,50, the logic of how the brain represents the myriad stimuli and their combinations present in the world should emerge.
Methods
Fly stocks.
Flies used to generate the IVIA and for functional-imaging experiments were of the following genotype: w/+; GMR57C10-LexA/+; 13xLexAop-GCaMP6s, 8xLexAop-mCD8tdTomato/+ (nsyb-LexA-G6s-tdtom), all ‘+’ chromosomes came from the NM91 wild-type strain9. Iav1 mutants had the following genotype: iav1/Y; GMR57C10-LexA/+; 13xLexAop-GCaMP6s, 8xLexAop-mCD8tdTomato/+. For imaging of olfactory PNs and Kenyon cells, flies had the w; GMR57C10-LexA, 8xLexAop-myr-tdTomato/GH146-Gal4; 20xUAS-GCaMP6s/+ genotype and w; GMR57C10-LexA, 8xLexAop-myr-tdTomato/+; 20xUAS-GCaMP6s/OK107-Gal4 genotype, respectively. For stalk segmentation of pC1, pC2l and pC2m Dsx-expressing neurons, we used the w; GMR57C10-LexA/20xUAS-GCaMP6s, 20xUAS-myr-tdTomato; 8xLexAop-mCD8tdTomato/dsxGal4 genotype. For stalk segmentation of P1 neurons and pMP7 neurons, we used the w; NP2631-Gal4, 20XUAS>STOP>CsChrimson.mVenus/GMR57C10-LexA; 8xLexAop-mCD8tdTo mato/fruFLP genotype. Other flies were acquired as follows: 13xLexAop-GCaMP6s (44590), GH146-Gal4(30026) and OK107-Gal4 (854) from the Bloomington Stock Center; iav1 from the Kyoto Stock Center; GMR57C10-LexA from G. Rubin at Janelia Research Campus; 20XUAS>STOP>CsChrimson.mVenus from V. Jayaraman at Janelia Research Campus; 8xLexAop-mCD8tdTomato from Y. N. Jan at the University of California San Francisco; NP2631-Gal4 from the Kyoto Stock Center; fruFLP from B. Dickson at Janelia Research Campus; dsxGal4 from S. Goodwin at University of Oxford; NM91 from the Andolfatto Group at Columbia University; and 20xUAS-myr-tdTomato from G. Turner at Janelia Research Campus.
Sound delivery and acoustic stimuli.
Sound was delivered as described in ref. 6. During experiments, the sound tube was positioned in front of the fly at a distance of 2 mm and an angle of ~17° from the midline (elevation of ~39°). In this configuration, we observed qualitatively similar responses from the AMMC in both hemispheres.
Acoustic stimuli were organized in two main protocols: (1) coarse-tuning protocol and (2) feature-tuning protocol (see Supplementary Table 2 for a list of all stimulus sets used). For both protocols, the intensity of auditory stimuli presented was within the dynamic range of JONs51 (0.1–6 mm s−1) and within the distribution of song intensities produced during natural courtship52 (pulse and sine song intensity of 5 mm s−1, and broadband white noise intensity of 2 mm s−1). Stimuli were preceded and followed by 10 or 4 s of silence for the coarse-tuning and feature-tuning protocol, respectively, to reduce cross-talk between responses to the subsequent stimulus. Stimuli were presented in blocks (for example, pulse, sine, white noise, repeated six times) or randomized for the coarse-tuning and feature-tuning protocol, respectively. We chose a long stimuli duration (10 or 4 s) to maximize the signal-to-noise ratio of auditory responses.
Head-fixed walking setup and behavioral tracking.
We shaped 9-mm diameter balls from polyurethane foam (FR-7120, General Plastics) similar to that described in ref. 53. The ball rested in an aluminum ball holder with a concave hemisphere 9.6 mm in diameter (Fig. 6a). The ball holder had a 1.27-mm channel drilled through the bottom of the hemisphere and connected to air flowing at ~75 ml min−1 and mounted on a manipulator (MS3, Thorlabs) to adjust the position of the ball under each fly.
We imaged the fly and ball from the side (90° to the anterior–posterior axis of the fly), illuminated by a pair of infrared 850-nm light-emitting diodes (M850F2, Thorlabs) coupled to optic fibers and collimator lenses (F810SMA-780, Thorlabs). Imaging was done with a monochromatic camera (Flea3 FL3-U3–13Y3M-C 1/2, FLIR), externally triggered at 100 Hz, with a zoom lens (C-Mount 15.5–20.4 mm Varifocal, Computar). The lens had a 770–790-nm narrow-band filter (BN785, MidOpt) attached to block a two-photon excitation laser (920 nm). This camera was used to position the fly and to track both the ball and fly movement.
To track the rotation of the ball, we used a Windows implementation of FicTrac software54 (see https://github.com/murthylab/fictrac#windows-installation). FicTrac calculates the angular position of the ball for each frame and reconstructs the x and y trajectory on a fictive two-dimensional (2D) surface. The ball was marked with non-repetitive shapes using a black permanent marker (PK10, Sarstedt) to allow tracking of the rotation of the ball. We used the x and y trajectories to calculate instantaneous ground velocities (forward and lateral velocity, and the speed in the fictive heading direction (forward plus lateral velocity)) and the angular position to calculate the instantaneous angular velocities (yaw, pitch and roll) (Fig. 6b).
To extract a high-dimensional representation of the body motion of the fly (that is, legs, abdomen and wings), we used the toolbox facemap34 (see https://github.com/MouseLand/facemap). Facemap applies singular value decomposition to the motion energy movie (the absolute difference between framet and framet+1) (Fig. 6c), extracting the top 500 coefficients of the components (spatial weights) and corresponding scores (temporal profiles) of the motion energy movie (Fig. 6d,e). For this analysis, the field of view of each recorded video was reduced to the smallest box containing the whole fly body, masking out pixels that belonged to the ball, and the head plus anterior thorax (pixels of these body parts were removed due to its contamination with the calcium-imaging-related signal). The top 500 components of the motion energy movie per fly were used as a summary of motor behaviors for later analysis (see the section “Linear modeling of neural activity from stimulus and behavior”).
Fly preparation and functional imaging.
Virgin female or male flies (3–7 days old) were mounted and dissected as previously described27,55, with minor differences for non-behaving animals. The angle of the thorax to the posterior side of the head was kept close to 90°, keeping the head as parallel as possible to the holder floor. Following dissection of the head cuticle, we removed the air sacks and tracheas using sharp forceps (Dumont number 5SF), and any additional fat or soft tissue was removed with suction using a sharp glass pipette. In addition, to enable imaging of ventral central brain regions close to the neck, we pushed the thorax posteriorly and fixed it using a tissue adhesive (3M vetbond) delivered using a sharp glass pipette. Proboscis or digestion-related motion artifacts were minimized by pulling the proboscis and waxing it at an extended configuration and removing muscle M16. In addition, during dissection, we kept the antennae dry and mobile for auditory stimulation. For VNC recordings, in addition to the previous steps, we removed thoracic tissue dorsal to the VNC (for example, the cuticle, indirect flight muscles and the digestive system), exposing the first and second segments of the VNC. For head-fixed behaving animals, flies were fixed to a custom holder (similar to ref. 56) by gluing the lateral and dorsal anterior edges of the thorax with light cured glue (Bondic), followed by waxing of the posterior side of the head (the head was pushed forward to end up with a similar configuration as for dissection for non-behaving animals). The rest of the dissection was the same as for non-behaving animals.
Imaging experiments were performed on a custom-built two-photon laser scanning microscope equipped with 5-mm galvanometer mirrors (Cambridge Technology), an electro-optic modulator (M350–80LA-02 KD*P, Conoptics) to control the laser intensity, a piezoelectric focusing device (P-725, Physik Instrumente), a Chameleon Ultra II Ti:sapphire laser (Coherent) and a water-immersion objective (Olympus XLPlan ×25, NA = 1.05). The fluorescence signal collected by the objective was reflected by a dichroic mirror (FF685 Dio2, Semrock), filtered using a multiphoton short-pass emission filter (FF01–680/sp-25, Semrock), split by a dichroic mirror (FF555 Dio3, Semrock) into two channels (green (FF02–525/40–25, Semrock) and red (FF01–593/40–25, Semrock)) and detected by GaAsP photo-multiplier tubes (H10770PA-40, Hamamatsu). We used a laser power below 20 mW to minimize photodamage. The microscope was controlled in Matlab using ScanImage 5.1 (Vidrio). Dissection chambers were placed beneath the objective, and perfusion saline was continuously delivered directly to the meniscus. The temperature of the perfusion saline was kept at 24 °C using a miniature perfusion cooler and heater unit (TC-RD, TC2–80-150-C, Biosciencetools). We chose GCaMP6s57 over faster sensors to maximize the signal-to-noise ratio of auditory responses. The experimenter was not blind to the animal sex or genotype, but males or females (of the same genotype) were randomly chosen for imaging experiments. For all flies not carrying the iav1 mutation, we only recorded from animals that first exhibited auditory-evoked responses in the AMMC of both hemispheres. Recordings typically lasted for 2–3 h. Flies whose global baseline fluorescence saturated and became homogeneous were discarded. We interpreted this as a sign of an unhealthy fly preparation.
Functional imaging analysis pipeline.
All steps from image preprocessing (step (1)) to registration of functional data to IVIA (step (4)) are schematized in Extended Data Fig. 1a, in which we take one example dataset step by step through the entire functional-imaging analysis pipeline.
Image preprocessing.
For all experiments, we recorded data from each animal using one of the following three protocols. For protocol one, 14–17 subvolumes at 1 Hz were recorded until the whole extent of the z axis per fly (posterior to anterior) was covered, with a 1.2 × 1.4 × 2 μm3 voxel size and an ~542 kHz pixel rate (Figs. 1–3 and 5 and Extended Data Figs. 3, 4 and 10d–g). The full volume scanned per fly was one-quarter of the central brain or ~300 × 300 × 250 μm3. For protocol two, 4–6 subvolumes from selected brain regions (not necessarily contiguous) at 2 Hz were recorded, with a 1 × 1 × 2 μm3, 1 × 1 × 3 μm3 or 1.2 × 1.2 × 5 μm3 voxel size and an ~542 kHz pixel rate (Fig. 4). For protocol three, one subvolume that extends from the LH in the posterior brain to the AMMC in the anterior brain at 1 Hz was recorded, with a 1.2 × 1.4 × 15 μm3 voxel size and an ~542 kHz pixel rate (Fig. 6). To generate the private whole brain atlas (per fly), we imaged 1–4 z-stack volumes (in both GCaMP and tdTomato channels) with a 0.75 × 0.75 × 1 μm3 voxel size to span the entire central brain (both hemispheres) in the z dimension.
For each volumetric time series or segment (x–y–z and time) (Extended Data Fig. 1a), we performed rigid motion correction on the x–y–z axes on the tdTomato signal using the NoRMCorre algorithm22, and this transformation was then applied to the GCaMP6s signal. Segments with an anisotropic x–y voxel size were spatially resampled to have an isotropic x–y voxel size (bilinear interpolation on the x and y axes), and temporally resampled to correct for different slice timing across planes of the same volume. Finally, segments were upsampled (linear interpolation) to twice the original sampling rate (2 Hz and 4 Hz for segments recorded at 1 Hz and 2 Hz, respectively) and aligned relative to the start of the first stimulus (linear interpolation) before analysis.
Segments consecutively recorded along the z axis (for protocol one) were stitched along the z axis using NoRMCorre (similarly, we used the average tdTomato signal as the reference image and applied the transformation to the GCAMP6s signal), obtaining a volume. For protocols two and three, each segment was treated as an independent volume. Volumes were then mirrored to the right hemisphere. Average tdTomato images were saved as individual NRRD (http://teem.sourceforge.net/nrrd) files for registration to IVIA (using the Matlab nrrdWriter function).
For the private whole brain atlases (per fly), z-stacks (~384 × 384 × 280 μm3 each) were stitched on the x–y axes using a pairwise stitching algorithm in Fiji58,59. These whole-brain images were smoothed on the x–y axes (Gaussian kernel size of [3, 3] voxels and Gaussian kernel standard deviation of [2, 2]), mirrored on the x axis as for the volumetric time series. Images were split into tdTomato and GCaMP6s channels and saved separately as NRRD files for registration to IVIA (using the Matlab nrrdWriter function).
ROI segmentation.
We used the constrained non-negative matrix factorization algorithm20 for ROI segmentation. The algorithm was implemented and generalized to 3D data in the Matlab version of the CaImAn toolbox23. We used the greedy initialization (Gaussian kernel size of [9, 9, 5] voxels and Gaussian kernel standard deviation of [4, 4, 2], and 10th percentile baseline subtraction) with 600 components per substack of 11 planes and walked through the entirety of the z axis with a substack overlap of 3 planes. After initialization, spatial components were updated using the ‘dilate’ search method, and default settings were used for updating temporal components. Then, spatially overlapping components were iteratively merged if the Pearson’s correlation of temporal components between them was greater than 0.9 (this allows stitching of contiguous patches of neuropil with similar temporal profiles). The combination of these two steps segmented soma-like and neuropil-like ROIs accordingly. We wrote custom code to compile spatial and temporal components from all substacks per fly (similar to run_CNMF_patches.m) and to calculate ΔF(t)/F0(t) ROI. Background fluorescence F0(t) and delta fluorescence ΔF(t) (that is, ROI activity) per ROI are directly modeled by CaImAn and provided as output variables. To calculate the ΔF(t)/F0(t), first, for each ROI, we detrended ΔF(t) (we performed a 20th percentile filtering over time, calculated from overlapping windows of 60 s each). Second, for each ROI we added to F0(t) its corresponding trend removed from ΔF(t). ΔF(t)/F0(t) was calculated as the detrended ΔF(t) divided by F0(t) (which is now the background fluorescence plus the trend from ΔF(t)) for each time point). See also https://github.com/murthylab/FlyCaImAn.
Identification of stimulus-modulated ROIs.
We linearly modeled each ROI signal as a convolution of the stimuli history and a set of three filters (one per stimulus) (Extended Data Fig. 1d). Filters were estimated using ridge regression60, with a filter length of 10 and 6 s for the coarse-tuning protocol (Figs. 1–3, 5 and 6) and the feature-tuning protocol (Fig. 4), respectively. For filter estimation, we partitioned each ROI raw signal (F(t)) into training (80%) and testing (20%) datasets. Estimated filters (q(Τ)) were then convolved with the stimulus history (f(t)) to generate the predicted signal (g(t)) for each ROI. The prediction goodness was measured as the Pearson’s correlation coefficient between raw and predicted signals ρ(F, g) using 15-fold cross-validation. The statistical significance of correlation coefficients (ρ(F, g)) was determined by bootstrapping. Each ROI raw signal was randomly shuffled in chunks of 10 s (sF(t)), and a distribution of 10,000 correlation coefficients (ρ(sF, g))) between each independent shuffle and the predicted signal was generated. P values for ρ(F, g) significance were calculated as the fraction of ρ(sF, g) with values greater than the 30th percentile of ρ(F, g), and adjusted using Benjamini–Hochberg false discovery rate (FDR) correction (FDR = 0.01).
Registration of functional data to IVIA.
In vivo functional data from each fly were registered to the IVIA using the structural channel (mCD8–tdTomato signal) in a two-step fashion: (1) volumes to private whole brain registration and (2) private whole brain to IVIA atlas registration (Extended Data Fig. 1a).
For volume to private whole brain registration, for each fly, we registered (linear followed by nonlinear registration) the volume tdTomato image to its own whole brain tdTomato image (z-stack volume, see the section “Image preprocessing”). Linear transformation included rotation, translation and anisotropic scaling (metric: normalized mutual information). For the nonlinear transformation (metric: normalized mutual information), we used a small grid spacing (20 μm, with 2 refinements), a Jacobian constraint weight of 0.01 and a smoothness constraint weight of 0.1.
For private whole brain to IVIA atlas registration, for each fly, we registered (linear followed by nonlinear registration) the private whole brain tdTomato image to the IVIA atlas. Linear transformation included rotation, translation, anisotropic scaling and shearing (metric: normalized mutual information). For the nonlinear transformation (metric: normalized mutual information), we used a bigger grid spacing (170 μm, with 5 refinements), a Jacobian constraint weight of 0.001 and a smoothness constraint weight of 1.
These volume-to-whole brain and whole brain-to-IVIA atlas transformations per fly were concatenated to transform spatial components of segmented ROIs from the native volume coordinates to in vivo atlas coordinates.
Construction of the average IVIA.
Recently, an in vivo D. melanogaster female atlas was built1; however, it does not account for known anatomical differences between female and male brains61. Therefore, we built an average IVIA using a pool of five male and eight female brains (expressing membranal tdTomato pan-neuronally (mCD8–tdTomato) (Extended Data Fig. 2a). In the first iteration, we picked a seed brain, half cropped on the x axis and stitched it to its mirror image to generate a symmetric initial brain seed. We then registered all the brains to the seed brain using the CMTK registration toolbox42,62. We used linear followed by nonlinear registration. We then generated a new average intensity and average deformation seed brain using an active deformation model (implemented in CMTK as avg_adm function). We iterated this process and obtained the IVIA at the end of the fifth iteration.
Construction of IVIA–IBNWB bridging registrations.
From all the fixed-brain atlases with anatomical labels (IBNWB and JFRC), we chose the IBNWB atlas based on the similar nsyb–GFP signal coverage (neuropil, fiber bundles and cell body rind) compared to the IVIA atlas (pan-neuronal mCD8–tdTomato). Despite this similarity, the two atlases have dissimilar distributions of fluorescence intensities per anatomical region and differences in signal-to-noise ratio (much higher for IBNWB), which limited the accuracy of intensity-based registration algorithms. Therefore, we used the rigid and non-rigid point set registration algorithm coherent point drift (CPD)26. We chose the surface of selected neuropils and neurite tracts and the entire central brain as sources of points to compare between atlases (Extended Data Fig. 2d). We converted segmented binary images from each of these brain regions to triangular surface meshes, generating a set of vertices and edges per brain region to be used for registration. We applied a rigid (translation, rotation and scaling) followed by a non-rigid CPD registration in both directions (IVIA-to-IBNWB and IBNWB-to-IVIA) between the two sets of vertices across all brain regions. For the non-rigid registration, we explored the space of the two main meta-parameters beta (Gaussian smoothing filter size) and lambda (regularization weight). We chose the non-rigid transformation (for either direction) with the highest Jaccard index across all segmented brain regions that had spatially smooth deformations (visually inspected) (Extended Data Fig. 2e).
Using the IVIA-to-IBNWB transformation, we mapped an anatomical segmentation of neuropils and neurite bundles25 to the IVIA coordinates (Fig. 1g). In this segmentation, the AMMC would also include JON projections. In addition, we merged the MB accessory calyx to the MB calyx due to its small size relative to the average ROI volume.
Registration accuracy.
To assess IVIA within-atlas accuracy, we segmented stalks from pC1, pC2l, pC2m and pM7 neurons (n = 20, 10, 12 and 7 hemibrains for pC1, pC2l, pC2m and pM7, respectively), which were consistently identified across individuals in flies expressing GCaMP6s via dsxGal4 or CsChrimson via NP2631–GAL4 and fruFLP combo. These stalks covered different dorsoventral regions from the posterior half of the brain (Extended Data Fig. 2b). We measured the within-atlas registration accuracy as the standard deviation of each trace class from the mean trace (all traces across animals). The overall within-atlas accuracy was calculated as the standard deviation across all trace classes (Extended Data Fig. 2b). This estimate is the sum of biological variability and algorithm-associated error. Although we obtained cellular accuracy (~4.3 μm), this is slightly lower compared with values reported for fixed-brain atlases (2–3 μm)42,63–65, and this might be due to both a richer spatial pattern and higher signal-to-noise ratio of the BRP (nc82) antibody staining in fixed-brain experiments.
To assess IVIA and IBNWB between-atlas accuracy, we used segmented pC1 stalks from the IVIA (described above) and compared them to pC1 neurons originally collected in the FCWB atlas (imaged via fruGal4 expression of GFP63), which were mapped to IBNWB (error of the transformation between FCWB and IBNWB is assumed to be negligible). Similar to the IVIA within-atlas accuracy, we calculated the mean pC1 trace for each atlas (μ-pC1-IVIA and μ-pC1-IBNWB) and the within-atlas accuracy (σ-pC1-IVIA and σ-pC1-IBNWB) (Extended Data Fig. 2f,g). We measured the IVIA-to-IBNWB registration accuracy as the difference between the σ-pC1-IBNWB and the standard deviation of pC1-IBNWB traces mapped to the IVIA relative to μ-pC1-IVIA. Moreover, we measured the IBNWB-to-IVIA registration accuracy as the difference between the σ-pC1-IVIA and the standard deviation of pC1-IVIA traces mapped to IBNWB relative to μ-pC1-IBNWB.
Anatomical cataloging of central brain activity.
For each fly, we combined all auditory ROIs and binarized these volumes to obtain binary auditory voxel maps per fly. Auditory maps were then transformed to IVIA coordinates. To account for the known fly-to-atlas error, we dilated these transformed maps ([1.25, 1.25, 2] μm3 in x–y–z). We then summed these maps across flies to generate a density volume (Fig. 2c and Extended Data Figs. 1b, 3b,c,i, 4f and 5d,e). To spatially map auditory activity, we transformed neuropils and neurite bundle segmentation from IVIA coordinates to imaged volume coordinates. We assigned ROIs to a given neuropil or neurite bundle using the following criteria: ROIs must have at least a 15% volume overlap with its assigned neuropil or neurite bundle; if overlapping with multiple neuropils or neurite bundles, ROIs are assigned to the neuropil or neurite bundle with the highest overlap. For auditory ROIs, we then quantified the volume they represent relative to the neuropil or neurite bundle for each fly (we used ROI, neuropil and neurite bundle volume in IVIA space) (Fig. 2e and Extended Data Fig. 3m).
Functional clustering of auditory responses to pulse, sine and white noise stimuli.
For all auditory-modulated ROIs from intact and iav1 flies (19,389 ROIs from 21 male (4 male flies carried iav1) and 17 female flies expressing tdTomato and GCaMP pen-neuronally) presented with the coarse-tuning protocol (this includes pulse, sine, white noise, natural song and Pfast-like pulse; see the section “Sound delivery acoustic stimuli”), we calculated the median response to each stimulus across trials, including 10 s before stimulus onset to 10 s after stimulus offset. We concatenated the median signal across pulse, sine and white noise stimuli only (183 time points) for each ROI and z-scored it. We then hierarchically clustered ROI signals (pooled from all intact and iav1 flies) using a Euclidean distance metric and inner square distance metric between clusters (Ward’s method). This clustering split ROI responses into inhibitory and excitatory responses (first branching of the hierarchical tree), and we then chose a distance (distance of 136) that would split both inhibitory and excitatory responses into the smallest number of distinct responses types (18 response types). We also required that each cluster be present in at least 18 flies (half the entire dataset). Although we could further split these clusters, we stopped at 18 given that subclusters within each of the 18 clusters were similar to the parent cluster, which suggested the occurrence of oversegmentation (Fig. 3a).
Response type kinetics and tuning.
For each mean trace per response type, we measured adaptation, half-time to peak and decay of auditory responses (Fig. 3d–f). Adaptation was measured as the percentage of the absolute baseline subtracted signal 2 s before stimulus offset with respect to the signal 2 s after stimulus onset (‘adaptation index’). Therefore, adaptation indices below 100% indicate some degree of adaptation (the lower the adaptation index, the higher the adaptation), while adaptation indices above 100% indicate a sustained increase in response (the higher the adaptation index, the higher the increase). Half-time to peak was measured as half the time to reach the peak signal during stimuli presentation. The decay time constant of auditory responses was determined by fitting an exponential to the 20 values after stimulus offset (10 s), except for response types 2 and 3, for which we fit an exponential to the 20 values after positive rebound.
To determine the coarse tuning of response types to pulse, sine or white noise, we measured the magnitude of stimulus-evoked responses. The response magnitude was defined as the absolute 80th or 20th percentile with the greater absolute value of baseline subtracted signal from stimulus onset to 2 s after stimulus offset (the baseline was defined as the mean signal from −4 to −0.5 s relative to stimulus onset, and dividing the response magnitude by the standard deviation during baseline gave us the signal-to-noise ratio). Response magnitudes were normalized by the maximum value across stimuli and multiplied by 100 to have units as a percentage. We considered a response type as tuned to a particular stimulus if the response magnitude to the other stimuli was below 85% of the response to the preferred stimulus, otherwise it was considered nonselective (Fig. 3g).
Clustering of tuning curves.
We generated tuning curves to the ‘feature tuning’ stimulus set (Fig. 4). For these experiments, we sampled a fraction of each neuropil selected (Extended Data Fig. 7a). For all auditory-modulated ROIs from intact flies (10,872 ROIs from 10 male and 11 female flies) presented with the feature-tuning protocol, we calculated the median response to each stimulus across trials, including 4 s before stimulus onset to 4 s after stimulus offset, and z-scored it. We measured the magnitude of responses to the auditory stimulus for each ROI. The baseline was defined as −2 to −0.25 s from stimuli onset, and we then calculated the mean μb of the signal during this time. The response magnitude was defined as the absolute 80th or 20th percentile with the greater absolute value of μb subtracted signal from stimulus onset to 2 s after stimulus offset. For each ROI, we generated a tuning curve by concatenating the absolute response magnitude across the presented sine song at different carrier frequencies, and the pulse song at different carrier frequencies, pulse pause and pulse durations (Extended Data Fig. 7b). We then hierarchically clustered ROI tuning curves using a Euclidean distance metric and inner square distance metric between clusters (Ward’s method). Examination of the tree revealed that tuning curves divided into three main classes: tuning to sine, tuning to pulses and nonselective. We then reduced the distance threshold (distance of 60) until we split these three tuning types into the smallest number of visually distinct clusters (Fig. 4c).
Measuring activity diversity across brain regions.
We collected ROIs belonging to the same neuropil across flies (separately for intact or iav1 flies). These ROIs were sorted by their response type within a neuropil. Using the volume of each ROI, we calculated the total number of voxels for each response per neuropil to obtain a distribution of voxel count per response type for each neuropil (we use ROI volume in IVIA space). These distributions were then max-normalized within neuropils; these are the values shown in Fig. 3b. We then calculated the entropy for each response type distribution as a measure of activity diversity, whereby regions with sparse response types will have a lower entropy value than regions with more broad or uniform distribution.
Measuring trial-to-trial variability.
To evaluate the variability of auditory responses across trials, we measured the residual of the ROI activity variance explained by the mean ROI activity across trials for pulse, sine and white noise stimuli (‘variability index’). Each trial included 10 s before stimulus onset to 10 s after stimulus offset. We concatenated the signal across stimuli and trials (183 time points) for each ROI and z-scored it. We computed the variability index for each ROI and then sorted these indices by neuropil (Fig. 5b).
Measuring across-individual variability.
We calculated response type distributions per neuropil (Fig. 3), but for each fly separately, obtaining an array of vectors of 18 dimensions per neuropil per fly. We normalized each vector to unit norm. We then used the cosine of the angle between normalized vectors as our measure of similarity of response distributions (‘similarity index’). Response distributions contained values greater than or equal to 0, thereby restricting the distribution of similarity indices to the 0–1 range.
Measuring sex-specific differences in activity.
Similar to the previous section, we calculated response-type distributions per neuropil for each fly separately. These distributions were then sorted by sex. We then evaluated differences in probability of each of the 18 response types across sexes for each neuropil (for significance of sex-specific differences see the section “Statistics”).
Identification of stimuli-modulated motor behaviors.
Similar to the previous section (“Identification of stimulus-modulated ROIs”), we linearly modeled each motor variable (velocity and motion energy component, X(t)) as a convolution of the stimuli history and a set of three filters (one per stimulus). Filters were estimated using ridge regression60, with a filter length of 10 s. The Pearson’s correlation coefficient between the raw motor variable (X(t)) to its linear prediction (g(t)), the Pearson’s correlation coefficient between the shuffled motor variable (sX(t)) to g(t) and the corrected P values were calculated as described in the section “Identification of stimulus-modulated ROIs”.
Linear modeling of neural activity from stimulus and behavior.
To measure the fraction of neural variance that could be predicted from stimulus presentation or behavior of the fly, we constructed a linear model y(t) = μ(t) + Beta × X(t) + ε(t) (Fig. 6g). The raw signal for each ROI (y(t)) was modeled as the sum of the stimulus component (the trial average across auditory stimuli pulse, sine and white noise, μ(t)), behavior component (behavior predictors—velocities and top 500—times a set of weights, Beta × X(t)) and noise (ε(t)). Behavior predictors were smoothed over time (Gaussian kernel size of ten time points and Gaussian kernel standard deviation of three) and downsampled to match the sampling rate of recorded activity (2 Hz). We computed the weights (Beta) of behavior predictors using ridge regression60. To quantify the explanatory power of the full model or reduced models (stimulus or behavior component only), we computed the explained variance using fivefold cross-validation.
Analysis of the hemibrain connectome.
To find connections between early auditory neuropils—AMMC/SAD and WED—and downstream neuropils, we used the new hemibrain connectome30 (Fig. 2g). The hemibrain project reconstructed 26,157 neurons and synaptic connections in an EM volume of approximately one-third of the Drosophila central brain. This volume includes a large portion of the AMMC, SAD and WED. We downloaded the connectivity matrix of all 26,157 neurons (https://www.janelia.org/project-team/flyem/hemibrain), and wrote a custom code in Python to find neuropils postsynaptic to a query neuropil, connected via one or two synapses. We considered a connection between two neurons if they had more than five synapses.
Statistics.
All statistical analyses were performed using Matlab 2018a. Effect sizes of female-to-male differences were measured using Cohen’s d (Extended Data Fig. 5c); data distribution was assumed to be normal, but this was not formally tested. Within a neuropil, we determined the significance of female-to-male differences for each response type using two-tailed two-sample t-tests with Benjamini–Hochberg FDR correction. For significance of stimuli modulation, see the section “Identification of stimulus-modulated ROIs”. No statistical methods were used to predetermine sample sizes, but our sample sizes were similar to those reported in previous publications4,18.
Reporting Summary.
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Extended Data
Supplementary Material
Acknowledgements
We thank J. Clemens for assistance on auditory stimuli delivery and linear modeling of calcium responses, as well as helpful discussions on auditory coding; A. Giovannucci for help with ROI segmentation of volumetric calcium signals and guidance with CaImAn toolbox usage; G. Jefferis and T. Rohlfing for help using the image registration toolbox CMTK and the neuroanatomy toolbox natverse; B. Cowley for help with linear modeling of calcium responses in behaving flies; A. Calhoun for help with analysis of the hemibrain connectome dataset; and T. Pereira for assistance with the CPD algorithm. We also thank G. Jefferis, S. Ahmed, C. Baker and J. Clemens for comments on the manuscript. D.A.P. was supported in part by a NSF Physics Frontier Center grant, and M.M. was supported by a NIH NINDS New Innovator award, NIH BRAIN Initiative R01s NS104899 and NS110060, and a Howard Hughes Medical Institute Faculty Scholar award.
Footnotes
Code availability
The Matlab code for the neural imaging data processing pipeline is available at https://github.com/murthylab/FlyCaImAn.
Online content
Any methods, additional references, Nature Research reporting summaries, source data, extended data, supplementary information, acknowledgements, peer review information; details of author contributions and competing interests; and statements of data and code availability are available at https://doi.org/10.1038/s41593-020-00743-y.
Competing interests
The authors declare no competing interests.
Additional information
Extended data is available for this paper at https://doi.org/10.1038/s41593-020-00743-y.
Supplementary information is available for this paper at https://doi.org/10.1038/s41593-020-00743-y.
Peer review information Nature Neuroscience thanks Ioana Carcea, Elizabeth Hillman, and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.
Reprints and permissions information is available at www.nature.com/reprints.
Data availability
Calcium imaging data from this study (all segmented ROIs) are publicly available at https://doi.org/10.34770/gv6w-5351.
References
- 1.Mann K, Gallen CL & Clandinin TR Whole-brain calcium imaging reveals an intrinsic functional network in Drosophila. Curr. Biol 27, 2389–2396.e4 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Chen X et al. Brain-wide organization of neuronal activity and convergent sensorimotor transformations in larval zebrafish. Neuron 100, 876–890.e5 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Bates AS et al. The natverse, a versatile toolbox for combining and analysing neuroanatomical data. eLife 9, e53350 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Deutsch D, Clemens J, Thiberge SY, Guan G & Murthy M Shared song detector neurons in Drosophila male and female brains drive sex-specific behaviors. Curr. Biol 29, 3200–3215.e5 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.von Schilcher F The role of auditory stimuli in the courtship of Drosophila melanogaster. Anim. Behav 24, 18–26 (1976). [Google Scholar]
- 6.Clemens J et al. Connecting neural codes with behavior in the auditory system of Drosophila. Neuron 87, 1332–1343 (2015); erratum 97, 475 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Yoon J et al. Selectivity and plasticity in a sound-evoked male–male interaction in Drosophila. PLoS ONE 8, e74289 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Clemens J et al. Discovery of a new song mode in Drosophila reveals hidden structure in the sensory and neural drivers of behavior. Curr. Biol. 28, 2400–2412.e6 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Coen P et al. Dynamic sensory cues shape song structure in Drosophila. Nature 507, 233–237 (2014). [DOI] [PubMed] [Google Scholar]
- 10.Coen P, Xie M, Clemens J & Murthy M Sensorimotor transformations underlying variability in song intensity during Drosophila courtship. Neuron 89, 629–644 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Göpfert MC & Robert D The mechanical basis of Drosophila audition. J. Exp. Biol 205, 1199–1208 (2002). [DOI] [PubMed] [Google Scholar]
- 12.Yorozu S et al. Distinct sensory representations of wind and near-field sound in the Drosophila brain. Nature 458, 201–205 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Kamikouchi A et al. The neural basis of Drosophila gravity-sensing and hearing. Nature 458, 165–171 (2009). [DOI] [PubMed] [Google Scholar]
- 14.Lai JS-Y, Lo S-J, Dickson BJ & Chiang A-S Auditory circuit in the Drosophila brain. Proc. Natl Acad. Sci. USA 109, 2607–2612 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Vaughan AG, Zhou C, Manoli DS & Baker BS Neural pathways for the detection and discrimination of conspecific song in D. melanogaster. Curr. Biol 24, 1039–1049 (2014). [DOI] [PubMed] [Google Scholar]
- 16.Zhou C et al. Central neural circuitry mediating courtship song perception in male Drosophila. eLife 4, e08477 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Matsuo E et al. Organization of projection neurons and local neurons of the primary auditory center in the fruit fly Drosophila melanogaster. J. Comp. Neurol 524, 1099–1164 (2016). [DOI] [PubMed] [Google Scholar]
- 18.Patella P & Wilson RI Functional maps of mechanosensory features in the Drosophila brain. Curr. Biol 28, 1189–1203.e5 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Wang K et al. Neural circuit mechanisms of sexual receptivity in Drosophila females. Preprint at bioRxiv 10.1101/2020.08.07.241919 (2020). [DOI] [PubMed] [Google Scholar]
- 20.Pnevmatikakis EA et al. Simultaneous denoising, deconvolution, and demixing of calcium imaging data. Neuron 89, 285–299 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Aimon S et al. Fast near-whole-brain imaging in adult Drosophila during responses to stimuli and behavior. PLoS Biol 17, e2006732 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Pnevmatikakis EA & Giovannucci A NoRMCorre: an online algorithm for piecewise rigid motion correction of calcium imaging data. J. Neurosci. Methods 291, 83–94 (2017). [DOI] [PubMed] [Google Scholar]
- 23.Giovannucci A et al. CaImAn an open source tool for scalable calcium imaging data analysis. eLife 8, e38173 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Yang HH et al. Subcellular imaging of voltage and calcium signals reveals neural processing in vivo. Cell 166, 245–257 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Ito K et al. A systematic nomenclature for the insect brain. Neuron 81, 755–765 (2014). [DOI] [PubMed] [Google Scholar]
- 26.Myronenko A & Song X Point set registration: coherent point drift. IEEE Trans. Pattern Anal. Mach. Intell 32, 2262–2275 (2010). [DOI] [PubMed] [Google Scholar]
- 27.Tootoonian S, Coen P, Kawai R & Murthy M Neural representations of courtship song in the Drosophila brain. J. Neurosci 32, 787–798 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Azevedo AW & Wilson RI Active mechanisms of vibration encoding and frequency filtering in central mechanosensory neurons. Neuron 96, 446–460. e9 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Chang AEB, Vaughan AG & Wilson RI A mechanosensory circuit that mixes opponent channels to produce selectivity for complex stimulus features. Neuron 92, 888–901 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Scheffer LK et al. A connectome and analysis of the adult Drosophila central brain. Elife 9, e57443 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Lehnert BP, Baker AE, Gaudry Q, Chiang A-S & Wilson RI Distinct roles of TRP channels in auditory transduction and amplification in Drosophila. Neuron 77, 115–128 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Senthilan PR et al. Drosophila auditory organ genes and genetic hearing defects. Cell 150, 1042–1054 (2012). [DOI] [PubMed] [Google Scholar]
- 33.Musall S, Kaufman MT, Juavinett AL, Gluf S & Churchland AK Single-trial neural dynamics are dominated by richly varied movements. Nat. Neurosci. 22, 1677–1686 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Stringer C et al. Spontaneous behaviors drive multidimensional, brainwide activity. Science 364, 255 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Baker CA et al. Neural network organization for courtship song feature detection in Drosophila. Preprint at bioRxiv 10.1101/2020.10.08.332148 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Harris DT, Kallman BR, Mullaney BC & Scott K Representations of taste modality in the Drosophila brain. Neuron 86, 1449–1460 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Naumann EA et al. From whole-brain data to functional circuit models: the zebrafish optomotor response. Cell 167, 947–960.e20 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Haesemeyer M, Robson DN, Li JM, Schier AF & Engert F A brain-wide circuit model of heat-evoked swimming behavior in larval zebrafish. Neuron 98, 817–831.e6 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Frechter S et al. Functional and anatomical specificity in a higher olfactory centre. eLife 8, e44590 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Auer TO & Benton R Sexual circuitry in Drosophila. Curr. Opin. Neurobiol 38, 18–26 (2016). [DOI] [PubMed] [Google Scholar]
- 41.Vogt K et al. Shared mushroom body circuits underlie visual and olfactory memories in Drosophila. eLife 3, e02395 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Jefferis GSXE et al. Comprehensive maps of Drosophila higher olfactory centers: spatially segregated fruit and pheromone representation. Cell 128, 1187–1203 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Bennet-Clark HC & Ewing AW Pulse interval as a critical parameter in the courtship song of Drosophila melanogaster. Anim. Behav 17, 755–759 (1969). [Google Scholar]
- 44.Murthy M, Fiete I & Laurent G Testing odor response stereotypy in the Drosophila mushroom body. Neuron 59, 1009–1023 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Chiappe ME, Seelig JD, Reiser MB & Jayaraman V Walking modulates speed sensitivity in Drosophila motion vision. Curr. Biol 20, 1470–1475 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Fujiwara T, Cruz TL, Bohnslav JP & Chiappe ME A faithful internal representation of walking movements in the Drosophila visual system. Nat. Neurosci 20, 72–81 (2017). [DOI] [PubMed] [Google Scholar]
- 47.Kim AJ, Fitzgerald JK & Maimon G Cellular evidence for efference copy in Drosophila visuomotor processing. Nat. Neurosci 18, 1247–1255 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Batchelor AV & Wilson RI Sound localization behavior in Drosophila melanogaster depends on inter-antenna vibration amplitude comparisons. J. Exp. Biol 222, jeb191213 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Rayshubskiy A et al. Neural control of steering in walking Drosophila. Preprint at bioRxiv 10.1101/2020.04.04.024703 (2020). [DOI] [Google Scholar]
- 50.Zheng Z et al. A complete electron microscopy volume of the brain of adult Drosophila melanogaster. Cell 174, 730–743.e22 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Effertz T, Wiek R & Göpfert MC NompC TRP channel is essential for Drosophila sound receptor function. Curr. Biol 21, 592–597 (2011). [DOI] [PubMed] [Google Scholar]
- 52.Morley EL, Jonsson T & Robert D Auditory sensitivity, spatial dynamics, and amplitude of courtship song in Drosophila melanogaster. J. Acoust. Soc. Am 144, 734 (2018). [DOI] [PubMed] [Google Scholar]
- 53.Seelig JD et al. Two-photon calcium imaging from head-fixed Drosophila during optomotor walking behavior. Nat. Methods 7, 535–540 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Moore RJD et al. FicTrac: a visual method for tracking spherical motion and generating fictive animal paths. J. Neurosci. Methods 225, 106–119 (2014). [DOI] [PubMed] [Google Scholar]
- 55.Murthy M & Turner G Whole-cell in vivo patch-clamp recordings in the Drosophila brain. Cold Spring Harb. Protoc 2013, 140–148 (2013). [DOI] [PubMed] [Google Scholar]
- 56.Maimon G, Straw AD & Dickinson MH Active flight increases the gain of visual motion processing in Drosophila. Nat. Neurosci 13, 393–399 (2010). [DOI] [PubMed] [Google Scholar]
- 57.Chen T-W et al. Ultrasensitive fluorescent proteins for imaging neuronal activity. Nature 499, 295–300 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Preibisch S, Saalfeld S & Tomancak P Globally optimal stitching of tiled 3D microscopic image acquisitions. Bioinformatics 25, 1463–1465 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Schindelin J et al. Fiji: an open-source platform for biological-image analysis. Nat. Methods 9, 676–682 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Park M & Pillow JW Receptive field inference with localized priors. PLoS Comput. Biol 7, e1002219 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Cachero S, Ostrovsky AD, Yu JY, Dickson BJ & Jefferis GSXE Sexual dimorphism in the fly brain. Curr. Biol 20, 1589–1601 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Rohlfing T & Maurer CR Jr. Nonrigid image registration in shared-memory multiprocessor environments with application to brains, breasts, and bees. IEEE Trans. Inf. Technol. Biomed 7, 16–25 (2003). [DOI] [PubMed] [Google Scholar]
- 63.Chiang A-S et al. Three-dimensional reconstruction of brain-wide wiring networks in Drosophila at single-cell resolution. Curr. Biol 21, 1–11 (2011). [DOI] [PubMed] [Google Scholar]
- 64.Peng H et al. BrainAligner: 3D registration atlases of Drosophila brains. Nat. Methods 8, 493–500 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Yu JY, Kanai MI, Demir E, Jefferis GSXE & Dickson BJ Cellular organization of the neural circuit that drives Drosophila courtship behavior. Curr. Biol 20, 1602–1614 (2010). [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Calcium imaging data from this study (all segmented ROIs) are publicly available at https://doi.org/10.34770/gv6w-5351.