Skip to main content
eLife logoLink to eLife
. 2021 Feb 11;10:e61408. doi: 10.7554/eLife.61408

Constructing and optimizing 3D atlases from 2D data with application to the developing mouse brain

David M Young 1,2, Siavash Fazel Darbandi 1, Grace Schwartz 1, Zachary Bonzell 1, Deniz Yuruk 1, Mai Nojima 1, Laurent C Gole 2, John LR Rubenstein 1, Weimiao Yu 2, Stephan J Sanders 1,
Editors: Joseph G Gleeson3, Kate M Wassum4
PMCID: PMC7994002  PMID: 33570495

Abstract

3D imaging data necessitate 3D reference atlases for accurate quantitative interpretation. Existing computational methods to generate 3D atlases from 2D-derived atlases result in extensive artifacts, while manual curation approaches are labor-intensive. We present a computational approach for 3D atlas construction that substantially reduces artifacts by identifying anatomical boundaries in the underlying imaging data and using these to guide 3D transformation. Anatomical boundaries also allow extension of atlases to complete edge regions. Applying these methods to the eight developmental stages in the Allen Developing Mouse Brain Atlas (ADMBA) led to more comprehensive and accurate atlases. We generated imaging data from 15 whole mouse brains to validate atlas performance and observed qualitative and quantitative improvement (37% greater alignment between atlas and anatomical boundaries). We provide the pipeline as the MagellanMapper software and the eight 3D reconstructed ADMBA atlases. These resources facilitate whole-organ quantitative analysis between samples and across development.

Research organism: Mouse

eLife digest

The research community needs precise, reliable 3D atlases of organs to pinpoint where biological structures and processes are located. For instance, these maps are essential to understand where specific genes are turned on or off, or the spatial organization of various groups of cells over time.

For centuries, atlases have been built by thinly ‘slicing up’ an organ, and then precisely representing each 2D layer. Yet this approach is imperfect: each layer may be accurate on its own, but inevitable mismatches appear between the slices when viewed in 3D or from another angle.

Advances in microscopy now allow entire organs to be imaged in 3D. Comparing these images with atlases could help to detect subtle differences that indicate or underlie disease. However, this is only possible if 3D maps are accurate and do not feature mismatches between layers. To create an atlas without such artifacts, one approach consists in starting from scratch and manually redrawing the maps in 3D, a labor-intensive method that discards a large body of well-established atlases.

Instead, Young et al. set out to create an automated method which could help to refine existing ‘layer-based’ atlases, releasing software that anyone can use to improve current maps. The package was created by harnessing eight atlases in the Allen Developing Mouse Brain Atlas, and then using the underlying anatomical images to resolve discrepancies between layers or fill out any missing areas. Known as MagellanMapper, the software was extensively tested to demonstrate the accuracy of the maps it creates, including comparison to whole-brain imaging data from 15 mouse brains. Armed with this new software, researchers can improve the accuracy of their atlases, helping them to understand the structure of organs at the level of the cell and giving them insight into a broad range of human disorders.

Introduction

Anatomical atlases have played a crucial role in research into organogenesis, anatomy, physiology, and pathology (Marquart et al., 2017; Chon et al., 2019; Liu et al., 2020) and have been used for a wide range of applications, including investigating developmental biology (Kunst et al., 2019), defining stereotactic coordinates (Franklin and Paxinos, 2013), identifying patterns of neural activity associated with behavior (Renier et al., 2016; Kim et al., 2017), and integrating morphological, transcriptomic, and neural activity data across species (Marquart et al., 2017). Ongoing initiatives continue to develop these resources, including the Human Cell Atlas (Regev et al., 2019), the BRAIN Initiative (Ecker et al., 2017; Koroshetz et al., 2018), the Human Brain Project (Amunts et al., 2016Bjerke et al., 2018), and centralized repositories of data, including the Allen Brain Atlas (Lein et al., 2007; Jones et al., 2009; Ng et al., 2007).

Macroscale imaging, such as magnetic resonance imaging (MRI), routinely generates data from intact whole organs (Huang and Luo, 2015; Thompson et al., 2014a; Jahanshad et al., 2013; Sudlow et al., 2015; Miller et al., 2016; Alfaro-Almagro et al., 2018; Walhovd et al., 2018) in contrast to microscale imaging, such as light microscopy, which has typically relied on physically cutting thin sections leading to misalignment artifacts. Technological advances in microscopy have facilitated a dramatic increase in imaging of high-resolution, intact whole organs using serial two-photon tomography (STPT) (Ragan et al., 2012) or tissue clearing techniques (e.g. CLARITY [Chung et al., 2013], 3DISCO [Ertürk et al., 2012], and CUBIC [Susaki et al., 2014]) and lightsheet microscopy (Chhetri et al., 2015; Meijering et al., 2016; Mano et al., 2018; Hillman et al., 2019; Cai et al., 2019; Ueda et al., 2020). These 3D microscopy methods have been applied to map cytoarchitecture and cellular connectivity in complex tissues, including the brain (Cai et al., 2019; Kim et al., 2015; Murakami et al., 2018), heart (Gilbert et al., 2019; Bai et al., 2015), and lung (Li et al., 2012; Ardini-Poleske et al., 2017).

Accurate analysis of imaging data from whole organs requires atlases of the corresponding organ, species, developmental age, resolution, and dimensions (Allen Institute for Brain Science, 2017; Susaki and Ueda, 2016; Niedworok et al., 2016; Watson and Watkins, 2019). The progression to 3D microscopy data necessitates 3D anatomical atlases to reference data to existing anatomically-defined labels (Puelles and Rubenstein, 2003; Watson et al., 2012). Although numerous atlases based on 2D physical sections exist (Franklin and Paxinos, 2013; Ardini-Poleske et al., 2017; Savolainen et al., 2009; de Boer et al., 2012; Thompson et al., 2014b), most suffer from label misalignments or insufficient resolution in the third dimension, limiting their utility for 3D imaging (Wang et al., 2020). The most recent version of the Allen Common Coordinate Framework atlas (CCFv3) of the adult (P56) C57BL/6J mouse brain represents one of the most complete 3D atlases at microscopic resolution to date (Wang et al., 2020). A team of neuroanatomists and illustrators redrew structures natively in 3D, sometimes using the original annotations as seeds to guide the redrawing process, producing 658 structural annotations (Wang et al., 2020). Even with this monumental effort, 23% of labels from the original Allen Reference Atlas were not included in the 3D version, largely due to the time and labor constraints of further manual curation (Wang et al., 2020). To date, this titanic effort to generate a 3D atlas at microscopic resolution has not been repeated for most combinations of organ, species, and developmental stage, despite the numerous existing 2D-based atlases.

Automated methods to convert existing 2D-derived atlases into fully 3D atlases could leverage existing, established atlases without the time-consuming effort required for manual 3D curation, or provide the substrate for more rapid manual fine-tuning. This conversion needs to address multiple limitations in atlases based on 2D physical sections, as demonstrated by the challenges of using the existing 3D reconstruction of eight mouse brains, each at a different developmental time point, in the Allen Developing Mouse Brain Atlas (ADMBA) (Thompson et al., 2014b; Figure 1A). First, all atlases in the ADMBA are missing labels across large regions, typically the lateral planes of one hemisphere and the entire opposite hemisphere. Second, warping and asymmetrical placement of samples in the histological images prevents simply mirroring labels from the annotated to the non-annotated hemisphere. Third, as noted in the original Allen Reference Atlas (Ng et al., 2007; Niedworok et al., 2016; Wang et al., 2020), the assembled 3D volume is smooth in the sagittal planes, in which the atlas annotations were drawn, but suffers from substantial edge artifacts in the coronal and axial planes (Figure 1A).

Figure 1. 3D atlas refinement pipeline overview and full example in the E13.5 atlas.

Figure 1.

(A) 2D-derived atlases, such as those in the Allen Developing Mouse Brain Atlas, are smooth and consistent in the sagittal plane in which they were annotated. However, in the 3D reconstructions of these 2D sagittal planes, the coronal and axial planes reveal missing sections and jagged edges. To improve their performance for annotating 3D data, the lateral edges are extended to complete the labeled hemisphere. A 3D rotation is applied to bring the brain parallel to the image borders, then both the completed hemisphere labels and underlying microscopy sections are mirrored across the sagittal midline to complete coverage in the opposite hemisphere. To improve anatomical registration, the labels are each eroded and re-grown through a 3D watershed, guided by the anatomical edge map. To smooth the final product, labels are iteratively filtered with a morphological opening operation, or a closing operation for small, disconnected labels that would otherwise be lost. (B) The pipeline illustrated in axial (top) and sagittal (bottom) views on the ADMBA E13.5 atlas, which requires the full pipeline shown in ‘A’, including an additional step to strip out non-CNS tissue from the original whole-embryo imaging. The nomenclature for pipeline steps shown here is used consistently throughout the manuscript. Insets of the ‘Original’ and ‘Smoothed (edge-aware)’ lateral regions highlight the label extension and smoothing. A spreadsheet mapping colors to label names and IDs in the Allen ontology for each atlas in this manuscript can be found in Supplementary file 1.

Several automated approaches have partially addressed some of these issues. The Waxholm Space rat brain atlas used two standard reference atlases as templates to yield an atlas with relatively smooth 3D structures through manual and semi-automated segmentation, although some label edge artifacts remain (Papp et al., 2014). Erö et al., 2018 performed non-rigid alignments between histological slice planes of the Common Coordinate Framework (CCFv2, 2011), applying the same deformations to the label images. This registration improved alignment; however, the labels often appear to extend beyond the underlying histological planes. Rather than registering planes, Niedworok et al., 2016 applied a smoothing filter on the annotations themselves to reduce label artifacts, though many label edge irregularities persisted.

Here, we present a pipeline to automatically generate fully 3D atlases from existing 2D-derived partial atlases and apply it to the full ADMBA. We performed three steps (Figure 1A): (1) Extension and mirroring of 2D data to approximate missing labels, (2) Smoothing with 3D watershed and morphological methods to align jagged edges between 2D sections, and (3) Edge detection from microscopy images to guide the 3D watershed to align boundaries with anatomical divisions, which we call ‘edge-aware’ refinement (example atlas shown in Figure 1B). We demonstrate qualitative and quantitative improvements over the existing 3D ADMBA. To assess the resulting 3D atlas, we generated 3D imaging data from 15 C57BL/6J wild-type mouse brains at postnatal day 0 (P0) using tissue clearing and lightsheet microscopy. The 3D labels in our refined atlas match the brain structures better than the initial atlas, as demonstrated by closer alignment between labels and anatomical edges and decreased variability within labels at the cellular level. We provide the pipeline as open source software called MagellanMapper to apply the algorithms to equivalent atlases and provide the resulting ADMBA 3D atlases as a resource for immediate application to automated whole-brain 3D microscopy analysis.

Results

Allen developmental mouse brain atlas

The ADMBA contains data for eight mouse brains, each at a different development stage: embryonic day (E)11.5, E13.5, E15.5, E18.5, postnatal day (P) 4, P14, P28, and P56. The data are generated from 158 to 456 serial 2D sagittal sections of 20–25 µm stained with Nissl or Feulgen-HP yellow and imaged with a light microscope at a lateral resolution of 0.99 to 1.049 µm. For the three earliest stages of development, E11.5, E13.5, E15.5, the entire embryo was imaged, rather than just the brain. Each of the eight atlases are annotated with expertly curated labels of brain structures (Thompson et al., 2014b; Allen Institute for Brain Science, 2013b) in a hierarchy starting with the largest structures (e.g. neural plate) extending through 13 sublevels to the smallest annotated structures (e.g. parvicellular part of Lat). Viewed sagittally, these labels cover the majority of tissue in each brain with smooth, anatomically aligned edges (Figure 2A); however, only sections on the left side of each brain are annotated and, for six atlases, labels are not present for the most lateral 14–24% sagittal planes of the brain (4–5% by volume; Figure 2—figure supplement 1; Figure 2B). Labels extend slightly beyond the midline for several atlases, helping to annotate brains with a midline skew, yielding a median label to atlas volume ratio of 51%.

Figure 2. Atlas lateral extension and mirroring for complete brain annotation.

(A) The original E18.5 atlas labels (bottom left), viewed sagittally, demonstrate smooth borders and close correspondence with the underlying microscopy images (upper left; overlaid on right). The dashed blue lines show the sections viewed in ‘B’ and ‘C’. (B) When viewed axially, the most lateral sections lack labels, one hemisphere lacks labels entirely, and label borders are jagged. Slight rotation of the underlying microscopy images leads to asymmetry between its two hemispheres. (C) Similar findings are apparent in the coronal view. (D) The Dice Similarity Coefficient (DSC), a measure of the completeness of labeling compared to the thresholded atlases, for the labeled hemispheres increased for all brains in the ADMBA after lateral edge extension (original median = 0.91, extended median = 0.93; p=0.02, Wilcoxon signed-rank test (WSRT); ‘Mus musculus’ level −1, ID 15564 in the Allen ontology). (E) To fill in the lateral edges using existing labels, a representative lateral label plane was iteratively resized to fit the underlying microscopy images. The plane for each subsequent microscopy plane was thresholded, the bounding box extracted, and the labels resized to fit this bounding box, followed by conforming labels to the underlying gross anatomical boundaries (Figure 2—figure supplement 2). A stretch of compressed planes was expanded (Figure 2—figure supplement 3), and the completed hemisphere of labels mirrored to the other hemisphere after rotation for symmetry to complete the labeling. (F) Coronal view after the lateral edge extension.

Figure 2.

Figure 2—figure supplement 1. Fraction of hemisphere that is unlabeled in each of the ADMBA atlases.

Figure 2—figure supplement 1.

(A) The fraction of unlabeled planes in a given hemisphere is taken as the number of sagittal planes without any labels over the total number of sagittal planes in the hemisphere that should have labels. By definition the extended atlases have no detected unlabeled planes because any plane determined to require labeling was filled. (B) Unlabeled volume fractions are measured by taking label volume over the thresholded histology foreground in the hemisphere. Each hemisphere is taken as the predominantly labeled hemisphere of the given atlas.
Figure 2—figure supplement 2. Lateral edge extension pipeline.

Figure 2—figure supplement 2.

(A) Example labeled sagittal section toward the lateral edge in the ADMBA E18.5 atlas. (B) Overview of the extension algorithm. Plane 38 represents one of the farthest lateral labeled planes. For each discrete structure, the labels are extracted to serve as templates for the subsequent plane. In plane 37, the corresponding template is resized to the bounding box of the microscopy image in that plane. To further fit labels to the underlying anatomy, gross anatomical edges are found in 3D for the full volumetric microscopy images. Labels are eroded, skeletons are added back to avoid loss of thin sections, and labels are regrown by a compact watershed guided by the anatomical edges in the plane. This refined plane of labels thus becomes the template for the next plane. As the extension progresses, some labels disappear during the erosion step, preferentially central labels, modeling the tapering of labels laterally. (C) Every fifth label in a stack after the original template plane in (left) the original atlas and (right) after label extension.
Figure 2—figure supplement 3. Expansion of compressed labels in a stretch of planes in the E18.5 atlas.

Figure 2—figure supplement 3.

(A) Example plane from a stretch of sagittal planes in the original ADMBA E18.5 atlas where the labels are compressed dorsoventrally. The original labels (top) were re-expanded (bottom) to match the extent of the underlying histology section. (B) The stack of original (left) and re-expanded (right) planes from just before until just after this stretch of compressed label planes.

Atlas similarity

To focus our analysis on the brain tissue, we excluded non-CNS tissue for the three embryos for which the brain was not dissected prior to imaging. Across all eight atlases the similarity between the microscopy images and the original label annotations was estimated by taking the threshold of each sagittal section on the left hemisphere of the brain as an approximation of ground truth and comparing the label coverage using the Dice Similarity Coefficient (DSC) (Dice, 1945), calculated by the Insight Segmentation and Registration Toolkit (ITK) (Tustison and Gee, 2009). Higher DSCs reflect greater similarity between the images and labels, with a maximum possible value of 1. The observed DSCs for the original labels in the annotated hemisphere ranged from 0.85 to 0.97 (median 0.91; Figure 2).

Atlas extension

To generate labels across the entire brain, we extended the existing labels to the lateral edges by following histological boundaries, before mirroring the labels on the opposite hemisphere (Figure 2). The most lateral labeled sagittal section provided the initial seed from which to grow labels laterally. First, we resized this plane to match the extent of corresponding microscopy signal in the next, unlabeled lateral sagittal section (Figure 2—figure supplement 2A,B). Next, we refined label boundaries by eroding each label and regrowing it along histological boundaries, which we call ‘edge-aware’ refinement. To model these boundaries, we generated gross anatomical maps in 3D using edge-detection methods on the volumetric histology images. Taking the Laplacian of Gaussian (Marr and Hildreth, 1980) of each histological volumetric image highlighted broad regions of similar intensities, using a relatively large Gaussian sigma of 5 to capture only well-demarcated boundaries. A zero-crossing detector converted this regional map into a binary anatomical edge map (Figure 2—figure supplement 2B). After eroding each label, we next used this anatomical map to guide the regrowth of labels by a compact watershed algorithm (Neubert and Protzel, 2014) step, which adds a size constraint to the classic watershed algorithm. Distances of each pixel to its nearest edge map pixel formed anatomically based watershed catchment areas, while each eroded label served as a large seed from which filling of the nearest catchment areas began and grew until meeting neighboring labels. Thus, we extended a labeled sagittal plane to an unlabeled one, refined by the histology.

This process was repeated iteratively across all remaining lateral sections (Figure 2—figure supplement 2B,C). By using a 3D anatomical map for the watershed and seeding it with the prior plane’s labels, we ensured continuity between planes. To model the tapering of labels laterally, we preferentially eroded central labels by weighting each label’s filter size based on the label’s median distance from the tissue section’s outer boundaries. Central labels thus eroded away faster, while the next most central labels grew preferentially to fill these vacated spaces. In the E18.5 atlas, for example, this approach allowed the basal ganglia to taper off and give way to the dorsal and medial pallium (Figure 2B). Although edges between planes remained somewhat jagged, this extension step creates a substrate for further refinement.

After completing label coverage for the left hemisphere, both the labels and underlying microscopy images were reflected across the sagittal midline to cover the remaining hemisphere. Care was taken to ensure that the sagittal midline was identified correctly by inspecting and rotating the 3D image and midline plane from multiple angles (Figure 2C). Recalculating the DSC between the microscopy images and labels for the left hemisphere showed greater similarity across all eight atlases with a median DSC improvement of 0.02 (p=0.02, WSRT) and a resulting DSC range of 0.87 to 0.97 (median 0.93, Figure 2D). Equivalent analysis of DSC for the whole brain would show substantial improvement due to the absence of labels on the right side in the original.

Label smoothing

The ADMBA atlases have been provided as 3D volumetric images, combined computationally from the original 2D sagittal reference plates; 2D sections can be generated from these 3D images in orthogonal dimensions to the original sagittal view. Visual inspection of labels in the axial and coronal planes reveals high-frequency artifacts along most edge borders, likely from the difficulty of drawing contiguous borders in dimensions orthogonal to the drawing plane (Figure 1Figure 3). To quantify the degree of label smoothness, we used the unit-less compactness metric (Ballard and Brown, 1982). The compactness measure applied in 3D incorporates both surface area and volume, allowing for quantification of smoothness to measure shape irregularity. Of note, compactness is independent of scale or orientation, facilitating comparison across all labels, despite size differences, and its sensitivity to noise allows finer detection of label irregularity (Bribiesca, 2008). Measuring the compactness of each label and taking the weighted mean based on label volume for each atlas gave a median compactness of 13,343 (mean: 34,099, standard deviation (SD): 44,988). For context, across all eight atlases the thresholded 3D whole-brain microscopy images were more compact (median compactness: 1895, mean: 2615, SD: 2305; p=0.02, WSRT, Bonferroni corrected for three comparisons; Figure 3—figure supplement 1A), consistent with the observed irregularity in the label images compared to anatomical images (Figure 2B and Figure 3).

Figure 3. Atlas smoothing and optimization by adaptive morphological filtering.

(A) Irregularities at the label edges in the original (mirrored) E18.5 atlas (top) were smoothed by applying an adaptive morphological opening filter iteratively to each label, starting from the largest to the smallest labels (bottom). 3D renderings (outer columns) are depicted for two representative labels before (top) and after (bottom) smoothing in the axial (middle left) and coronal (middle right) views. (B, C) The view in ‘A’ is repeated for E15.5 and P14, respectively. (D) To identify the optimal filter structuring element size, we devised an ‘atlas-wide smoothing quality’ metric to incorporate the balance between smoothing (compaction) and changes in size and shape (displacement). While compaction continued to improve with increasing structuring element size, displacement eventually caught up, giving an optimal atlas-wide smoothing quality with a structuring element size of 4 for the E18.5 atlas (left). We also assessed the number of labels that were lost (none in this example) and the surface area to volume ratio, for which lower values reflect smoother shapes (right). Vertical dashed lines indicate the optimal filter structuring element size, based on ‘atlas-wide smoothing quality’. (E, F) This plot is repeated for E15.5 and P14, respectively.

Figure 3.

Figure 3—figure supplement 1. Compactness of whole brain histology and labels.

Figure 3—figure supplement 1.

(A) Compactness of the whole brain histology can serve as a proxy for compactness of a biological structure since the whole brain is a well-defined structure whose boundaries are broadly accepted, unlike those of individual labels. After thresholding the microscopy images and taking their compactness, we found this compactness to be significantly lower (i.e. more compact) compared with that of the original (stripped) labels (p=0.02, WSRT, Bonferroni corrected across all three comparisons) but (B) similar to that of smoothed labels (p=1.0). (C) Original (stripped) were also significantly less compact (higher compactness) than smoothed labels were (p.=0.02).
Figure 3—figure supplement 2. Smoothing labels by Gaussian blur.

Figure 3—figure supplement 2.

(A) Example of the E18.5 atlas before (left) and after (right) smoothing by a Gaussian filter with a sigma of 0.25. (B) Smoothing quality metrics show increased compaction with a relatively slower increase in displacement with increasing filter sizes, leading to a peak smoothing quality at sigma of 1. This smoothing comes at a cost of label loss at all tested sigmas, including 15% at even the lowest tested sigma of 0.25. (C) Compared with Gaussian smoothing at this sigma across all ADMBA atlases, the adaptive opening filter approach showed a significant increase in atlas-wide smoothing quality (median 0.53 by Gaussian vs. 0.62 by adaptive opening filter; p=0.008, WSRT; mean 0.53 vs. 0.63).
Figure 3—figure supplement 3. Smoothing quality across filter sizes for all ADMBA atlases.

Figure 3—figure supplement 3.

Smoothing quality metrics for each atlas in the ADMBA across filter structuring elements sizes, with size 0 corresponding to the original (mirrored) atlas. For each atlas, the top plot depicts the overall, brain-wide smoothing quality along with its separate compaction and displacement components, with peak smoothing quality lying between filter sizes 2–7. The bottom plot for each atlas shows the overall surface area to volume across labels as well as fraction of labels lost during smoothing, typically occurring at larger sizes, starting at four or higher.

To reduce this irregularity, we applied a smoothing filter iteratively to the 3D image of each label. Prior approaches to this problem used a Gaussian filter (kernel SD of 0.5, applied in two passes) (Niedworok et al., 2016). While this visually improved the smoothness of label edges, we observed sharp angles along the edges, presumably from the limitation of rounding blurred pixel values to the integers required for label identities rather than allowing the subtler floating-point gradients from the Gaussian kernel (Figure 3—figure supplement 2A). In addition, the Gaussian filter expanded the volume of each label, leading to label loss as larger labels enveloped smaller ones (Figure 3—figure supplement 2B).

Optimal smoothing would maximize the smoothness of each label (i.e. compactness; Bribiesca, 2008) whilst minimizing changes in shape and location. It would also not lead to label loss. To calculate the improvement in compactness, we defined ‘compaction’ as the difference in compactness between the original and smoothed label over the compactness of the original label with a range from 0 or 1, with 1 being optimal. For changes in shape and location, we defined ‘displacement’ as the fraction of the smoothed label that was outside of the original label with a range from 0 or 1, with 0 being optimal. We defined a ‘smoothing quality’ metric to reflect the balance of compaction and displacement, calculated as the difference between these two measures with a range from −1 or 1, with 1 being optimal. To estimate atlas-wide smoothing quality, we took the weighted sum by volume of smoothing quality for all labels in the atlas. Assessing the quality of smoothing using Gaussian blur with increasing Gaussian sigmas, we observed label loss in all atlases in the ADMBA, even with a small sigma where labels remained visibly jagged. At this sigma value, the median atlas-wide smoothing quality across all eight atlases was 0.54 (mean 0.53), rising to a peak of 0.57 (mean 0.56) at sigma 0.5 and 0.49 (mean 0.53) at a sigma of 1.0, but with substantial loss of labels (median 14% lost at sigma 0.25, rising to 42% lost at sigma 1; mean 11% and 47%, respectively Figure 3—figure supplement 2B).

To refine smoothing while minimizing label loss, we changed the filter from a Gaussian to a morphological opening filter (Serra, 1983). This filter first erodes each label to remove artifacts, followed by dilation to restore its original volume. To avoid label loss caused by excessive erosion of small labels, we halved the size of its structuring element for labels with 5000 pixels. A few labels were split into numerous tiny fragments that would disappear with an opening filter. For these small labels, the opening filter was replaced by a closing filter, reversing the process by dilating before eroding to reconnect these components. With this adaptive opening filter approach, labels became more compact with smoother edges, while retaining their overall shape as seen in both 2D and 3D visualizations (Figure 3A).

Smoothing quality across filter sizes for all ADMBA atlases

Quantifying the improvement with the adaptive opening filter approach, using only filter sizes that completely eliminated label loss, we obtained a median atlas-wide smoothing quality of 0.61 (mean 0.62) across all eight atlases, and improvement over the Gaussian filter approach (sigma 0.25; p=0.008; Figure 3—figure supplement 2C). The optimal filter size varied between atlases, ranging from 2 to 7 (E15.5 and P14 shown in Figure 3C,D; all ADMBA shown in Figure 3—figure supplement 3). The median overall compactness improved significantly (13,343 (SD = 44,988) for unsmoothed labels vs. 2527 (SD = 2634) for smoothed labels, p=0.02, WSRT, Bonferroni corrected for the three comparisons; mean 34,099 vs. 3172) to a level that did not differ from that observed for the microscopy images of whole brains (p=1.00, WSRT, Bonferroni corrected for the three comparisons; Figure 3—figure supplement 1B,C).

Because morphological filters such as erosion classically operate globally, a drawback to these filters is the potential loss of thin structures, such as loss of the thin portion of the alar plate of the evaginated telencephalic vesicle (Figure 3A). Smoothing in-place also does not address gross anatomical misalignment. We address these issues in subsequent steps.

Label refinement by detected anatomical edges

The extending, mirroring, and smoothing steps lead to a more complete set of labels for the ADMBA and correct the irregular borders in orthogonal planes to which the original labels were drawn; however, in several locations the labels do not align closely to the anatomical edges seen in the underlying histology images, for example the basal ganglia do not follow the curve of the lateral septal nuclei (Figure 2C). To better map the anatomical fidelity of annotations in all dimensions, without manual relabeling, we leveraged our method for extending labels laterally based on gross anatomical edges to further refine all labels (Figure 4).

Figure 4. Edge-aware atlas reannotation based on underlying anatomical boundaries.

(A) Edge detection of the volumetric histology image delineated gross anatomical edges (left), shown here with the E18.5 atlas. To compare these histology-derived anatomical edges with the extended and mirrored but unsmoothed label edges (center left), we used a distance transform method to find the distance from each label edge pixel to the nearest anatomical edge (center right), shown here as a heat map of edge distances (right). (B) Eroded labels served as seeds (left) from which to grow edge-aware labels through a watershed algorithm (center left), guided by the gross anatomical edges. After smoothing, borders matched the anatomical edges more closely (center right), as shown in the edge distance heat map for the modified labels (right), using the same intensity scale. (C) The final E18.5 atlas after edge-aware reannotation and label smoothing to minimize edge artifacts. (D) To evaluate the level of edge match by label, we mapped differences in the intensity coefficient of variation weighted by relative volume for each label before and after label refinement onto each corresponding label for the E18.5 atlas (left) and across all ADMBA atlases (right). For both, the anatomical map depicts this metric as a color gradient across all of the sublevel labels present in a cross-section of the E18.5 atlas. Improvements of this metric with the refined atlas are colored in blue, minimal change is shown in white, while red represents better performance with the original atlas. (E) Applied across the full ADMBA, edge-aware reannotation and smoothing led to a significant improvement in the overall variation of intensities, taken as a weighted mean across all labels to incorporate parcellation changes while weighting by label volume (central nervous system, or ‘neural plate,’ level 0, ID 15565 in the Allen ontology; p=0.008, n = 8 atlases, WSRT). (F) Distances from labels to anatomical edges taken as the sum across all labels similarly showed a significant improvement across atlases (p=0.008, n = 8, WSRT).

Figure 4.

Figure 4—figure supplement 1. Edge distances with smoothing, edge-aware reannotation, or both.

Figure 4—figure supplement 1.

Distances between labels and gross anatomical edges in the ADMBA E18.5 atlas show improvement from the original (top left) with either smoothing (top right) or edge-aware reannotation (bottom left) alone, but even more so when combining the two approaches through edge aware reannotation followed by label smoothing (bottom right).
Figure 4—figure supplement 2. Volume changes by label from edge-aware smoothing across the ADMBA.

Figure 4—figure supplement 2.

The change in volume from each mirrored to its smoothed (edge-aware) atlas is shown by total volume for each label. Scaling is log-modulus transformed to log-scale all values including non-positive volumes. 3132 labels are present where x < 0.2.
Figure 4—figure supplement 3. Internal control using the asymmetric, contralateral hemispheres from the original atlases.

Figure 4—figure supplement 3.

(A) The asymmetry of the right hemisphere, contralateral to that of the left, originally annotated hemisphere, in the original atlas microscopy image is apparent when the mirrored labels are overlaid as illustrated here in the E18.5 atlas, notably in the olfactory bulb (top, axial view) and lateral cortex (bottom, coronal view). (B–E) Image registration of the mirrored (top, illustrated on the E18.5 atlas) or smoothed (edge-aware, bottom) atlases allows separate assessment of the original hemisphere (left) as an internal control for the contralateral (right) hemisphere within each atlas. In the original, left hemispheres, the (B) intensity coefficient of variation decreased from a median of 0.340 to 0.331 (p=0.008, WSRT) across the ADMBA, and (C) edge distances decreased from a median of 91 million to 50 million µm (p=0.008, WSRT), similarly to the full mirrored atlases. Findings were similar in the contralateral, right hemispheres, from which neither the original labels nor the 3D reconstruction were derived, where both (D) variation decreased from a median of 0.298 to 0.291 (p=0.008, WSRT) and (E) edge distances decreased from a median of 89 million to 41 million (µm, p=0.008, WSRT).
Figure 4—figure supplement 4. Lateral extension comparison with partial annotations and ground truth.

Figure 4—figure supplement 4.

(A–C) The original P28 atlas contains partially labeled lateral planes that were not included in the 3D reconstruction as seen in the axial (A), coronal (B), and sagittal (C) views. (D) To measure the accuracy of the computationally generated labels with these partially labeled planes, we manually annotated the most clearly demarcated structures in these planes (center column; brown = hippocampus, green = cortical layer 1). (E–G) Computationally generated labels in the lateral planes of the smoothed (edge-aware) atlases as seen in the sagittal (E), axial (F), and coronal planes (G). The similarity between the composite set of labels corresponding to each manual annotation as measured by the Dice Similarity Coefficient (DSC) for the original and generated labels increased with the computationally generated labels, more prominently in layer 1 and the average weighted by volume. Manual annotations were cropped to the extent of all labels to only compare the partially labeled original planes where they were annotated.
Figure 4—figure supplement 5. Edge-aware adjustment in delicate sub-regions with subtle contrast.

Figure 4—figure supplement 5.

(A) The ventromedial hypothalamus (VMH) is a region with low contrast as shown here in the microscopy images (top row) in each orthogonal view of the P28 atlas along with its original labels (upper middle row; originally annotated, sagittal planes are shown in the left column alongside their orthogonal planes in the center and right columns). Erosion of each label (lower middle row) preserves the core of highest confidence annotation while removing artifacts that were particularly visible in the axial and coronal views. Eroded label edges near anatomical boundaries (light purple lines) take advantage of them during watershed-based regrowth (bottom row; see the green structure), while edges far from these boundaries meet in-between neighboring labels (green and tan structures). The relatively coarse anatomical edge map captures only clearly demarcated boundaries with strong contrast differential, limiting the number of artifactual boundaries around low-contrast regions, while the erosion and regrowth remove artifacts in the axial and coronal views. (B) Similar views in the pontine hindbrain.
Figure 4—figure supplement 6. Preservation of regions demarcated by in-situ hybridization markers.

Figure 4—figure supplement 6.

(A) Wnt3a is selectively expressed in the cortical hem at E13.5 (orange structure) and used by Thompson et al., 2014b to define its boundaries based on in-situ hybridization (ISH). The region corresponding to Wnt3a (dashed circle) remains contained within the cortical hem in the original (middle row) and 3D reconstructed (bottom row) atlases. (B) Similarly, the Hoxa2 ISH signal was used to demarcate the border between the pontine hindbrain (PH) and prepontine hindbrain (PPH) in the E15.5 atlas. This boundary remains well-demarcated before (lower middle row) and after (bottom row) 3D reconstruction.

Using the same gross anatomical map in 3D (shown for an example plane in Figure 4A, second from left), we first quantified the distances from 3D label edges to the expected anatomical position. Assuming that the nearest gross anatomical edge was the correct one, we measured the distance from each label edge to the nearest gross anatomical edge. We can visualize this distance as a color gradient, in which higher intensity of color represents a greater distance to each anatomical edge (Figure 4A, right columns).

To modify the labels in light of the gross anatomical edge map, we again incorporated the edge-aware algorithm, this time in 3D. We made the assumption that the core of each label is annotated accurately, whereas label edges are more prone to inaccuracies from plane-to-plane misalignments or the difficulty of assessing histological edges. To preserve the core while reannotating the periphery, we first eroded each label to remove edges. These eroded labels became the seed for the watershed step, which re-grew labels back toward their original size but now guided by the gross anatomical edge map.

Normally, the erosion step would lead to loss of thin structures within labels because erosion operates globally on the entire label. To preserve these thin structures, we skeletonized each label in 3D, which thins the label to its core structure (Lee et al., 1994), and added the skeleton back to the eroded label. We used a much more lightly eroded version of the label for the skeletonization to avoid retaining label edges in the skeleton. By combining erosion with skeletonization when generating seeds for the watershed, we retained thin structures such as the alar plate of the evaginated telencephalic vesicle located anterior to the basal ganglia.

After performing this edge-aware step, we ran the adaptive morphological opening filter smoothing step (Figure 4B). Because the edge-aware step partially smooths structures, we could use smaller filter sizes for smoothing to avoid loss of thin structures. The resulting labels show considerable improvement, for example the basal ganglia now curve around the lateral septal nuclei (Figure 4C). By adapting the morphological filter sizes during both the edge-aware and final smoothing steps, we avoid label loss and minimize volume changes relative to label size as seen in the smaller labels (Figure 4—figure supplement 2). Visualization of the color gradient of distances to anatomical edges also confirms substantial improvement in label alignment compared with the original labels or smoothing or edge-aware steps alone (Figure 4—figure supplement 1B). To quantify this improvement brain-wide, we calculated the sum of edge distances for each pixel at label surfaces across the ADMBA. We observed a significant reduction from a median of 187 million to 86 million µm (p=0.008, WSRT; Figure 4B,C,F), with a median Dice Similarity Coefficient between original (mirrored) and smoothed (edge-aware) labels of 0.76 (mean 0.80) and 9% median (mean 12%) volume reassignment. Example planes from all atlases in the ADMBA before and after refinement are depicted in Figure 5, and movies across all planes are shown in Figure 5—videos 116.

Figure 5. Original and 3D reconstructed atlas comparison across the ADMBA.

(A-H) Representative axial planes from all atlases. For each pair of images, a plane of the original (left) atlas is depicted next to the refined (right) atlas after undergoing the full refinement pipeline. Complete atlases before and after refinement are shown as movies in Figure 5—videos 116.

Figure 5.

Figure 5—figure supplement 1. Piecewise 3D affine transformation of the spinal cord in the E11.5 atlas.

Figure 5—figure supplement 1.

(column 1) The distal cord of the ADMBA E11.5 atlas specimen is skewed laterally as seen in the axial view (top), preventing mirroring without duplicating the distal cord (bottom). (columns 2–4) A series of three successive piecewise affine transformations straightened the cord to avoid duplication with mirroring.
Figure 5—video 1. Original ADMBA E11.5 Atlas.
Download video file (1.3MB, mp4)
Figure 5—video 2. 3D reconstructed ADMBA E11.5 Atlas.
Download video file (930.7KB, mp4)
Figure 5—video 3. Original ADMBA E13.5 Atlas.
Download video file (6.4MB, mp4)
Figure 5—video 4. 3D reconstructed ADMBA E13.5 Atlas.
Download video file (1.7MB, mp4)
Figure 5—video 5. Original ADMBA E15.5 Atlas.
Download video file (15.2MB, mp4)
Figure 5—video 6. 3D reconstructed ADMBA E15.5 Atlas.
Download video file (2.3MB, mp4)
Figure 5—video 7. Original ADMBA E18.5 Atlas.
Download video file (2.5MB, mp4)
Figure 5—video 8. 3D reconstructed ADMBA E18.5 Atlas.
Download video file (1.8MB, mp4)
Figure 5—video 9. Original ADMBA P4 Atlas.
Download video file (5.6MB, mp4)
Figure 5—video 10. 3D reconstructed ADMBA P4 Atlas.
Download video file (3.2MB, mp4)
Figure 5—video 11. Original ADMBA P14 Atlas.
Download video file (6.5MB, mp4)
Figure 5—video 12. 3D reconstructed ADMBA P14 Atlas.
Download video file (4.4MB, mp4)
Figure 5—video 13. Original ADMBA P28 Atlas.
Download video file (10.5MB, mp4)
Figure 5—video 14. 3D reconstructed ADMBA P28 Atlas.
Download video file (9.2MB, mp4)
Figure 5—video 15. Original ADMBA P56 Atlas.
Download video file (8MB, mp4)
Figure 5—video 16. 3D reconstructed ADMBA P56 Atlas.
Download video file (6MB, mp4)

Anatomical edges in microscopy images reflect differences in intensity between regions. Therefore, we would expect accurate labels to have smaller variation in intensity in the underlying microscopy images than inaccurate labels, although such a difference would need to be apparent even though the majority of the label is unchanged. We used the coefficient of variation of intensity values within each label to quantify this expectation and demonstrated significantly lower variation with edge-aware labeling (median from all labels weighted by size decreased from 0.290 to 0.282, p=0.008, for all eight atlases, WSRT; Figure 4D,E). Furthermore, edge-aware labeling decreased the absolute coefficient of variation for 92 of the 100 individual labels represented by all atlases. The few labels that showed increased variation were frequently in regions of relatively subtle intensity changes, such as the hindbrain, where histological edges were less well-defined (Figure 4D).

As an internal control, we also registered the atlases to their original, asymmetric microscopy images to compare the impact of atlas reconstruction in the originally labeled hemispheres and the contralateral, asymmetric hemispheres on which the atlases were not derived and found similar improvement in anatomical fidelity (Figure 4—figure supplement 3). We also compared the computationally generated labels and partially labeled lateral regions not included in the 3D reconstruction with manually annotated regions in the P28 atlas, which showed an increase in the Dice Similarity Coefficient in the reconstructed labels (Figure 4—figure supplement 4). To explore how the edge-aware approach handles delicate sub-regions, which may not have clearly demarcated anatomical boundaries, we focused in greater depth on two complicated sub-regions in the P28 atlas (Figure 4—figure supplement 5) and two regions described in the original ADMBA paper (Thompson et al., 2014b) as demarcated by in-situ hybridization markers in the E13.5 and E15.5 atlases (Figure 4—figure supplement 6), showing that if anatomical data does not support moving the boundary, our approach smooths the boundary in the axial and coronal planes without moving the location in any of the planes.

Application to tissue-cleared whole brains

A major goal of 3D atlas refinement is for automated label propagation to optically sectioned, volumetric microscopy images generated from intact tissue using recently developed tissue clearing techniques. As the accuracy of atlas registration is ultimately dependent on the fidelity of the underlying atlas in all dimensions, we sought to test and quantify improvement from our atlas refinements in cleared mouse whole brains.

Among the many available methods to clear whole organs, we chose CUBIC (Susaki et al., 2014; Susaki et al., 2015) given its balance of overall transparency, morphological retention with minimal tissue expansion, and ease of handling as a passive aqueous technique (Kolesová et al., 2016; Xu et al., 2019a; Bossolani et al., 2019). After clearing C57BL/6J WT mouse pup brains (age P0) with simultaneous incubation in SYTO 16 nuclear dye for 2 weeks, we imaged intact whole brains by lightsheet microscopy at 5x to obtain volumetric images at cellular resolution (Figure 6—figure supplement 1), taking approximately 3 hr to image and generating approximately 500 GB of data per brain (n = 15; 10 male, 5 female).

To detect nuclei throughout cleared whole brains automatically, we implemented a 3D blob detection algorithm using the Laplacian of Gaussian filter, which has been shown to work well in equivalent image data (Schmid et al., 2010; Figure 6). To make the nuclei approximately spherical for blob detection, we interpolated the images axially to match the lateral resolution. Due to the large quantity of data, processing was performed in parallel on small chunks. Preprocessing and detection settings were optimized using hyperparameter tuning against a ‘truth set’ of 1118 nuclei selected from multiple brain regions that had been verified by manual visualization (Figure 6—figure supplement 2A,B). The resulting model achieved a recall (sensitivity) of 90% and precision (positive predictive value) of 92%. Furthermore, the model showed high correlation with total lightsheet microscopy intensity levels brain-wide (r > 0.99, p≤1 × 1016, Spearman’s rank correlation coefficient (SRCC) for both original (mirrored) and smoothed (edge-aware) atlases) and nuclei vs. intensity densities (original: r = 0.89; smoothed: r = 0.93; p≤1 × 1016 for both, SRCC), suggesting the performance was accurate outside the narrow target of the truth set (Figure 6—figure supplement 2C–E).

Figure 6. 3D whole-brain nuclei detection pipeline.

(A) Overview of the nuclei detection and assignment pipeline. After tissue clearing by CUBIC and nuclear staining, we imaged intact whole P0 mouse brains by lightsheet microscopy (Figure 6—figure supplement 1). The E18.5 mouse brain atlas was registered to volumetric sample images. Nuclei were identified brain-wide with a 3D blob detector algorithm, and the nuclei coordinates were used to map to specific region labels, allowing quantification of nuclei counts and density for each label. (B) Axial view of the hippocampus of a P0 mouse brain at the original magnification of 5x (left) and zoomed to 20x (right), with the same region of interest (ROI) depicted by the yellow box in both. (C) Serial 2D planes along the ROI shown by the yellow box in part ‘B’ show the emergence and disappearance of nuclei while progressing vertically (z-axis) through the ROI. Each of four individual nuclei is assigned a unique colored circle. The most central image of the nuclei is indicated by a solid circle and asterisk, while images above or below have dashed circles. (D) 3D schematic of the four nuclei from part ‘C’ demonstrating their spatial orientation.

Figure 6.

Figure 6—figure supplement 1. Custom 3D printed specimen mount for suspending samples in the Zeiss Lightsheeet Z.1 microscope.

Figure 6—figure supplement 1.

(A) Mount schematic and orientation in the chamber. The lightsheet chamber requires the specimen to be suspended between the two illumination objectives producing a plane of light and detection objective perpendicular to this plane. To hold the brain in place while immersed in the refractive index matching solution, we devised a mount with a platform to maximize surface contact with the ventral surface of the brain, attached by super glue. To allow full latitude of the specimen within the chamber, we added a mount neck to center the specimen between the illumination objectives when the manipulator was approximately centered. (B) Photograph of mount and custom mount holder to facilitate gluing the specimen to the mount platform. Cotton buds were used to dry the ventral surface of the brain before attachment. We used the rod end of another mount to gently press the brain into the glue. (C) Photograph of specimen suspended and centered between the illumination objections, facing the detection objective.
Figure 6—figure supplement 2. Nuclei detection optimization and comparison with intensity.

Figure 6—figure supplement 2.

(A) Example receiver operating characteristic (ROC) curves. (top) Example ROC of nuclei detection across varying isotropic scaling factors along the z-axis. Scaling is shown as the fraction of isotropy along the z-axis, where 1 = isotropic. For example, 0.5 indicates that each ROI was resized so that the z-axis was half of the size it would need to be for full isotropy with the x- and y-axes. (bottom) Example ROC evaluating two hyperparameters, the minimum and maximum standard deviations for the Gaussian kernel used for multi-scale detection, to find the optimal size bounds for blob detection. Inset shows a zoomed view of the clustered points. (B) Total nuclei vs. intensity by label showed a linear relationship using labels defined by either the original (mirrored; r = 0.997, p≤1×1016) or smoothed (edge-aware; r = 0.997, p≤1×1016) atlases. In the correlation plot (bottom), the size of the circle indicates the size of the correlation coefficient (r). (C) Similarly, nuclei density vs. intensity density relationships were approximately linear in the original (mirrored; r = 0.894, p≤1×1016) and smoothed (edge-aware; r = 0.929, p≤1×1016) atlases. (D) Sensitivity and False Discovery Rate stratified by nuclei density in manually annotated truth sets as seen in the histogram (left) showed a similar distribution by region (middle) and overall metrics (right) by density group, with numerically lower recall and slightly higher precision in denser regions.
Figure 6—figure supplement 3. Registration of atlas to WT brains.

Figure 6—figure supplement 3.

(A) Overlays of each registered atlas microscopy image on its corresponding downsampled wild-type brain. All examples are shown at the same plane. (B) Composite image of wild-type sample brains registered in reverse, using the same settings but registering the sample to the atlas to evaluate overall alignment of registered brains. (C) Overall registration quality as expressed by the DSC between the foreground of each sample and its registered atlas showed a median of 0.91 (mean 0.90).

Using the E18.5 brain as the closest atlas to our P0 C57BL/6J wild-type mouse brains (birth typically at E19 Murray et al., 2010), we then registered the atlas volumetric histology images to each sample brain. We chose to register brains using the Elastix toolkit (Klein et al., 2010; Shamonin et al., 2014), which has been validated on CUBIC cleared brains and balances computational efficiency and accuracy (Nazib et al., 2018). After downsampling the volumetric images to the same dimensions as the atlas for efficiency, we applied rigid, followed by non-rigid, deformable registration based on a multi-resolution scheme, optimizing the cross-correlation similarity metric as implemented by the SimpleElastix (Nazib et al., 2018; Marstal et al., 2016) programmable interface to the Elastix toolkit (Klein et al., 2010; Hammelrath et al., 2016). After registration, the median DSC between the registered atlas volumetric histology images and the lightsheet microscopy images of each sample brain was 0.91 (mean: 0.91; 95% confidence interval: 0.90–0.92; Figure 6—figure supplement 3), although inspection of registered brains also revealed slight misalignments with the atlas, mostly in caudal regions including the cerebellum, a structure known to be challenging for registration (Nazib et al., 2018). Variations in sample preparation, including the completeness of the dissected hindbrain and expansion of the third ventricle during tissue clearing may also have contributed to these misalignments.

To evaluate whether our updated, smoothed 3D labels improved the analysis of true volumetric data, we registered both the original and refined labels to these 15 tissue-cleared, nuclei-labeled whole wild-type mouse brains (10 male, five female). We would expect improved 3D labels to correspond more closely to the underlying anatomy. To test this expectation, we generated gross anatomical edge maps for each cleared brain and measured distances from label borders to anatomical edges. Edge distances significantly improved for almost all labels (overall decrease from a median of 153 million µm to 96 million µm, p=0.007, WSRT, Bonferroni corrected for all 120 labels across all hierarchical levels, mean 152 million to 99 million µm; Figure 7A,C–D; Figure 7—figure supplement 1A). We would also expect improved 3D labels to have lower variation in image intensity within each label, as we observed in assessing the refined labels with the original brain images from which they were derived (Figure 4E). We observed a small, but significant decrease in the intensity coefficient of variation at the whole-brain level (0.309 to 0.301, p=0.007, WSRT, Bonferroni corrected for the 120 comparisons, mean 0.311 to 0.304) and most sub-labels (Figure 7—figure supplement 1B). For 22 labels (18% of all labels) the variability worsened, however this was only significant (p=0.04) for a single label. The majority of these 22 labels describe small structures (combined 7% of total volume) and therefore sensitive to slight perturbations in border location, and located ventrally in the brain, where signal was prone to being distorted by glue artifacts from the mount for lightsheet imaging. For context, variability improved for 98 labels (82% of all labels) with 27 labels showing significant improvement.

Figure 7. 3D reconstructed atlas validation in wild-type mouse brains at cellular resolution.

(A) Example of the original (mirrored, left) and smoothed (edge-aware, right) E18.5 atlases registered to a representative cleared, optically-sectioned P0 wild-type mouse brain. Edge distances between the registered labels and the brain’s own anatomical edge map are reduced for this example brain, shown by the color gradient for each edge (bottom). (B) Label alignment at higher resolution. The top row depicts the registered original (mirrored) atlas at 20x and 100x around a region of interest highlighted by a yellow box. This same brain region is depicted in the bottom row, but overlaid with the refined atlas registered using identical transformations as in the original atlas. (C) Summation of the edge distances across all 15 wild-type brains with color gradient showing the edge distances with the original (top) and smoothed (bottom) labels. (D) Total edge distance at the whole-brain level before and after atlas refinement for each of these brains. (E) Density-based nuclei clustering within each label. Many of the isolated nuclei that could not be clustered in the original labels were clustered in the refined labels, with differences in unclustered nuclei shown from 16 regions selected across the hierarchy of labels from the grossest (neural plate, L0, left) to the finest (e.g. dorsal pallidum, L7, right). Error bars represent 95% confidence intervals. (F) The differences between the original and refined labels’ unclustered nuclei are depicted on an anatomical map showing this metric as a color gradient across all the sublevel labels present in a cross section. Improvements with the refined atlas are colored in blue, while red represents better performance with the original atlas. A complete list of differences for each metric in each label is provided in Supplementary file 2.

Figure 7.

Figure 7—figure supplement 1. Region homogeneity in wild-type brains based on the original vs.refined atlases.

Figure 7—figure supplement 1.

(A) The differences between the original (mirrored) and smoothed (edge-aware) labels’ edge distance sum and (B) intensity and (C) nuclei density coefficients of variation as reflections of region homogeneity are quantified with bar plots (left) for selected regions across the hierarchy of labels. Error bars represent 95% confidence intervals. Anatomical maps depict these metrics as a color gradient across all of the finest labels present in a cross section (right). Increases in a metric with the original atlas are colored in blue, while red represents increases with the refined atlas.
Figure 7—figure supplement 2. Changes in volumes, nuclei, and densities across wild-type brains based on atlas labels.

Figure 7—figure supplement 2.

(A) The differences between the original (mirrored) and smoothed (edge-aware) labels’ volume, (B) nuclei, and (C) nuclei densities (bottom) are quantified with bar (right) and volcano (left) plots. Selected regions are shown across the hierarchy of labels from the grossest (neural plate, L0, left) to the finest (e.g. dorsal pallium, L7, right). Positive values indicate decreased size in the refined atlas. Error bars represent 95% confidence intervals. Volcano plots depict each label as a separate point with point size correlating with label size and colors corresponding to major parent structure (neural plate = light blue, forebrain = dark blue, midbrain = light green, hindbrain = dark green). Points higher and farther from 0 have stronger statistical significance and effect size, respectively. p-Values are calculated by the WSRT, Bonferroni corrected for all 120 labels within each measurement. Effect sizes are standardized based on the WSRT Z-statistic.
Figure 7—figure supplement 3. Nuclei clustering in the original vs. refined atlases.

Figure 7—figure supplement 3.

(A-B) To determine the appropriate cluster neighborhood distance parameter, the fifth nearest neighbor distances for all nuclei in each wild-type brain are plotted in ascending order (right), following the convention in DBSCAN of taking the 2·ndim−1th neighbor (Schubert et al., 2017). (left) The zoomed view allows identification of the distance at the ‘elbow’ or ‘knee,’ or maximum curvature, typically at 20 µm or higher. (C) Nuclei clustering within each label in the smoothed (edge-aware) atlas as a fraction of that of the original (mirrored) atlas across a range of distance values. The number of isolated, unclustered nuclei and total clusters, indicative of nuclei isolation and cluster fragmentation, respectively, decreased in the smoothed atlas relative to the original atlas, starting even prior to the minimum distance based on the max curvature seen in part ‘B.’ (E) Example using a conservative neighbor distance of 20 µm for clustering in a brain before (top) and after (bottom) label refinement. Light gray lines show label boundaries. Unclustered nuclei are depicted as white points. All other points are clustered nuclei. Within each region, nuclei are colored by cluster size, where blue represents nuclei in the largest cluster, followed by orange, green, and red for the next successively smaller clusters. (D) Differences in unclustered nuclei and (E) number of clusters by region across all wild-type brains with this distance setting for selected regions across label hierarchies in bar plots (left) and as color gradients on an anatomical map (right).
Figure 7—figure supplement 4. MagellanMapper software GUI screenshots.

Figure 7—figure supplement 4.

(A) ROI serial 2D viewer and annotation editor GUI. The viewer provides overview images at increasing magnifications, zooming into the ROI outlined in yellow. Overview images are scrollable to show how the z-planes change through the ROI in place. Sequential z-planes of the ROI alone depict the original 2D images side-by-side through the entire 3D ROI, with a few additional opacified planes shown above and below the ROI for context. Overlaid labels show an example of automated nuclei segmentation. (B) Another example of the ROI viewer, here in nuclei annotation and verification mode. Automated nuclei centroid ‘blob detections’ are depicted as interactive circles corresponding to blob positions and radii. For building truth sets, circles provide drag-n-drop and copy-paste controls to reposition, add, or subtract these detections for accuracy. Green and red flags allow scoring for detection correctness. To compare with truth sets, blue and purple dots depict previously annotated nuclei positions pulled in from a database. Automated verification of current blob detections against the truth set shows blue dots as correctly detected truth blobs, purple as missed truth blobs, green circles as correct detections based on a matched truth blob, and red as incorrect detections (no remaining truth blob matches). (C) ROI selector and 3D visualization GUI. The right panel displays ROI offset and size controls, 2D and 3D display options, label controls including ontology depth, and an editable blob table. The left panel shows an example whole atlas 3D visualization using the ADMBA E11.5 atlas. (D) Atlas editor GUI. The GUI provides simultaneous orthogonal views in all three dimensions, with planes corresponding to the crosshairs. With ‘Editing’ selected, labels can be painted into adjacent labels as shown in the intermediate stratum of Str label in the left hemisphere of the z-plane. Editing a second, non-contiguous z-plane in the same label will enable the ‘Fill’ button to perform edge interpolation for this label through all intervening planes.

These two analyses support an overall improved performance of the refined 3D labels with the volumetric images, however we can also test nuclei density - a key use case for whole-brain imaging methods (Murakami et al., 2018). After detecting nuclei throughout each brain (Figure 6), we assigned each nucleus to an atlas label based on the 3D location of the nucleus. Using the numbers of nuclei per label and the volume per registered label, we calculated nuclei densities using the original and refined labels (Figure 7—figure supplement 2). As with the volumetric image data, we would expect accurate labels to encapsulate regions with more constant nuclei density. We therefore assessed the coefficient of variation for nuclei density and observed a small, but significant improvement with the refined labels overall (median 0.629–0.625, p=0.007, WSRT, Bonferroni corrected for the 120 hierarchical labels, mean 0.629–0.625) and across the majority of labels in the hierarchy (Figure 7—figure supplement 1C). A median of 13.5% (4.7 million) nuclei (mean 13.2% [4.6 million] nuclei) were reassigned from original to refined labels. Examples of these nuclei reassignments in a wild-type mouse forebrain are depicted in Figure 7B. As an independent assessment of label alignment based on nuclei density alone, we measured nuclei clustering within each label under the expectation that well-aligned labels would group nuclei of similar densities, whereas poorly aligned labels would form isolated pockets of nuclei along borders between regions of contrasting nuclei densities. We employed Density-Based Spatial Clustering of Applications with Noise (DBSCAN) (Ester et al., 1996), an algorithm useful for both clustering and noise or anomaly detection (Ali et al., 2010), and indeed found that label refinement reduced isolated, unclustered nuclei by 30% (median 4.4 × 105 to 3.1 × 105 nuclei; p=0.007, WSRT, Bonferroni corrected for the 120 labels; mean 4.4 × 105 to 3.1 × 105; Figure 7E–F; Figure 7—figure supplement 3), suggesting greater nuclei assignment to labels with nuclei clusters of similar density.

MagellanMapper as a tool

To facilitate both automated atlas refinement and visual inspection of large 3D microscopy images, we developed the open-source MagellanMapper software suite. This Python-based (Oliphant, 2007; Millman and Aivazis, 2011) image processing software consists of both a graphical user interface (GUI) for visualization and a command-line interface for automation (Figure 7—figure supplement 4). A major goal of the suite’s GUI is to enable unfiltered, raw visualization and annotation of original 2D images in a 3D context. The GUI consists of three main components: (1) A region of interest (ROI) selector and 3D visualizer with 3D rendering, (2) A serial 2D ROI visualizer and annotator, tailored for building truth sets of 3D nuclei positions, and (3) A simultaneous orthogonal plane viewer with atlas label editing tools, including painting designed for use with a pen and edge interpolation to smoothly fill in changes between two edited, non-adjacent planes. The command-line interface provides automated pipelines for processing of large (terabyte-scale) volumetric images, including 3D nuclei detection at full image resolution.

The suite makes extensive use of established computer vision libraries such as scikit-image (van der Walt et al., 2014), ITK (Yoo et al., 2002) (via SimpleITK Lowekamp et al., 2013 and SimpleElastix Klein et al., 2010; Marstal et al., 2016), visualization toolkits such as Matplotlib (Hunter, 2007) and Mayavi (Ramachandran and Varoquaux, 2011), and statistical and machine learning libraries such as Numpy (van der Walt et al., 2011), SciPy (Virtanen et al., 2020), Pandas (McKinney, 2010), and scikit-learn (Pedregosa et al., 2018). The cross-platform capabilities of Python allow the suite to be available in Windows, MacOS, and Linux. We also leverage additional open-source imaging ecosystems such as the ImageJ/Fiji (Schindelin et al., 2012; Rueden et al., 2017) suite for image stitching through the BigStitcher plugin (Hörl et al., 2019), integrated into our automated pipelines through Bash scripts.

Discussion

3D reference atlases are critical resources for interpreting and comparing anatomically complex tissues, especially as techniques for whole-tissue microscopy (Ragan et al., 2012; Murakami et al., 2018) and in-situ single-cell transcriptome methodologies are refined (Nowakowski et al., 2017; Polioudakis et al., 2019). Considerable effort has been invested in generating 2D reference atlases, while manually building atlases in 3D is time-consuming and laborious; building 3D atlases from these 2D resources will save time and improve comparability between analyses. Here, we have presented a pipeline to automate the completion and conversion of atlases from 2D sections to complete, anatomically aligned 3D volumes and applied these methods to the eight developing mouse brain atlases in the ADMBA (Figure 5).

Key features of our pipeline include, first, an approach to adaptive morphology for label smoothing while preserving image detail by combining existing skeletonization, morphological erosion, and watershed segmentation methods and optimizing their parameters for application to 3D atlas reconstruction. Second, we utilized this method as an edge-aware label refinement approach, which incorporates anatomical edge detection from an underlying intensity-based image to perform two functions: (1) Label extension into incompletely labeled regions, and (2) Anatomically guided label curation. Third, we integrated these approaches into an automated pipeline as a general way to refine serial 2D labeled datasets into 3D atlases. We next applied the pipeline to complete and refine the full suite of 8 atlases in the Allen Developing Mouse Brain Atlas series from ages E11.5 to P56 as a resource to the neuroscience and neurodevelopmental communities (Figure 5). Specifically, these tools extended incomplete labels in each atlas to cover the entire brain and smooth artifacts between 2D sections, whilst respecting anatomical boundaries. We showed that the refined labels are more complete (Figure 2), the interface between labels match anatomical images more closely (Figure 4), and the image intensity within these labels is more homogeneous (Figure 4). Using whole-brain lightsheet microscopy of 15 mice at P0, we identified the 3D location of 35 million nuclei per brain (median across 15 samples; Figure 6). We showed that the refined labels match anatomical boundaries better at cellular resolution, providing a key step towards comparative analysis of whole-brain cytoarchitecture (Figure 7). Fifth, we have provided the pipeline as an open source, graphical software suite with an end-to-end automated image processing pipeline for refining atlases, detecting nuclei, and quantifying nuclei counts by anatomical label in raw microscopy images of whole tissues.

Building a 3D reference atlas from a physically sectioned atlas presents three main challenges: (1) Incomplete labels resulting in unidentified tissues, (2) Smoothing jagged edge artifacts between sections in two of the three dimensions, and (3) Maintaining anatomical boundaries through the smoothing process. Prior work has tackled these challenges through two approaches. The first is automated smoothing of label boundaries (Niedworok et al., 2016); however, this approach has well recognized limitations (Chen et al., 2019), including labels crossing anatomical boundaries, loss of label detail, and complete loss of small labels (Figure 3—figure supplement 2). The second approach uses a combination of novel data generation and copious, manual curation as applied by neuroanatomists and illustrators to version 3 of the Allen Common Coordinate Framework (CCFv3, 2017) (Allen Institute for Brain Science, 2017; Wang et al., 2020). The resulting 3D reference atlas is highly accurate; however, an even greater large-scale effort would be required for the eight stages of development in the ADMBA.

Our method automates additional curation steps to extend and smooth labels guided by the underlying anatomy, providing an automated method closer to the accuracy of manual curation. To address loss of detail and loss of labels, we used the combination of adaptive morphology, an edge-aware watershed, and skeletonization to preserve small structures during automated smoothing. By fitting and extending labels to anatomical boundaries, we improved both the accuracy and label coverage of the resulting 3D reference atlas, without requiring additional data. While manual curation remains the gold standard, as performed in the Allen CCFv3, even this monumental effort translated only 77% of labels because of resource limitations (Wang et al., 2020). As suggested by Wang et al., future atlas development lies in automated, unbiased approaches (Wang et al., 2020). We demonstrate qualitatively (Figure 5) and quantitatively (Figure 4) that our automated approach is a substantial refinement over existing reference atlases and creates a complete resource that is immediately applicable for registration of 3D imaging data (Figure 7). Furthermore, it acts as a foundation for further improvements, substantially reducing the work required to generate a manually curated atlas in the future while freeing up anatomists and neuroscientists for the role of oversight in such endeavors.

Despite these substantial improvements, we are aware of some specific limitations of the approach. We expected less variability in image intensity within a region than between regions, and therefore improved region labels should reduce image intensity variability. While this expectation was correct for the majority of labels (92% of labels present in all atlases), a small number of regions showed a modest increase in variability with our refined labels. Most of these labels were in the hindbrain, where histological contrast is lower, leading to less well-defined anatomical edges. Adjusting the weight of edge-detected boundaries, based on the degree of underlying intensity contrast, could lessen the impact of these edges on reannotation. A few more labels exhibited this paradox when registered to the wild-type brains, likely owing to artifacts from tissue preparation and mounting, the slight age mismatch between the closest available E18.5 atlas and the P0 wild-type brains, registration misalignments, and inaccuracies of the blob detector.

Another limitation is the use of a single filtering kernel size for smoothing across the entire brain (Figure 3). In labels with both thick and thin portions, the thin portions are prone to being lost during the erosion step of smoothing. While a fully adaptive morphological approach would be beneficial, our use of skeletonization and the edge-aware watershed allowed us to preserve the core of such thin portions. More generally, the combination of erosion, skeletonization, and watershed could serve as an adaptive form of morphological operations to better preserve any label’s overall footprint and topological consistency.

While the refined labels are superior to the original ones in both the ADMBA images and whole-brain lightsheet microscopy images, we observed more registration artifacts with the whole-brain lightsheet microscopy images (Figure 7A). Undoubtedly, biological variation, the slight age mismatch (E18.5 in ADMBA vs. P0 for lightsheet), and the different preparations and imaging modality (Nissl sections vs. nuclei-stained whole brain) are substantial factors. Generating a group-wise composite of nuclei-stained whole brain could help reduce biological variation. Deferring the edge-aware reannotation step until after registration, using boundaries identified in the nuclei-stained whole brain, may also reduce these artifacts.

Recent methodological advances may further improve the refined 3D labels. Chon et al., 2019 integrated the Franklin-Paxinos atlas with the CCFv3 atlas and additional cell-type specific markers, addressing discrepancies between these standard references. Integrating our anatomical edge maps with in-situ hybridization expression patterns (Allen Institute for Brain Science, 2017; Thompson et al., 2014b) may similarly improve label boundaries. A deep learning model utilized histological textures to generate labels (Chen et al., 2019). Training the model required substantial manual annotation to serve as the initial ground truth; in this context, our automated approach to atlases generation may be able to replace or augment this manual step. Another recent approach used fractional differential operators to preserve texture as well as edge information for segmentation of single-cell resolution data (Xu et al., 2019b). This could be incorporated into the 3D anatomical edge map generation step to further delineate labels at higher or even full resolution.

Conclusion

Mouse whole-brain imaging has been used to understand models of human traits, including sexual dimorphism (Kim et al., 2017) and models of human disorders, including Alzheimer’s disease (Liebmann et al., 2016; Delafontaine-Martel et al., 2018; Whitesell et al., 2019), serotonin dysregulation (Ellegood et al., 2018), epilepsy (Intson et al., 2019), and autism spectrum disorder (ASD) (Suetterlin et al., 2018). Such analyses would be augmented by accurate 3D reference atlases, allowing the detection of subtle quantitative changes not readily appreciable in individual slices. As we have described and demonstrated here, anatomically guided completion and refinement of reference atlases more fully leverages the many existing and established atlases by expanding them to full coverage and greater accuracy at cellular resolution. The completed and refined 3D ADMBA serves as a resource to help identify biological differences in the multiple models of human disorders and traits across brain development.

Materials and methods

Key resources table.

Reagent type
(species) or resource
Designation Source or reference Identifiers Additional information
Genetic reagent
(Mus. musculus)
C57BL/6J Jackson
Laboratory
Stock #: 000664
RRID:MGI_3028467
Commercial assay or kit SYTO 16 Thermo Fisher
Scientific
Cat. #: S7578 1 µM
Software, algorithm FIJI/ImageJ software FIJI/ImageJ RRID:SCR002285 Version 1.52
Software, algorithm MagellanMapper software This paper
Software, algorithm Zeiss Zen software Zeiss RRID:SCR_018163 Version 2014

Original atlases

The ADMBA series provides atlases for multiple developmental time points from embryonic through adult stages (Thompson et al., 2014b). Each atlas consists of two main images given in a volumetric format (.mhd and its associated .raw file), a microscopy image (‘atlasVolume’) and a labels image (‘annotation’), which are each composed of multiple 2D sagittal planes. As outlined in the ADMBA technical white paper (Allen Institute for Brain Science, 2013a) and online application programming interface (API) documentation (Allen Institute for Brain Science, 2018), each microscopy plane is from imaging of a C57BL/6J mouse brain specimen cryosectioned into 20 µm (E11.5-P4) or 25 µm (P14-P56) thick sagittal sections stained with Feulgen-HP yellow nuclear stain (E11.5-E18.5) or Nissl staining (P4-P56), and imaging planes were assembled into a 3D volume. To create the atlas labeling, an expert anatomist used Adobe Illustrator CS to annotate 2D planes, which were interpolated to create annotations in 3D (Thompson et al., 2014b; Allen Institute for Brain Science, 2013b; Allen Institute for Brain Science, 2018). The annotation planes correspond to each microscopy plane, with integer values at each pixel denoting the label at the finest structural level annotated for the corresponding microscopy voxel.

Lateral labels extension

While the ADMBA covers a large proportion of unique areas within each brain, the atlas labels leave the lateral edges of one hemisphere and the entire other hemisphere unlabeled. To fill in missing areas without requiring further manual annotation, we made use of the existing labels to sequentially extend them into the lateral unlabeled histology sections. This process involves (1) resizing the last labeled plane to the next, unlabeled plane, (2) refitting the labels to the underlying histology, and (3) recursively labeling the next plane until all unlabeled planes are complete (Figure 2—figure supplement 2).

To extend lateral edges of the labeled hemisphere, we first identified the lateral-most labeled plane of the atlas microscopy images. We started from the first sagittal plane on the labeled side of the image and moved toward and into the brain, checking each plane for the presence of any label. Once we identified a contiguous stretch of planes with labels, we used the most lateral labeled plane as the template for subsequent labels to be extended out laterally, in the opposite direction. In a few atlases (e.g. P28), the lateral-most labeled planes are only partially complete, in which case the most lateral completely labeled plane was manually specified instead.

This last labeled lateral plane contained one or more discrete structures to extend label coverage, typically the cortex and sometimes the cerebellum. To find each structure and its associated labels, we first generated a mask of the histology by taking a slightly dilated version of the labels (morphology.binary_dilation method in scikit-image, typically with a disk shaped structuring element of size 5) to capture nearby unlabeled histological structures, thresholded this histology plane by a value of 10, removed small objects of size less than 200 connected pixels (default connectivity of 1 in morphology.remove_small_objects), and identified bounding boxes of connected histology components (measure.regionprops) and their matching labels. These bounding boxes contain the discrete structures that we will follow laterally, using the corresponding labels for each structure as templates to extend into the rest of each structure.

After identifying each structure and its labels, we fit the labels to the next lateral plane and recursively generated a new template for the subsequent plane. To fit the labels to the next plane, we found the histology bounding box for each structure and resized its labels to this box. We assumed for simplification that each structure is the same or smaller size in each subsequent plane, such as the tapering profile of the cortex laterally, and took only the single largest object found within the structure’s bounding box.

Spline interpolation, anti-aliasing, and range rescaling were turned off during the resize operation to avoid introducing new label values. Some atlas labels contain empty space such as ventricles. To ensure that the ventricles close as they progress laterally, we employed an in-painting approach. Using the Euclidean distance transform method from the Scipy (Virtanen et al., 2020) library (ndimage.distance_transform_edt), we identified the indices of the nearest neighboring pixel for any unlabeled pixel whose corresponding histology intensity value was above threshold and filled this missing pixel label with the value of this neighbor.

Edge map generation

While the labels from one plane could serve as an approximate template for the next plane, we curated this template to fit the underlying anatomy. To map this anatomy, we generated gross anatomical edge maps of the histology images through 3D edge detection. First, we smoothed the volumetric microscopy image using a Gaussian filter with a sigma of 5 followed by an edge detection with a Laplacian filter using a default operator size of 3, using both filters implemented in scikit-image (filters.gaussian and filters.laplace, respectively). This relatively large Gaussian sigma allowed for capture of broad anatomical edges such as the cortex and basal ganglia while minimizing detection of artifactual boundaries. To enhance edge detection of the outermost boundaries, we identified and removed background by thresholding the original microscopy image with an Otsu threshold (filters.threshold_otsu) and combining it with a mask of all the original labels to fill in any missing holes in the thresholded image.

Finally, we reduced the edge-detected image to a binary image of edges alone by applying a zero-crossing detector. This detector separately erodes and dilates (morphology.erosion and morphology.dilation, respectively, with a ball-shaped structuring element of size 1) the edge-detected image to find borders, taking all pixels where this image changed signs as gross anatomical edges (Figure 4A).

Anatomically guided serial 2D reannotation

This edge map allowed us to conform labels to local anatomy. To refit labels, we eroded and regrew them in a watershed transformation guided by the anatomical edge map.

First we eroded labels individually (morphology.binary_erosion in scikit-image with a ball-shaped structuring element of manually determined radii for each atlas). These eroded labels served as seeds for a compact watershed (morphology.watershed implemented in scikit-image with a compactness parameter of 0.005) (Neubert and Protzel, 2014; Kornilov and Safonov, 2018), guided by the anatomical edge map. We used the Euclidean distance transform of the anatomical edge map to define the catchment basins for the watershed transformation, where voxels farther from anatomical edges are at the bottoms of basins and fill faster, guiding the regrowth of eroded labels. Labels typically crossed several anatomical edges but tended to meet neighboring labels at common edges (Figure 4B). To limit the watershed to atlas foreground, which prevents label spillover across empty spaces, we set the watershed mask parameter to the original total labels foreground smoothed by an opening filter (morphology.binary_opening with a ball structuring element of radius 0 [off] to two determined for each atlas to avoid label loss and minimize volume changes) to remove artifacts around small ventricular spaces that might otherwise be crossed. The eroded labels thus regrew to fit anatomical guides and became the new, anatomically refined template to extend labels into the next plane.

To model the tapering and disappearance of labels laterally, we allowed labels to erode completely. We weighted erosion toward central labels within each structure by multiplying the erosion filter size for each label by the label’s median distance to the structure perimeter divided by the maximum distance. Instead of simply eroding away the smallest labels, this approach preferentially eroded labels farthest from the perimeter, typically central labels.

In atlases such as the ADMBA E18.5 atlas, a few spurious labeled pixels from other regions regrew to create artifacts. To filter these spurious pixels, we applied smoothing to the initial labels template using the smoothing approach described below. Also, we applied skeletonization as outlined below to avoid loss of thin structures during erosion.

Each plane of labels thus conformed to its underlying anatomy and became the template for the next plane, keeping labels inherently connected from one plane to the next. While this approach is in serial 2D rather than fully 3D because the starting labeled plane is 2D, a subsequent step will further refine labels in 3D.

Atlas 3D rotation and mirroring

To fill the missing hemisphere in the labels image, we initially simply mirrored the present labels and underlying histology planes across the first unlabeled sagittal plane on the opposite side. We noticed, however, that this mirroring frequently duplicated midline structures because the labels extend slightly past the true sagittal midline. When we shifted the mirroring to the sagittal plane closest to midline, we found that many labels were lost in the final image.

Many of the developing atlases contain at least a few labels positioned solely across the midline in the otherwise unlabeled hemisphere. While it is possible that these labels represent structures unique to one hemisphere, at least some of these labels are on the other side of the midline in other atlases of the ADMBA and thus more likely represent artifact. In some cases (e.g. P4 and P14), mirroring just slightly past the sagittal midline preserved these labels, whereas other atlases (e.g. P28) contained over 100 labels past the midline. One approach to preserve these labels would be to compress the near-midline labels to bring all labels into a single hemisphere at the expense of potentially misaligning otherwise appropriately positioned labels. To avoid this side effect, we elected to cut off a few labels to preserve the placement of the majority of labels.

Upon closer inspection, many of the brains are slightly rotated in two or three dimensions, which contributed to loss of midline labels during mirroring. To rotate images volumetrically in 3D, we applied the scikit-image rotate function to all 2D planes along any given axis for each rotation. For each atlas, we applied this volumetric rotation along all necessary axes until the sagittal midline was parallel to an image edge, manually inspecting each brain in our 2D/3D orthogonal viewer to ensure symmetry. In some cases, such as the E18.5 atlas, the brain skewed laterally along its rostrocaudal axis. To avoid introducing a gap between midline labels and the corrected midline after rotation, we filled all planes on the unlabeled hemisphere side with the last labeled plane before rotation. Mirroring would then overwrite all of these repeated labels except those that filled in potential gaps left by rotation. After rotation, we noticed that mirroring reduced label loss in at least some atlases (e.g. E13.5).

Piecewise 3D affine transformation

The distal spinal cord of the E11.5 atlas is strongly skewed and would be duplicated during mirroring. The skew is complicated by the cord’s coiling back on itself and progressive skew along its length. To bend the distal cord to the sagittal midline, we developed a method for piecewise 3D affine transformation. This transformation allows a cuboid ROI within a volumetric image to be sheared while maintaining a specified attachment point along another axis, reducing discontinuity with neighboring areas.

First, we specify an axis along which to shear and the degree of shearing for a given ROI. Each full plane along this axis is shifted in the indicated direction by a progressively larger amount, overwriting pixels into which the plane is shifted and filling vacated pixels with background to shear the stack smoothly in 3D. If an axis of attachment is also specified, each plane is sheared line-by-line along this axis so that the resulting parallelogram remains fully connected at one end of each axis. The resulting ROI is thus sheared in 3D while remaining smoothly connected to its surrounding space along two orthogonal faces of the original cuboid ROI to minimize disruption.

Applied to the skewed spinal cord, this approach allowed us to shear the cord one section at a time as it curved back along itself (Figure 5—figure supplement 1, left column), with the connected faces ensuring that each piece remains attached to one another. First we sheared the entire distal cord from the start of the skew (Figure 5—figure supplement 1, middle left column). This shift brought the proximal section toward midline, but the more distal cord remained skewed where it curved back on itself. To correct this skew, we sheared again but starting from a more distal point and along an orthogonal angle to follow the cord’s curve (Figure 5—figure supplement 1, middle right column). Now most of the cord lined up with the sagittal midline, but the shear exacerbated the skew of the most distal cord. Finally, we applied a third affine, this time on only the most distal section and in the opposite direction as the prior affine (Figure 5—figure supplement 1, right column).

After mirroring the resulting brain and cord, no duplication could be seen. This piecewise, overlapping affine of targeted regions allowed straightening of the cord without breakages or alteration of surrounding areas. Although surrounding non-CNS tissue suffered noticeable breakage, they were stripped out as described below.

Stripping non-CNS signal

The extended labels allowed us to mask and crop out non-CNS tissue, including the rest of the embryo present in several of the embryonic stage atlases. While distinguishing this tissue based on intensity characteristics alone would be challenging, especially for areas where the spinal cord extends the length of the embryo, the extended, mirrored labels provide a map of relevant CNS tissue.

For each of the atlases with non-CNS tissue (E11.5-E15.5), we first cropped the atlas to the bounding box of these labels along with a small padding (5px) to remove much of the non-CNS tissue, such as the entire unlabeled body in the E15.5 atlas. We removed non-CNS tissue remaining within the cropped areas by using the labels as a mask to remove all histology pixels outside of the labels mask, including the body surrounding the labeled spinal cord in E11.5. To avoid missing tissue that may have been unlabeled, we dilated the labels mask slightly (morphology.binary_dilation with a ball-shaped structuring element of size 2) so that the mask encompasses both the labels and its immediate surroundings. The resulting histology thus contains all pixels in the near vicinity of labels, including pixels that should be labeled but are not.

Label smoothing

While labels appear generally smooth when viewed from the sagittal plane, label edges are noticeably jagged when seen from the orthogonal directions. A previous smoothing solution proposed by Niedworok et al., 2016 utilized a Gaussian blur with a sigma of 0.5 in two iterations to minimize ragged edges in an atlas derived from the Allen Reference Atlas P56 mouse brain atlas. We applied this approach by iteratively applying the Gaussian filter implemented in scikit-image (filters.gaussian) with a range of sigmas (0.25–1.25) to each label in 3D (Figure 3—figure supplement 2).

To extract each label in 3D, we found the bounding box of the label using the scikit-image measure.regionprops method and added additional empty padding space around the box. The Gaussian filter was applied to the label, and the original label in the bounding box was replaced with the smoothed label. Each label smoothing left small gaps vacated by the previously ragged borders. To fill in these gaps, we employed the in-painting approach described above. Finally, we replaced the original bounding box with that of the smoothed label in the volumetric labels image. We repeated the process for all labels from largest to smallest to complete the smoothing.

To enhance smoothing while retaining the original underlying contour of each label, we devised an adaptive opening morphological filter approach. The morphological opening filter first erodes the label to remove artifacts such as ragged edges, followed by dilation of the smoothed label to bring it back toward its original size. In place of the Gaussian filter, we applied this opening filter (morphology.binary_opening in scikit-image). For smaller labels, the filter occasionally caused the label the disappear, particularly for sparsely populated labels with disconnected pixels. To avoid label loss, we employed an adaptive filter approach by halving the size of the filter’s structuring element for small labels (5000 pixels). For any label lost in spite of this filter size reduction, we switched the filter to a closing morphological filter (morphology.binary_closing), which dilates first before eroding and tends to preserve these sparse labels at the expense of potentially amplifying artifact.

Smoothing quality metric

To evaluate the quality of smoothing and to optimize morphological filter structuring element sizes, we developed a smoothing quality metric with the goal of balancing smoothness while maintaining the overall original shape and placement. Since the major effect of label smoothing is to minimize high frequency aberrations at label borders and thus make each label more compact, we measured this amount of smoothing by the 3D compactness metric. We termed the difference in compactness before and after smoothing as ‘compaction.’ As the goal of smoothing is to remove these artifacts while preserving the overall shape of the label, we introduced a penalty term of ‘displacement,’ measured by the volume shifted outside of the label’s original bounds.

We used the classical measure of volumetric compactness, where lower values are more compact (Ballard and Brown, 1982):

Compactness=SA3/Vol2 (1)

To measure the surface area of each label in 3D, we employed the marching cubes algorithm (Lorensen and Cline, 1987) as implemented in scikit-image (measure.marching_cubes_lewiner), which also accounts for anisotropy. The algorithm extracts surfaces in 3D by dividing the volume into cubes and marching through these cubes to find where isosurfaces intersect with each cube, forming triangular surfaces within each cube wherever an isosurface passes through the cube. Using a mask of each label, we obtained the edge surface as a mesh from which we measured the surface area (measure.mesh_surface_area). We took the volume as the total number of mask pixels and multiplied by the product of the voxel spacing to account for anisotropy. With the surface area and volume, we calculated the labels compactness (1). To quantify the fractional change in compactness before and after smoothing, we took the original compactness minus the smoothed compactness and normalized the difference to the original compactness to give the unitless value that we termed ‘compaction’:

Compaction=CorigCsmoothCorig (2)

where C is the 3D compactness given above.

To measure ‘displacement,’ we measured the volume shifted outside of the label’s original bounds. Taking the mask of the smoothed label, we combined it with the inverse of the mask of the original label through a boolean ‘and’ before totaling the number of pixels. Similarly to the compactness measure, we normalized the displacement volume to generate a unitless value constrained between 0 and 1, dividing this displaced smoothed volume by the total smoothed volume:

Displacement=VolsmoothVolorigVolsmooth (3)

where Vol is the volume of the given label.

As a measure of smoothness quality, we took the difference of the compaction and displacement:

Smoothingquality=CompactionDisplacement (4)

To quantify the smoothing quality for the entire atlas, we took the weighted arithmetic mean of all the labels’ smoothing qualities, weighting by each label’s volume. The atlas-wide smoothing quality metric can be summarized in the following equation:

AtlasWideSmoothingQuality=i=1NSmoothingQualityiVoli,origi=1NVoli,orig (5)

where N is the total number of original labels. While maximizing compaction would reduce surface area the most by transforming the label into a perfect sphere, the displacement from the label’s original space would penalize this over-compaction, allowing us to target the balance of compaction and displacement to find the optimal smoothing quality.

Anatomical to label edge distance quantification

As an automated method of quantifying the correspondence between labels and anatomical edges, we used the anatomical edge maps generated earlier to measure the distance between anatomical and label edges. We reduced each label to its edges in 3D by eroding the very outer surface of each label (morphology.binary_erosion with a cross-shaped structuring element with connectivity of one) and subtracting this eroded label from the original label.

To measure distances between label and anatomical borders, we performed the Euclidean distance transform (ndimage.distance_transform_edt in Scipy) on the anatomical edge map to measure distances from any given voxel to the anatomical edges, using the sampling parameter to specify the microscopy image spacing in µm. Using the labels edge map, we next took only the voxels in the distance map corresponding to these label borders as a map of distances from each label voxel to its nearest anatomical edge. As overall measures of distance from label borders to the nearest anatomical border, we summed the edge distances for each label to compare before and after label refinement.

Anatomically guided 3D reannotation

Generating edge images provided a map of the boundaries between anatomically distinct regions not only to measure distances between borders, but also to curate the labels themselves with these anatomical edges. To reannotate the existing unsmoothed labels, we eroded and regrew them through an anatomically guided watershed transformation similar to the approach described above but now in 3D. We tested seeds with multiple erosion filter sizes and found that a structuring element size of 8 reduced the seed size sufficiently to correct label bounds around several grossly abnormal labels, including the lateral septal nuclei and basal ganglia.

As a global operator, erosion typically leads to loss of thin structures, especially with larger structuring elements. To avoid this loss, we first extracted the core structure of each label by finding its 3D skeleton (morphology.skeletonize_3d in scikit-image). We added the skeletonized image back to the eroded label to recover the location of thin structures, allowing the watershed to regrow these labeled areas in addition to the eroded label. To limit branches in the skeleton, which could counter the effect of the erosion, we input a lightly eroded version of the labels (structuring element half the size of that for the main erosion) to the skeletonization.

The resulting watershed segmentation also served as preliminary smoothing but introduced its own label edge artifacts, though of lower frequency than in the original labels. We thus deferred the smoothing algorithm until after this watershed step and could use smaller filter sizes to generate a final smooth image.

Application to the full ADMBA series

Each atlas in the ADMBA required a different set of refinement features and settings, including specialized adjustments such as the 3D piecewise affine only for a specific atlas (E11.5). To allow for customized settings, we defined separate profiles of parameters within our software suite for each atlas.

E11.5

The microscopy images depict a specimen in embryonic stage with the caudal end including the spinal cord wrapping around itself and deviated laterally from the rest of the body. As with most other atlases in this series, one half of the labels were missing. Making the image symmetric would allow us to mirror labels from the existing side onto this missing side and minimize bias when using the atlas for automated registration tasks. We started by rotating the image by 5° in the axial planes toward the left of the brain and 1° in the coronal planes to raise the right side of the brain, bringing the sagittal midline parallel to an image edge. To shift the distal end of the spinal cord back toward the sagittal midline, we applied the piecewise 3D affine transformation described above (Figure 5—figure supplement 1). Prior to the affine, we applied an additional rotation of 30° in the sagittal planes to position the skewed distal cord within a single cuboid ROI parallel to the image.

With the embryo now symmetric, we mirrored the labels and microscopy signal along the embyro’s midline, measured as 52% across the sample along the z-axis. To highlight the central nervous system and remove breakages introduced by the affine transformations in non-CNS tissue, we stripped out non-CNS tissue as described earlier.

The E11.5 atlas uniquely contains labeled ventricles. To ensure that the Laplacian of Gaussian edge-detection algorithm appropriately found ventricular edges, we used only the thresholded atlas rather than incorporating the labels to find the background. We also included the ventricular space as foreground for purposes of the DSC calculations between microscopy and labels to account for the ventricular labeling. This atlas did not require lateral extension.

E13.5

This atlas also contains the full embryo, but the spinal cord is symmetric along the sagittal midline of the brain and did not require the affine transformations as in the E11.5 atlas. After extending the lateral edges, we rotated the atlas by 4° in the axial planes toward the left of the brain and 2° in the coronal planes to lift the right side of the brain, making the atlas symmetric before mirroring the atlas at 48% along the sagittal planes. We again stripped non-CNS tissue the same way as for the E11.5 atlas.

E15.5

This atlas is the final one to include the complete embryo, although only the very rostral end of the spinal cord includes labels. We extended the lateral edges, rotated the images by 4° in the axial planes toward the left side of the brain, mirrored the atlas at 49% along the sagittal planes, and stripped away the entire embryo outside of the brain. A stepwise shift in sagittal planes is apparent at several planes (e.g. 104 and 129) in both the histology and label images in the original atlas, which we smoothed slightly in the labels during the smoothing step.

E18.5

A small subset of labels from sagittal planes 103–107 were compressed along the dorsoventral axis. To match them with their neighboring label planes and the underlying atlas, we resized them similarly to the lateral edge extension. Starting with the first plane in this subset, we thresholded the microscopy image, removed small objects, and obtained the largest bounding box of connected histology components. Taking only this largest connected structure allowed us to avoid including extraneous tissue visible on the ventral aspect of the brain, which was unlabeled. We repeated the process on the labels to obtain its compressed bounding box and resized it to the size of the microscopy bounding box. Finally, we repeated the entire process on the rest of the planes in this subset (Figure 2—figure supplement 3).

For the lateral edge extension, we noted that the basal ganglia in the most lateral planes are slightly larger than in more medial planes. Under the assumption that the basal ganglia would be tapering laterally, we skipped these planes and started the extension at 13.7% along the sagittal planes to take the plane with smallest basal ganglia label. After rotating the atlas by 1.5° in the axial planes toward the right of the brain and 2° in the coronal planes to lift the left side of the brain, we mirrored microscopy and labels at 52.5% along the sagittal planes for symmetry.

P4

This atlas is reminiscent of E18.5 but without the necessity of setting the lateral edge extension starting plane explicitly or expanding any compressed labels. After lateral edge extension and rotating the atlas by 0.22° in the axial planes toward the right of the brain, we mirrored microscopy and labels at 48.7% along the sagittal planes for symmetry.

P14

The most laterally labeled plane is discontinuous with the rest of the labeled planes in this atlas, so we skipped this plane during lateral extension and started extending only from the first contiguous set of planes. After rotating the atlas by 0.4° in the axial planes toward the left of the brain, we mirrored the microscopy and labels images at the 50% mark along the sagittal planes.

P28

The most lateral planes had incomplete labels, requiring use of a more medial plane with complete labels at 11% along the sagittal planes for the lateral edge extension. After rotating the atlas by 1° in the axial planes toward the right of the brain, we mirrored the microscopy and labels images at the 48% mark along the sagittal planes.

P56

The ADMBA contains a P56 mouse similar to the adult P56 but following the same ontological labeling scheme as in the rest of the ADMBA. This atlas uniquely contains bilateral labels, although the far lateral section is still missing. We again extended the lateral edges and mirrored the microscopy and labels images, starting extension at 13.8% and mirroring at 50% along the sagittal planes, respectively. The most lateral labeled plane contains two distinct labeled structures, the cortex and cerebellum, requiring separate extension for each distinct structure as outlined in our method above.

Animals and tissue clearing

All procedures and animal care were approved and performed in accordance with institutional guidelines from the University of California, San Francisco Laboratory Animal Research Center (LARC). All strains were maintained on a C57BL/6J background. Animals were housed as one breeding pair per cage in a vivarium with a 12 hr light, 12 hr dark cycle. For timed pregnancies, noon on the day of the vaginal plug was counted as embryonic day 0.5. Pups were harvested at P0 (postnatal day 0). For assessing the eight atlases in the ADMBA, sample size was determined by the number of atlases available in the original 2D resource (n = 8). The size of our validation cohort (P0 wild-type mouse brains, n = 15) was chosen to detect an effect size of 1.5 after correction for multiple comparisons using a power calculation based on a paired t-test.

At the time of experiment, P0 pups were anesthetized on ice and perfused transcardially with ice-cold 1X PBS supplemented with 10 U/mL heparin and then with 4% PFA in 1X PBS, followed by brain isolation. P0 brains were post-fixed overnight at 4°C in 4% PFA in 1X PBS. The next day the excess fixative was removed by washing the brains with 1X PBS supplemented with 0.01% (wt/vol) sodium azide (Sigma-Aldrich) for at least 2 hr at room temperature (RT).

Samples were cleared using the advanced CUBIC clearing protocol for whole-brain and whole-body clearing (Susaki et al., 2015). In short, samples were immersed in 1/2-water-diluted Reagent-1 containing 1 µM SYTO 16 (Thermo Fisher) and incubated at 37°C for 6 hr (Reagent-1: 25 weight% (w%) Urea, 25 wt% Quadrol, 15 wt% Triton X-100, and dH2O). 1/2 diluted Reagent-1 was replaced with Reagent-1 containing 1 µM SYTO 16 and incubated overnight at 37°C on a rotator. The next day, solution was replaced with fresh Reagent-1 containing 1 µM SYTO 16 and was incubated overnight at 37°C. Reagent-1 was replaced with fresh Reagent-1 containing 1 µM SYTO 16 every 48 hr for a total of 8 days. Tissue clearing was stopped by washing the sample with 1X PBS supplemented with 0.01% (wt/vol) sodium azide at RT once for 2 hr, once overnight and again once for 2 hr. After the wash step, samples were immersed in 10 mL of 1/2-PBS-diluted Reagent-2 (Reagent-2: 25 wt% Urea, 50 wt% sucrose, 10 wt% Triethanolamine, and dH2O). Vials containing brains in 1/2-PBS-diluted Reagent-2 were placed in a vacuum desiccator with gentle shaking overnight at RT. The following day, the solution was replaced with Reagent-2 and incubated at 37°C overnight on a rotator. The next day, Reagent-2 solution was replaced with fresh Reagent-2 and incubated at 37°C overnight on a rotator. This step was repeated four times. We imaged and processed a total of n = 15 from 5 female and 10 male mice.

Lightsheet imaging

Samples were imaged within 7 days of completing the clearing protocol at the Gladstone Institutes Histology and Light Microscopy Core on a Zeiss Lightsheet Z.1 microscope. Lightsheet imaging of a mouse P0 brain required approximately 50 tiles (5 × 10) per brain at 5x (Zeiss EC Plan-Neofluar 5x, NA 0.16, RI 1.45, WD 5.6 mm), which afforded a level of resolution that allowed for nuclei detection (0.913 µm/px lateral resolution).

The microscope requires specimen to be suspended between fixed illuminator and detector objectives, typically using capillaries to embed small specimens. To suspend the relatively larger whole brain, we designed a custom 3D-printed rod with a platform to maximize surface area for gluing the ventral surface of the brain to the mount (Figure 7—figure supplement 1). The rod also contains a neck to position the brain directly under the mount holder, allowing full movement of the brain throughout the chamber to capture the brain it its entirety. The mount oriented the brain axially toward the detector to minimize the path of the illuminators on opposite sides laterally through the tissue as well as the emission path superiorly to the detector. We designed the mount in Onshape and printed it using a Stratasys uPrint 3D printer.

To glue the brain to the mount, we placed the brain, ventral surface facing up, on a custom sieve to drain reagent and rolled cotton bud applicators on the ventral surface of the brain to dry it. We applied glue (Scotch Super Glue Liquid) with a brush applicator to the mount platform surface and attached it the brain. After flipping the brain to dorsal side up, we placed the mount in a custom holder to allow the glue to dry over 3 min and dribbled Reagent-2 media on the dorsal surface to ensure that it did not dry out. After the glue dried, we suspended the mount from the microscope manipulator before immersing the mounted brain into the microscope chamber filled with Reagent-2.

With the lasers initiated, we aligned the two opposite illuminators along the z-axis by visual inspection at a central region of the brain. We ranged the z-stack from just beyond the first and last z-planes with visible nuclei, typically 800–1000 planes using a slice interval of 4.935 µm with a lightsheet thickness of 10.44 µm, and set the image tiling to include the farthest lateral and anterior-posterior nuclei with a 10% overlap per tile. We illuminated the brain with 488 nm excitation and 30 ms dwell time through the Z.1 LSFM 5x/0.1 illuminators, LBF 405/488/561/640 laser blocking filter, SBS LP 560 secondary beam splitter, and BP 505–545 band pass filter. The microscope paused for 20 s between each tile to allow tissue settling after repositioning for the next tile. We controlled the microscope through the Zeiss Zen microscopy software suite and saved images in a CZI multi-tile format with all tiles stored in a single archive.

Image stitching

To stitch the tiled microscopy images in an automated fashion, we used the Fiji/ImageJ (Schindelin et al., 2012; Rueden et al., 2017) BigStitcher plugin (Hörl et al., 2019), a successor to the Stitching plugin (Preibisch et al., 2009) that allows for better memory management and multiprocessing as well as a graphical interface to verify alignments. As we needed to stitch multiple brains, we accessed this plugin headlessly through its scriptable interface, manually intervening only to visually verify alignments before proceeding with the fusion step.

After importing the CZI file into the BigStitcher HDF5-based format, the plugin auto-detected the brightest illuminator for each tile, discarding planes from the other illuminator. We chose to simply select planes from the optimal illuminator rather than fusing planes from both illuminators after our inspection revealed that illuminators rarely if ever aligned perfectly throughout the tile, leading to artifacts such as apparent elongation of nuclei from imperfect overlap of the same nuclei from different illuminators.

The plugin calculated tile shifts using a phase correlation method, and we filtered out links below a correlation threshold of r = 0.8 before applying shifts through the two-round iterative global optimization strategy. To account for occasional tile misalignments, we manually inspected every pre-stitched brain to reposition any misaligned tiles, which occurred in approximately 10% of brains. Once tiles aligned, the plugin fused them into a single large TIFF image per channel. We imported this fused file via Python-Bioformats and Javabridge, libraries that allow access to life science formats via Bio-Formats, to a Numpy array format for image processing in our Python-based software as outlined below.

Automated nuclei detection

We detected nuclei in cleared mouse brains using a 3D Laplacian of Gaussian blob detection technique throughout each whole brain. To perform detections in large images several hundred gigabytes (GB) to over a terabyte (TB) in size, we subdivided the image into many smaller chunks to reduce RAM requirements and maximize parallel processing. We loaded images through the Numpy library’s memory mapped method (load with the mmap_mode option) to load only the necessary parts of the image on-the-fly, allowing us to load small images a chunk at a time without reading the entire volumetric image into memory. We divided the image shape into overlapping chunks to ensure that nuclei at borders would not be missed, with overlap size of approximately the nucleus diameter. After determining the offset and shape of each chunk, we set the image array as a class attribute, initiated multiprocessing (the multiprocessing.Pool in the standard Python library), and accessed each chunk as a view in a separate process via class methods to avoid duplicating arrays in memory. Thus, we could control total memory usage by the size of chunks and the number of CPU (central processing unit) cores available for separate processes.

3D cell detection poses a number of challenges including adapting to local variation such as staining inhomogeneity, background variation, and autofluorescence, in addition to overlapping cells in dense tissue (Shuvaev et al., 2017). To address these issues, we analyzed images in a local manner by further subdividing each chunk for preprocessing based on its immediate surroundings. We split each chunk into sub-chunks using the same approach as above but ran each sub-chunk serially within each CPU process. In each sub-chunk, we first clipped intensity values at the 5th and 98.5th percentiles (percentile in Numpy) to remove extreme outliers, rescaled the intensities from 0 to 1, and further saturated signal by clipping the rescaled intensities at the 50th percentile. We next enhanced edges by using unsharp masking with a Gaussian sigma of 8 (filters.gaussian in scikit-image) to identify sharp details as the difference between an image and its blurred version (which we amplified by a factor of 0.3) and adding back those details to the original image. We mildly eroded the resulting signal with an erosion filter using an octahedron structuring element of size 1 (morphology.erosion with morphology.octahedron in scikit-image) to separate out blobs.

To detect blobs, we implemented the 3D Laplacian of Gaussian blob detector from the scikit-image library (feature.blob_log) as a multi-scale interest point operator (Lindeberg, 1993). We set the minimum and maximum sigma based on the microscopy resolution, with 10 intermediate values, detection threshold of 0.1, and overlap fraction threshold of 0.55 below which duplicated blobs are eliminated. Initially we missed many nuclei positioned above one another in the z-direction, likely because the anisotropy necessitated by the relatively thick lightsheet at 5x in our setup limited resolution in the z-direction. To improve detection along the z-axis, we interpolated the images in this direction to near isotropy before detection (transform.resize in scikit-image). The blob detector had a tendency to cluster detections in the bottom and topmost z-planes in each ROI from nuclei visible within the ROI but whose centroids are outside. To avoid this clustering and minimize duplication with adjacent ROIs, we cropped nuclei from these planes on the assumption that they would be captured better in the adjacent, overlapping ROIs.

Overlapping chunks minimized missing nuclei at edges but also necessitated pruning blobs duplicately detected in adjacent chunks. Pruning involves checking for duplicates within all potentially overlapping regions. Since the overlapping portions of the regularly spaced chunks collectively form a grid pattern throughout the full volumetric image, we could limit our search to these grid planes along each axis. After completing detection on the whole image, we first pooled all detected blobs into a single array. Along a given axis of the full image, we determined the boundaries for each overlapping region and all of its blobs. Within each overlapping region, we found all blobs close to another blob by taking the absolute value of the difference between all blobs with one another and finding blobs within a given tolerance in all dimensions. The tolerance was titrated so that the ratio of final blobs in overlapping regions to the next adjacent regions of same volume was about 1:1. For each close pair of blobs found, we replaced both blobs with a new blob that took the mean of their coordinates. To minimize memory usage, we checked smaller groups of blobs against one another until completing all comparisons. We checked overlapping regions simultaneously in multiprocessing along a given axis for efficiency, re-pooled all blobs, and pruned along the next axis to account for blobs that may have been duplicately detected in overlapping chunks along multiple axes, at grid intersections.

Parameters for preprocessing, detections, and pruning steps were optimized through a Grid Search approach, a type of hyperparameter tuning, to check combinations of parameters systematically. To evaluate the accuracy of each set of parameters, two students in our lab generated truth sets of nuclei locations and radii using our serial 2D nuclei annotation tool, taking ROIs of size 42×42×32 pixels (x, y, z) from representative ROIs of all major brain structures (n = 15 forebrain, eight midbrain, and seven hindbrain; n = 2766 nuclei). They separately generated additional truth sets at a slightly lower magnification (4x), size 60×60×14 pixels (n = 40 ROIs, 1116 nuclei) to increase representation. After detecting nuclei on these images with a given set of parameters, matches between detections and ground truth was determined using the Hungarian algorithm, a combinatorial optimization method to determine optimal assignments between two sets (Kuhn, 1955), as implemented in optimize.linear_sum_assignment in Scipy. After scaling nuclei coordinates for isotropy, we found the Euclidean distances between detected and truth nuclei points through distance.cdist, which serves as the cost matrix input to optimize.linear_sum_assignment to find optimal pairings between points based on closest distance. We took correctly identified detections, or true positives (TP), as pairings within a given tolerance distance. Unpaired detections or those in pairs exceeding this threshold were considered false positives (FP), and the same for ground truth were false negatives (FN). Since a match for a given nucleus within the ROI may lie outside of it and thus go unseen, we first searched for pairings only within an inner sub-ROI, followed by a secondary search for pairings between only unmatched inner sub-ROI objects and the rest of the ROI (Ho et al., 2017). This approach reduced the total number of nuclei available (n = 1118 nuclei) but avoided missed border matches. As measures of performance of our detection compared with ground truth, we used the following standard equations:

Sensitivity(Recall)=TPTP+FN (6)
PositivePredictiveValue(PPV,orPrecision)=TPTP+FP (7)

Image downsampling and registration

To assign nuclei to the proper brain label, we employed automated label propagation by registering the E18.5 atlas to each of our imaged mouse brains using SimpleElastix (Marstal et al., 2016), a toolkit that combines the programmatic access of SimpleITK (Lowekamp et al., 2013) to the Insight Segmentation and Registration Toolkit (ITK) (Yoo et al., 2002) with the Elastix (Klein et al., 2010; Shamonin et al., 2014) image registration framework. Elastix has been recently validated as a computationally efficient and accurate tool for registration of mouse brains cleared by CUBIC (Nazib et al., 2018).

As stitched images are typically several hundreds of GBs per file, image downsampling was necessary to reduce memory utilization during image registration. To reduce file size efficiently in both time and memory usage, we employed the same chunking strategy as used during nuclei detection except with larger, non-overlapping units to resize multiple sections of the image simultaneously. We also reduced memory required for the output array by saving directly to disk with a memory-mapped array (lib.format.open_memmap in Numpy). We matched the target final size to the E18.5 atlas, which would be registered to each downsampled image.

Registration involved rigid followed by non-rigid alignment. For rigid registration, we employed a translation (translation parameter map in SimpleElastix) with default settings except increasing to 2048 iterations (MaximumNumberOfIterations setting) followed by an affine (affine parameter map) with 1024 iterations, applied with an ElastixImageFilter, to shift, resize, and shear the atlas microscopy image to the same space as that of the sample brain. For non-rigid alignment, we employed a b-spline strategy (bspline parameter map) guided by the AdvancedNormalizedCorrelation similarity metric (Nazib et al., 2018; Hammelrath et al., 2016) with grid spacing of size 60, measured in voxels rather than physical units (FinalGridSpacingInVoxels setting in place of FinalGridSpacingInPhysicalUnits), over 512 iterations. The TransformixImageFilter in SimpleElastix allowed us to apply the identical registration transformation to the atlas labels image, except that we set the final b-spline interpolation order (FinalBSplineInterpolationOrder) to 0 to avoid interpolating any new values, preserving the labels’ specific set of integer values. We applied this identical transformation to both the mirrored and edge-refined atlas labels.

To evaluate the level of alignment from registration, we measured the similarity between each registered atlas histology and its corresponding sample image using a Dice Similarity Coefficient (DSC) (Dice, 1945) as implemented by the GetDiceCoefficient function in SimpleElastix/SimpleITK, given by the equation (Tustison and Gee, 2009):

DiceSimilarityCoefficient(DSC)=2|ST||S|+|T| (8)

where S and T are two different sets of voxels. We took the foreground of each atlas and sample microscopy image to be its mean threshold (filters.threshold_mean function in scikit-image) and input them to a LabelOverlapMeasuresImageFilter to take the DSC.

Whole-brain nuclei measurements by label

Registration of the atlas to our volumetric nuclear-stained brain microscopy images allowed us to quantify nuclei per label for comparison with the original (mirrored) and smoothed (edge-aware) atlases. We first measured volumes per label by taking a mask of each registered label and summing the foreground pixels within each mask before multiplying this volume by the microscopy pixel resolution (scaled for downsampling) to obtain volumes in physical units (mm3).

For nuclei densities, we first constructed a nuclei heat map by converting the nuclei coordinates to nuclei per voxel within the downsampled image. We scaled the coordinates to the scaling of the downsampled image, rounding to the nearest integer, and found the counts of nuclei at each coordinate (unique with return_counts option in Numpy). We next indexed these coordinates directly into an empty Numpy array of the same shape as that of the downsampled image to assign them to the corresponding nuclei counts at each voxel. We used the same label mask previously obtained to find the number of nuclei within each given label. Dividing the number of nuclei by the volume within each label gave the label nuclei density.

To measure the variability of nuclei within each label before and after label reannotation, we measured the coefficient of variation within each label given by the standard equation:

Coefficientofvariation(CV)=σμ (9)

where σ is the standard deviation, and µ is the mean. A lower coefficient of variation indicates lower variability and thus tighter capture of a more homogeneous label. As a raw proxy for nuclei variation, we first measured the variation of intensities within the nuclear-stained lightsheet images. Using the same label masks, we took the standard deviation of voxel intensities and divided it by the mean of intensities within the label (std and mean, respectively, in Numpy) to obtain the intensity coefficient of variation. Similarly, we measured the nuclei coefficient of variation by measuring the standard deviation and mean values of nuclei counts per label within the nuclei heat map. We also stratified nuclei ROIs into low/medium and high-density regions by taking the Otsu threshold of their density, which corresponded with their density histogram, and separately measured recall and precision in each group (Figure 6—figure supplement 2D).

Nuclei clustering

As another measure of label alignment at the nuclei level, we measured nuclei clustering using Density-Based Spatial Clustering of Applications with Noise (DBSCAN) (Ester et al., 1996) as implemented by the scikit-learn library (Pedregosa et al., 2018). DBSCAN clusters tightly packed points within a neighbor distance defined by the parameter E and a minimum number of points given as another parameter, with isolated points in lower density regions that cannot be clustered considered ‘outliers’ or ‘noise.’ The minimum number of samples is typically taken as 2 · ndim, where ndim is the number of dimensions (Schubert et al., 2017), thus giving 6 for our 3D nuclei point cloud. To find E, the nearest-neighbor distance of the 2 · ndim − 1 neighbor for each point is sorted and plotted to find the distance at the ‘elbow’ point, the point of maximum curvature (Schubert et al., 2017), which we found to be at least 20 µm (Figure 7—figure supplement 3A).

For each label in the original (mirrored) atlas registered to each wild-type brain, we extracted the nuclei coordinates within the label and clustered them by DBSCAN to find the number of clusters, nuclei per cluster, and ‘noise,’ or number of isolated nuclei that remained unclustered. We repeated the same process using the same nuclei coordinates for each brain but with the smoothed (edge-aware) atlas registered identically to the brain. As the ‘elbow’ distance of maximum curvature can be difficult to define, and E can strongly influence the clustering, we repeated this process for a range of E values through the elbow region (Figure 7—figure supplement 3B) and highlighted a conservative distance of 20 µm (Figure 7—figure supplement 3C–E).

Statistics

Statistical tests used for comparisons are described in the relevant Results sections and figure legends. Bonferroni correction for p-values is applied when multiple statistical tests are performed for each figure sub-panel as indicated for the number of tests.

MagellanMapper software suite

We provide the MagellanMapper image software suite as a tool to assist with visualization, annotation, and automated processing of volumetric images (Figure 7—figure supplement 4). The suite consists of a graphical user interface (GUI) to aid visualization of 2D images in a 3D context and command-line interface (CLI) for non-interactive processing in workstation and cloud environments.

Graphical interface

The main GUI integrates ROI selection with 3D point and surface rendering through the Mayavi toolkit (Ramachandran and Varoquaux, 2011). Users can load volumetric image files and specify ROI boundaries through sliders and text boxes or load a previously saved ROI. 3D point rendering provides a voxel-based visualization of the ROI with minimal filtering, whereas the 3D surface rendering utilizes VTK (The Visualization Toolkit) (Schroeder et al., 1998) for cleaner images. The interface is implemented in TraitsUI for integration with Mayavi.

To inspect raw images, the user can launch two types of mixed 2D/3D interfaces from the main GUI to display and annotate the original 2D images for each plane. The first 2D interface is a serial 2D ROI viewer that shows each successive 2D plane within the ROI side-by-side, allowing the user to follow objects such as nuclei that come and go from plane to plane. Larger overview images at different magnifications show context and synchronize with the smaller views to effectively zoom in on a given plane. These overview images are also scrollable along the z-axis to visualize subtle object shifts in place. The interface also provides annotation tools geared toward blob detection such as nuclei. Detected blobs appear as circles along with optional segmentations, and the user can drag, resize, or cut/copy/paste circles to improve placement and flag their correctness. We have used this simplified blob annotator to generate truth sets for blob detection verification and optimization.

In addition to ROI viewers, we provide a simultaneous orthogonal viewer to visualize and annotate atlases in all three dimensions. It displays orthogonal planes of the full volumetric image in three separate panels, with crosshairs in each panel denoting the corresponding planes in other panels. Clicking on or scrolling within any panel updates crosshairs and synchronizes the other displayed planes. The user can also load label maps to overlay directly on atlas microscopy images with an adjustable labels opacity to allow close inspection of annotation alignment with anatomical structures. To distinguish an arbitrary number of labels from one another, we use a custom discrete colormap with randomly generated colors, each assigned to a single label. We display all images as views of one another to minimize memory requirement and update all orthogonal planes in real-time.

As a method for simple, rapid editing of these labels, the user can enter an editing mode to simply click and drag on individual labels to paint them into other spaces. We have used the interface with a tablet and electronic pen to edit labels by drawing. As hand-editing any given plane likely introduces edge artifacts seen in other orthogonal directions, we also designed a method to interpolate contours between two distant planes. After the user edits the same label ID at start and ending planes and initiates the interpolation, it takes the signed distance transforms of the label masks (ndimage.distance_transform_edt in Scipy combined with a mask of the original label to identify distances inside versus outside the label) and interpolates those distances for each intervening plane (interpolate.interpn applied across a mesh-grid of the planes). The interpolation provides label border extensions that are generally smooth in all dimensions after manually editing only the first and last plane.

The annotation interfaces are implemented in Matplotlib (Hunter, 2007). Blobs are stored in an SQLite (Hipp et al., 2021) database, while atlas edits are saved directly to their underlying 3D image file.

Headless pipelines

In addition to a GUI for interactive visualization and verification, MagellanMapper provides automated pipelines for non-interactive image processing such as whole brain nuclei detection in cloud-based work environments. Users can access the suite via its CLI, and the suite provides Bash scripts to connect MagellanMapper with other tools such as Fiji/ImageJ for image stitching and Amazon Web Services (AWS) in a platform-independent manner.

For input/output (I/O), the suite utilizes standard 3D image formats for portability with other software libraries. Microscopy images stored in proprietary formats such as Zeiss CZI format can be imported to a standard Numpy array archive using the Bio-Formats library (Linkert et al., 2010) (via the Javabridge and Python- Bioformats libraries developed for CellProfiler [Lamprecht et al., 2007] libraries) along with a separate Numpy archive containing image metadata extracted from the original file. Loading the imported Numpy array as a memory-mapped file (load function with mmap_mode option) allows users to access small parts of large files to minimize load time and memory usage by only loading the requested ROI rather than the potentially TB-sized full volumetric image. In addition to Numpy array archives, other 3D image formats such as MetaImage, NIfTI, and DICOM are supported through the SimpleITK/SimpleElastix library. Annotated images are saved to their respective formats.

As an example of an automated pipeline for image import, a user can launch the script from a cloud-based server instance to first retrieve and decompress an archived microscopy file previously saved in a cloud storage location. After extracting the microscopy image, the pipeline script launches Fiji/ImageJ to run a custom headless script for the BigStitcher plugin, which stitches the tiled microscopy image as described above. After allowing the user to verify and adjust tile placements through the BigStitcher interface, the script continues the BigStitcher tile fusion operation, resulting in TIFF formatted images for each channel. The pipeline script next calls MegellanMapper to import these TIFF images into a single Numpy array archive. Subsequent pipelines can be run to process the imported image for automated nuclei detections, downsample or transpose the image, register images to an atlas, or automatically refine new atlases in 3D.

Software access

We provide MagellanMapper as open-source software in the hope of facilitating both interactive and headless processing of large volume microscopy images and the atlases to which they will be registered.

The suite API includes 3D image processing functions designed to be useful as library methods for other applications. Library functions for the Python and R plots depicted here are also provided to reproduce similar graphs. As an open-source, Python-based tool, our vision for the suite is that it will work alongside, integrate with, and itself become refined by the many other excellent image processing software suites and libraries available to the scientific community.

Software versions

The MagellanMapper suite is written in Python with the following versions and Python libraries: Python 3.6, Numpy 1.15, Scipy 1.1, scikit-image 0.14, scikit-learn 0.21, Matplotlib 3.0, Matplotlib ScaleBar 0.6, Mayavi 4.6, TraitsUI 6.0, PyQt 5.11, Python-Bioformats 1.1, Javabridge 1.0, SimpleElastix 1.1, and Pandas 0.23. We used Conda 4.7 for Python library management. We stitched images with Fiji/ImageJ 1.52 using BigStitcher 0.2.10. We computed additional statistics in R 3.5 with RStudio 1.1. To interact with AWS, we used AWS CLI 1.16 and Boto3 1.9. For image acquisition, we used Zeiss Zen 2014.

We conducted most software development on a Mid-2014 MacBook Pro (Intel Core i7-4980HQ, 16 GB RAM, 1TB SSD) with MacOS 10.13, image stitching and nuclei detections on AWS EC2 c5.9xlarge instances (36 vCPUs, 72 GB RAM, typically configured with 1TB SSD) running RHEL 7.5 and Ubuntu 18.04, and volume measurements and aggregation on a Dell Precision T7500 workstation (64 GB RAM, 256 GB SSD and 2.2TB HDD) with Ubuntu 18.04. We conducted additional cross-platform compatibility testing on a Microsoft Surface Pro (2017) laptop (8 GB RAM, 256 GB SSD) with Microsoft windows 10 (build 1803), Windows Subsystem for Linux running Ubuntu 18.04, and Windows 10 virtual machines in VirtualBox 6.0.

Acknowledgements

We would like to thank the reviewers for their constructive feedback, including Harold Burgess. We are grateful for the microscopy advice and expertise from Meredith Calvert, director of the Gladstone Institutes Histology and Light Microscopy Core. We greatly appreciate Eric Lam from the KAVLI-PBBR Fabrication and Design Center at UCSF for assistance with the initial design of the custom 3D printed microscopy brain mount. We are deeply indebted to Hwee Kuan Lee from the Agency for Science, Technology and Research (A*STAR) for invaluable imaging informatics advice. We appreciate support from the Institute of Molecular and Cell Biology at A*STAR.

Funding Statement

The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.

Contributor Information

Stephan J Sanders, Email: stephan.sanders@ucsf.edu.

Joseph G Gleeson, Howard Hughes Medical Institute, The Rockefeller University, United States.

Kate M Wassum, University of California, Los Angeles, United States.

Funding Information

This paper was supported by the following grants:

  • Brain and Behavior Research Foundation NARSAD Young Investigator Grant to Stephan J Sanders.

  • National Institute of Mental Health U01 MH122681 to Stephan J Sanders.

  • National Institute of Mental Health R01 MH109901 to Stephan J Sanders.

  • National Institute of Neurological Disorders and Stroke R01 NS099099 to John LR Rubenstein.

Additional information

Competing interests

No competing interests declared.

Author contributions

Data curation, Software, Formal analysis, Validation, Investigation, Visualization, Methodology, Writing - original draft, Writing - review and editing, Designed all experiments, developed the 3D reconstruction tool, performed and optimized imaging.

Data curation, Investigation, Methodology, Performed mouse husbandry and tissue clearing, assisted with imaging.

Investigation, Optimized tissue clearing and sample mounting.

Methodology, Piloted tissue clearing and imaging.

Data curation, Validation, Annotated ground truth sets, software usability testing.

Data curation, Validation, Annotated ground truth sets, software usability testing.

Methodology, Computer vision and image processing guidance, figure review.

Resources, Supervision, Provided expertise on mouse development and anatomy.

Supervision, Methodology, Writing - review and editing.

Conceptualization, Resources, Data curation, Formal analysis, Supervision, Funding acquisition, Investigation, Visualization, Methodology, Writing - original draft, Project administration, Writing - review and editing, Supervised the entire project.

Ethics

Animal experimentation: All procedures and animal care were approved and performed in accordance with institutional guidelines from the University of California, San Francisco Laboratory Animal Research Center (LARC). All animal handling complied with the approved Institutional Animal Care and Use Committee (IACUC) protocol (AN180174-02) at the University of California, San Francisco.

Additional files

Supplementary file 1. Spreadsheet mapping colors to label names and IDs in the Allen ontology for each atlas in the ADMBA.
elife-61408-supp1.xlsx (206.1KB, xlsx)
Supplementary file 2. Spreadsheet of metric differences by label for WT brains before and after edge-aware atlas refinement.
elife-61408-supp2.xlsx (135.8KB, xlsx)
Transparent reporting form

Data availability

The full 3D generated atlases and wild-type brain images are being deposited with the Human Brain Project EBRAINS data platform. All data analyses are included in the manuscript and supporting files.

The following datasets were generated:

Young D, Yu W, Sanders SJ. 2020. E11.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. E13.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. E15.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. E18.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. P4 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. P28 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

Young D, Yu W, Sanders SJ. 2020. P56 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

References

  1. Alfaro-Almagro F, Jenkinson M, Bangerter NK, Andersson JLR, Griffanti L, Douaud G, Sotiropoulos SN, Jbabdi S, Hernandez-Fernandez M, Vallee E, Vidaurre D, Webster M, McCarthy P, Rorden C, Daducci A, Alexander DC, Zhang H, Dragonu I, Matthews PM, Miller KL, Smith SM. Image processing and quality control for the first 10,000 brain imaging datasets from UK biobank. NeuroImage. 2018;166:400–424. doi: 10.1016/j.neuroimage.2017.10.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Ali T, Asghar S, Sajid NA. Critical analysis of DBSCAN variations. In 2010 International Conference on Information and Emerging Technologies; 2010. pp. 1–6. [Google Scholar]
  3. Allen Institute for Brain Science . Allen Institute for Brain Science; 2013a. http://help.brain-map.org/download/attachments/4325389/DevMouse_Overview.pdf?version=1&modificationDate=1370999986503&api=v2 [Google Scholar]
  4. Allen Institute for Brain Science . Allen Institute for Brain Science; 2013b. http://help.brain-map.org/download/attachments/4325389/ReferenceAtlas_DevMouse.pdf?version=1&modificationDate=1370999986479&api=v2 [Google Scholar]
  5. Allen Institute for Brain Science Overview of the Design and Implementation of the Allen Mouse Common Coordinate Framework. [June 14, 2019];Allen Institute for Brain Science. 2017 http://help.brain-map.org/display/mousebrain/Documentation
  6. Allen Institute for Brain Science Allen Developing Mouse Brain Atlas API. [July 3, 2019];Allen Institute for Brain Science. 2018 http://help.brain-map.org/display/devmouse/API
  7. Amunts K, Ebell C, Muller J, Telefont M, Knoll A, Lippert T. The human brain project: creating a european research infrastructure to decode the human brain. Neuron. 2016;92:574–581. doi: 10.1016/j.neuron.2016.10.046. [DOI] [PubMed] [Google Scholar]
  8. Ardini-Poleske ME, Clark RF, Ansong C, Carson JP, Corley RA, Deutsch GH, Hagood JS, Kaminski N, Mariani TJ, Potter SS, Pryhuber GS, Warburton D, Whitsett JA, Palmer SM, Ambalavanan N, LungMAP Consortium LungMAP: the molecular atlas of lung development program. American Journal of Physiology-Lung Cellular and Molecular Physiology. 2017;313:L733–L740. doi: 10.1152/ajplung.00139.2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Bai W, Shi W, de Marvao A, Dawes TJ, O'Regan DP, Cook SA, Rueckert D. A bi-ventricular cardiac atlas built from 1000+ high resolution MR images of healthy subjects and an analysis of shape and motion. Medical Image Analysis. 2015;26:133–145. doi: 10.1016/j.media.2015.08.009. [DOI] [PubMed] [Google Scholar]
  10. Ballard DH, Brown CM. Computer Vision. First Edition. Prentice Hall Professional Technical Reference; 1982. [Google Scholar]
  11. Bjerke IE, Øvsthus M, Papp EA, Yates SC, Silvestri L, Fiorilli J, Pennartz CMA, Pavone FS, Puchades MA, Leergaard TB, Bjaalie JG. Data integration through brain atlasing: human brain project tools and strategies. European Psychiatry. 2018;50:70–76. doi: 10.1016/j.eurpsy.2018.02.004. [DOI] [PubMed] [Google Scholar]
  12. Bossolani GDP, Pintelon I, Detrez JD, Buckinx R, Thys S, Zanoni JN, De Vos WH, Timmermans JP. Comparative analysis reveals Ce3D as optimal clearing method for in toto imaging of the mouse intestine. Neurogastroenterology & Motility. 2019;31:e13560. doi: 10.1111/nmo.13560. [DOI] [PubMed] [Google Scholar]
  13. Bribiesca E. An easy measure of compactness for 2D and 3D shapes. Pattern Recognition. 2008;41:543–554. doi: 10.1016/j.patcog.2007.06.029. [DOI] [Google Scholar]
  14. Cai R, Pan C, Ghasemigharagoz A, Todorov MI, Förstera B, Zhao S, Bhatia HS, Parra-Damas A, Mrowka L, Theodorou D, Rempfler M, Xavier ALR, Kress BT, Benakis C, Steinke H, Liebscher S, Bechmann I, Liesz A, Menze B, Kerschensteiner M, Nedergaard M, Ertürk A. Panoptic imaging of transparent mice reveals whole-body neuronal projections and skull-meninges connections. Nature Neuroscience. 2019;22:317–327. doi: 10.1038/s41593-018-0301-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Chen Y, McElvain LE, Tolpygo AS, Ferrante D, Friedman B, Mitra PP, Karten HJ, Freund Y, Kleinfeld D. An active texture-based digital atlas enables automated mapping of structures and markers across brains. Nature Methods. 2019;16:341–350. doi: 10.1038/s41592-019-0328-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Chhetri RK, Amat F, Wan Y, Höckendorf B, Lemon WC, Keller PJ. Whole-animal functional and developmental imaging with isotropic spatial resolution. Nature Methods. 2015;12:1171–1178. doi: 10.1038/nmeth.3632. [DOI] [PubMed] [Google Scholar]
  17. Chon U, Vanselow DJ, Cheng KC, Kim Y. Enhanced and unified anatomical labeling for a common mouse brain atlas. Nature Communications. 2019;10:5067. doi: 10.1038/s41467-019-13057-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Chung K, Wallace J, Kim SY, Kalyanasundaram S, Andalman AS, Davidson TJ, Mirzabekov JJ, Zalocusky KA, Mattis J, Denisin AK, Pak S, Bernstein H, Ramakrishnan C, Grosenick L, Gradinaru V, Deisseroth K. Structural and molecular interrogation of intact biological systems. Nature. 2013;497:332–337. doi: 10.1038/nature12107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. de Boer BA, van den Berg G, de Boer PA, Moorman AF, Ruijter JM. Growth of the developing mouse heart: an interactive qualitative and quantitative 3D atlas. Developmental Biology. 2012;368:203–213. doi: 10.1016/j.ydbio.2012.05.001. [DOI] [PubMed] [Google Scholar]
  20. Delafontaine-Martel P, Lefebvre J, Tardif P-L, Lévy BI, Pouliot P, Lesage F. Whole brain vascular imaging in a mouse model of Alzheimer’s disease with two-photon microscopy. Journal of Biomedical Optics. 2018;23:1–501. doi: 10.1117/1.JBO.23.7.076501. [DOI] [PubMed] [Google Scholar]
  21. Dice LR. Measures of the amount of ecologic association between species. Ecology. 1945;26:297–302. doi: 10.2307/1932409. [DOI] [Google Scholar]
  22. Ecker JR, Geschwind DH, Kriegstein AR, Ngai J, Osten P, Polioudakis D, Regev A, Sestan N, Wickersham IR, Zeng H. The BRAIN initiative cell census consortium: lessons learned toward generating a comprehensive brain cell atlas. Neuron. 2017;96:542–557. doi: 10.1016/j.neuron.2017.10.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Ellegood J, Yee Y, Kerr TM, Muller CL, Blakely RD, Henkelman RM, Veenstra-VanderWeele J, Lerch JP. Analysis of neuroanatomical differences in mice with genetically modified serotonin transporters assessed by structural magnetic resonance imaging. Molecular Autism. 2018;9:24. doi: 10.1186/s13229-018-0210-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Erö C, Gewaltig M-O, Keller D, Markram H. A cell atlas for the mouse brain. Frontiers in Neuroinformatics. 2018;12:1662–5196. doi: 10.3389/fninf.2018.00084. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Ertürk A, Becker K, Jährling N, Mauch CP, Hojer CD, Egen JG, Hellal F, Bradke F, Sheng M, Dodt HU. Three-dimensional imaging of solvent-cleared organs using 3disco. Nature Protocols. 2012;7:1983–1995. doi: 10.1038/nprot.2012.119. [DOI] [PubMed] [Google Scholar]
  26. Ester M, Kriegel H-P, Sander J, Xu X. A density-based algorithm for discovering clusters a density-based algorithm for discovering clusters in large spatial databases with noise. Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, Ser. KDD’96; 1996. pp. 226–231. [Google Scholar]
  27. Franklin KBJ, Paxinos G. The Mouse Brain in Stereotaxic Coordinates. Academic Press; 2013. [Google Scholar]
  28. Gilbert K, Bai W, Mauger C, Medrano-Gracia P, Suinesiaputra A, Lee AM, Sanghvi MM, Aung N, Piechnik SK, Neubauer S, Petersen SE, Rueckert D, Young AA. Independent left ventricular morphometric atlases show consistent relationships with cardiovascular risk factors: a UK biobank study. Scientific Reports. 2019;9:1130. doi: 10.1038/s41598-018-37916-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Hammelrath L, Škokić S, Khmelinskii A, Hess A, van der Knaap N, Staring M, Lelieveldt BPF, Wiedermann D, Hoehn M. Morphological maturation of the mouse brain: an in vivo MRI and histology investigation. NeuroImage. 2016;125:144–152. doi: 10.1016/j.neuroimage.2015.10.009. [DOI] [PubMed] [Google Scholar]
  30. Hillman EMC, Voleti V, Li W, Yu H. Light-Sheet microscopy in neuroscience. Annual Review of Neuroscience. 2019;42:295–313. doi: 10.1146/annurev-neuro-070918-050357. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Hipp DR, Kennedy D, Mistachkin J. SQLite development team. 3.34.1Sqlite. 2021 https://www.sqlite.org/
  32. Ho DJ, Fu C, Salama P, Dunn KW, Delp EJ. Nuclei segmentation of fluorescence microscopy images using three dimensional convolutional neural networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW); 2017. pp. 834–842. [DOI] [Google Scholar]
  33. Hörl D, Rojas Rusak F, Preusser F, Tillberg P, Randel N, Chhetri RK, Cardona A, Keller PJ, Harz H, Leonhardt H, Treier M, Preibisch S. BigStitcher: reconstructing high-resolution image datasets of cleared and expanded samples. Nature Methods. 2019;16:870–874. doi: 10.1038/s41592-019-0501-0. [DOI] [PubMed] [Google Scholar]
  34. Huang ZJ, Luo L. It takes the world to understand the brain. Science. 2015;350:42–44. doi: 10.1126/science.aad4120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Hunter JD. Matplotlib: a 2D graphics environment. Computing in Science & Engineering. 2007;9:90–95. doi: 10.1109/MCSE.2007.55. [DOI] [Google Scholar]
  36. Intson K, van Eede MC, Islam R, Milenkovic M, Yan Y, Salahpour A, Henkelman RM, Ramsey AJ. Progressive neuroanatomical changes caused by Grin1 loss-of-function mutation. Neurobiology of Disease. 2019;132:104527. doi: 10.1016/j.nbd.2019.104527. [DOI] [PubMed] [Google Scholar]
  37. Jahanshad N, Kochunov PV, Sprooten E, Mandl RC, Nichols TE, Almasy L, Blangero J, Brouwer RM, Curran JE, de Zubicaray GI, Duggirala R, Fox PT, Hong LE, Landman BA, Martin NG, McMahon KL, Medland SE, Mitchell BD, Olvera RL, Peterson CP, Starr JM, Sussmann JE, Toga AW, Wardlaw JM, Wright MJ, Hulshoff Pol HE, Bastin ME, McIntosh AM, Deary IJ, Thompson PM, Glahn DC. Multi-site genetic analysis of diffusion images and voxelwise heritability analysis: a pilot project of the ENIGMA-DTI working group. NeuroImage. 2013;81:455–469. doi: 10.1016/j.neuroimage.2013.04.061. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Jones AR, Overly CC, Sunkin SM. The allen brain atlas: 5 years and beyond. Nature Reviews Neuroscience. 2009;10:821–828. doi: 10.1038/nrn2722. [DOI] [PubMed] [Google Scholar]
  39. Kim Y, Venkataraju KU, Pradhan K, Mende C, Taranda J, Turaga SC, Arganda-Carreras I, Ng L, Hawrylycz MJ, Rockland KS, Seung HS, Osten P. Mapping social behavior-induced brain activation at cellular resolution in the mouse. Cell Reports. 2015;10:292–305. doi: 10.1016/j.celrep.2014.12.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Kim Y, Yang GR, Pradhan K, Venkataraju KU, Bota M, García Del Molino LC, Fitzgerald G, Ram K, He M, Levine JM, Mitra P, Huang ZJ, Wang XJ, Osten P. Brain-wide maps reveal stereotyped Cell-Type-Based cortical architecture and subcortical sexual dimorphism. Cell. 2017;171:456–469. doi: 10.1016/j.cell.2017.09.020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Klein S, Staring M, Murphy K, Viergever MA, Pluim JP. Elastix: a toolbox for intensity-based medical image registration. IEEE Transactions on Medical Imaging. 2010;29:196–205. doi: 10.1109/TMI.2009.2035616. [DOI] [PubMed] [Google Scholar]
  42. Kolesová H, Čapek M, Radochová B, Janáček J, Sedmera D. Comparison of different tissue clearing methods and 3D imaging techniques for visualization of GFP-expressing mouse embryos and embryonic hearts. Histochemistry and Cell Biology. 2016;146:141–152. doi: 10.1007/s00418-016-1441-8. [DOI] [PubMed] [Google Scholar]
  43. Kornilov A, Safonov I. An overview of watershed algorithm implementations in open source libraries. Journal of Imaging. 2018;4:123. doi: 10.3390/jimaging4100123. [DOI] [Google Scholar]
  44. Koroshetz W, Gordon J, Adams A, Beckel-Mitchener A, Churchill J, Farber G, Freund M, Gnadt J, Hsu NS, Langhals N, Lisanby S, Liu G, Peng GCY, Ramos K, Steinmetz M, Talley E, White S. The state of the NIH BRAIN initiative. The Journal of Neuroscience. 2018;38:6427–6438. doi: 10.1523/JNEUROSCI.3174-17.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Kuhn HW. The hungarian method for the assignment problem. Naval Research Logistics Quarterly. 1955;2:83–97. doi: 10.1002/nav.3800020109. [DOI] [Google Scholar]
  46. Kunst M, Laurell E, Mokayes N, Kramer A, Kubo F, Fernandes AM, Förster D, Dal Maschio M, Baier H. A Cellular-Resolution atlas of the larval zebrafish brain. Neuron. 2019;103:21–38. doi: 10.1016/j.neuron.2019.04.034. [DOI] [PubMed] [Google Scholar]
  47. Lamprecht MR, Sabatini DM, Carpenter AE. CellProfiler: free, versatile software for automated biological image analysis. BioTechniques. 2007;42:71–75. doi: 10.2144/000112257. [DOI] [PubMed] [Google Scholar]
  48. Lee TC, Kashyap RL, Chu CN. Building skeleton models via 3-D medial surface axis thinning algorithms. CVGIP: Graphical Models and Image Processing. 1994;56:462–478. doi: 10.1006/cgip.1994.1042. [DOI] [Google Scholar]
  49. Lein ES, Hawrylycz MJ, Ao N, Ayres M, Bensinger A, Bernard A, Boe AF, Boguski MS, Brockway KS, Byrnes EJ, Chen L, Chen L, Chen TM, Chin MC, Chong J, Crook BE, Czaplinska A, Dang CN, Datta S, Dee NR, Desaki AL, Desta T, Diep E, Dolbeare TA, Donelan MJ, Dong HW, Dougherty JG, Duncan BJ, Ebbert AJ, Eichele G, Estin LK, Faber C, Facer BA, Fields R, Fischer SR, Fliss TP, Frensley C, Gates SN, Glattfelder KJ, Halverson KR, Hart MR, Hohmann JG, Howell MP, Jeung DP, Johnson RA, Karr PT, Kawal R, Kidney JM, Knapik RH, Kuan CL, Lake JH, Laramee AR, Larsen KD, Lau C, Lemon TA, Liang AJ, Liu Y, Luong LT, Michaels J, Morgan JJ, Morgan RJ, Mortrud MT, Mosqueda NF, Ng LL, Ng R, Orta GJ, Overly CC, Pak TH, Parry SE, Pathak SD, Pearson OC, Puchalski RB, Riley ZL, Rockett HR, Rowland SA, Royall JJ, Ruiz MJ, Sarno NR, Schaffnit K, Shapovalova NV, Sivisay T, Slaughterbeck CR, Smith SC, Smith KA, Smith BI, Sodt AJ, Stewart NN, Stumpf KR, Sunkin SM, Sutram M, Tam A, Teemer CD, Thaller C, Thompson CL, Varnam LR, Visel A, Whitlock RM, Wohnoutka PE, Wolkey CK, Wong VY, Wood M, Yaylaoglu MB, Young RC, Youngstrom BL, Yuan XF, Zhang B, Zwingman TA, Jones AR. Genome-wide atlas of gene expression in the adult mouse brain. Nature. 2007;445:168–176. doi: 10.1038/nature05453. [DOI] [PubMed] [Google Scholar]
  50. Li B, Christensen GE, Hoffman EA, McLennan G, Reinhardt JM. Establishing a normative atlas of the human lung: computing the average transformation and atlas construction. Academic Radiology. 2012;19:1368–1381. doi: 10.1016/j.acra.2012.04.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Liebmann T, Renier N, Bettayeb K, Greengard P, Tessier-Lavigne M, Flajolet M. Three-Dimensional study of Alzheimer's Disease Hallmarks Using the iDISCO Clearing Method. Cell Reports. 2016;16:1138–1152. doi: 10.1016/j.celrep.2016.06.060. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Lindeberg T. Detecting salient blob-like image structures and their scales with a scale-space primal sketch: a method for focus-of-attention. International Journal of Computer Vision. 1993;11:283–318. doi: 10.1007/BF01469346. [DOI] [Google Scholar]
  53. Linkert M, Rueden CT, Allan C, Burel J-M, Moore W, Patterson A, Loranger B, Moore J, Neves C, MacDonald D, Tarkowska A, Sticco C, Hill E, Rossner M, Eliceiri KW, Swedlow JR. Metadata matters: access to image data in the real world. Journal of Cell Biology. 2010;189:777–782. doi: 10.1083/jcb.201004104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Liu C, Ye FQ, Newman JD, Szczupak D, Tian X, Yen CC-C, Majka P, Glen D, Rosa MGP, Leopold DA, Silva AC. A resource for the detailed 3D mapping of white matter pathways in the marmoset brain. Nature Neuroscience. 2020;23:271–280. doi: 10.1038/s41593-019-0575-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Lorensen WE, Cline HE. Marching cubes: a high resolution 3d surface construction algorithm. Proceedings of the 14th Annual Conference on Computer Graphics and Interactive Techniques, Ser. SIGGRAPH’ 87; 1987. pp. 163–169. [Google Scholar]
  56. Lowekamp BC, Chen DT, Ibáñez L, Blezek D. The design of SimpleITK. Frontiers in Neuroinformatics. 2013;7:1662–5196. doi: 10.3389/fninf.2013.00045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Mano T, Albanese A, Dodt HU, Erturk A, Gradinaru V, Treweek JB, Miyawaki A, Chung K, Ueda HR. Whole-Brain analysis of cells and circuits by tissue clearing and Light-Sheet microscopy. The Journal of Neuroscience. 2018;38:9330–9337. doi: 10.1523/JNEUROSCI.1677-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Marquart GD, Tabor KM, Horstick EJ, Brown M, Geoca AK, Polys NF, Nogare DD, Burgess HA. High-precision registration between zebrafish brain atlases using symmetric diffeomorphic normalization. GigaScience. 2017;6:gix056. doi: 10.1093/gigascience/gix056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Marr D, Hildreth E. Theory of edge detection. Proceedings of the Royal Society of London. Series B, Biological Sciences. 1980;207:187–217. doi: 10.1098/rspb.1980.0020. [DOI] [PubMed] [Google Scholar]
  60. Marstal K, Berendsen F, Staring M, Klein S. SimpleElastix: a user-friendly, multi-lingual library for medical image registration. 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW); 2016. pp. 574–582. [DOI] [Google Scholar]
  61. McKinney W. Data structures for statistical computing in Python. Presented at the Proceedings of the 9th Python in Science Conference; 2010. pp. 51–56. [Google Scholar]
  62. Meijering E, Carpenter AE, Peng H, Hamprecht FA, Olivo-Marin JC. Imagining the future of bioimage analysis. Nature Biotechnology. 2016;34:1250–1255. doi: 10.1038/nbt.3722. [DOI] [PubMed] [Google Scholar]
  63. Miller KL, Alfaro-Almagro F, Bangerter NK, Thomas DL, Yacoub E, Xu J, Bartsch AJ, Jbabdi S, Sotiropoulos SN, Andersson JL, Griffanti L, Douaud G, Okell TW, Weale P, Dragonu I, Garratt S, Hudson S, Collins R, Jenkinson M, Matthews PM, Smith SM. Multimodal population brain imaging in the UK biobank prospective epidemiological study. Nature Neuroscience. 2016;19:1523–1536. doi: 10.1038/nn.4393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Millman KJ, Aivazis M. Python for scientists and engineers. Computing in Science & Engineering. 2011;13:9–12. doi: 10.1109/MCSE.2011.36. [DOI] [Google Scholar]
  65. Murakami TC, Mano T, Saikawa S, Horiguchi SA, Shigeta D, Baba K, Sekiya H, Shimizu Y, Tanaka KF, Kiyonari H, Iino M, Mochizuki H, Tainaka K, Ueda HR. A three-dimensional single-cell-resolution whole-brain atlas using CUBIC-X expansion microscopy and tissue clearing. Nature Neuroscience. 2018;21:625–637. doi: 10.1038/s41593-018-0109-1. [DOI] [PubMed] [Google Scholar]
  66. Murray SA, Morgan JL, Kane C, Sharma Y, Heffner CS, Lake J, Donahue LR. Mouse gestation length is genetically determined. PLOS ONE. 2010;5:e12418. doi: 10.1371/journal.pone.0012418. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Nazib A, Galloway J, Fookes C, Perrin D. Performance of registration tools on high- resolution 3d brain images, in 2018 40th. Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); 2018. pp. 566–569. [DOI] [PubMed] [Google Scholar]
  68. Neubert P, Protzel P. Compact watershed and preemptive SLIC: on improving trade- offs of superpixel segmentation algorithms. In 2014 22nd International Conference on Pattern Recognition.2014. [Google Scholar]
  69. Ng L, Pathak S, Kuan C, Lau C, Dong HW, Sodt A, Dang C, Avants B, Yushkevich P, Gee J, Haynor D, Lein E, Jones A, Hawrylycz M. Neuroinformatics for genome-wide 3D gene expression mapping in the mouse brain. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2007;4:382–393. doi: 10.1109/tcbb.2007.1035. [DOI] [PubMed] [Google Scholar]
  70. Niedworok CJ, Brown APY, Jorge Cardoso M, Osten P, Ourselin S, Modat M, Margrie TW. aMAP is a validated pipeline for registration and segmentation of high-resolution mouse brain data. Nature Communications. 2016;7:11–879. doi: 10.1038/ncomms11879. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Nowakowski TJ, Bhaduri A, Pollen AA, Alvarado B, Mostajo-Radji MA, Di Lullo E, Haeussler M, Sandoval-Espinosa C, Liu SJ, Velmeshev D, Ounadjela JR, Shuga J, Wang X, Lim DA, West JA, Leyrat AA, Kent WJ, Kriegstein AR. Spatiotemporal gene expression trajectories reveal developmental hierarchies of the human cortex. Science. 2017;358:1318–1323. doi: 10.1126/science.aap8809. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Oliphant TE. Python for scientific computing. Computing in Science & Engineering. 2007;9:10–20. doi: 10.1109/MCSE.2007.58. [DOI] [Google Scholar]
  73. Papp EA, Leergaard TB, Calabrese E, Johnson GA, Bjaalie JG. Waxholm space atlas of the sprague dawley rat brain. NeuroImage. 2014;97:374–386. doi: 10.1016/j.neuroimage.2014.04.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Mu¨ller A, Nothman J, Louppe G, Prettenhofer P, Weiss R, Dubourg V, Vanderplas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E. Scikit-learn: machine learning in Python. arXiv. 2018 https://arxiv.org/abs/1201.0490
  75. Polioudakis D, de la Torre-Ubieta L, Langerman J, Elkins AG, Shi X, Stein JL, Vuong CK, Nichterwitz S, Gevorgian M, Opland CK, Lu D, Connell W, Ruzzo EK, Lowe JK, Hadzic T, Hinz FI, Sabri S, Lowry WE, Gerstein MB, Plath K, Geschwind DH. A Single-Cell transcriptomic atlas of human neocortical development during Mid-gestation. Neuron. 2019;103:785–801. doi: 10.1016/j.neuron.2019.06.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Preibisch S, Saalfeld S, Tomancak P. Globally optimal stitching of tiled 3D microscopic image acquisitions. Bioinformatics. 2009;25:1463–1465. doi: 10.1093/bioinformatics/btp184. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Puelles L, Rubenstein JL. Forebrain gene expression domains and the evolving prosomeric model. Trends in Neurosciences. 2003;26:469–476. doi: 10.1016/S0166-2236(03)00234-0. [DOI] [PubMed] [Google Scholar]
  78. Ragan T, Kadiri LR, Venkataraju KU, Bahlmann K, Sutin J, Taranda J, Arganda-Carreras I, Kim Y, Seung HS, Osten P. Serial two-photon tomography for automated ex vivo mouse brain imaging. Nature Methods. 2012;9:255–258. doi: 10.1038/nmeth.1854. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Ramachandran P, Varoquaux G. Mayavi: 3d visualization of scientific data. Computing in Science & Engineering. 2011;13:40–51. doi: 10.1109/MCSE.2011.35. [DOI] [Google Scholar]
  80. Regev A, Teichmann SA, Lander ES, Amit I, Benoist C, Birney E, Bodenmiller B, Campbell P, Carninci P, Clatworthy M, Clevers H, Deplancke B, Dunham I, Eberwine J, Eils R, Enard W, Farmer A, Fugger L, Göttgens B, Hacohen N, Haniffa M, Hemberg M, Kim S, Klenerman P, Kriegstein A, Lein E, Linnarsson S, Lundberg E, Lundeberg J, Majumder P, Marioni JC, Merad M, Mhlanga M, Nawijn M, Netea M, Nolan G, Pe'er D, Phillipakis A, Ponting CP, Quake S, Reik W, Rozenblatt-Rosen O, Sanes J, Satija R, Schumacher TN, Shalek A, Shapiro E, Sharma P, Shin JW, Stegle O, Stratton M, Stubbington MJT, Theis FJ, Uhlen M, van Oudenaarden A, Wagner A, Watt F, Weissman J, Wold B, Xavier R, Yosef N, Human Cell Atlas Meeting Participants The human cell atlas. eLife. 2019;6:e27041. doi: 10.7554/eLife.27041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Renier N, Adams EL, Kirst C, Wu Z, Azevedo R, Kohl J, Autry AE, Kadiri L, Umadevi Venkataraju K, Zhou Y, Wang VX, Tang CY, Olsen O, Dulac C, Osten P, Tessier-Lavigne M. Mapping of brain activity by automated volume analysis of immediate early genes. Cell. 2016;165:1789–1802. doi: 10.1016/j.cell.2016.05.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Rueden CT, Schindelin J, Hiner MC, DeZonia BE, Walter AE, Arena ET, Eliceiri KW. ImageJ2: imagej for the next generation of scientific image data. BMC Bioinformatics. 2017;18:529. doi: 10.1186/s12859-017-1934-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Savolainen SM, Foley JF, Elmore SA. Histology atlas of the developing mouse heart with emphasis on E11.5 to E18.5. Toxicologic Pathology. 2009;37:395–414. doi: 10.1177/0192623309335060. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Schindelin J, Arganda-Carreras I, Frise E, Kaynig V, Longair M, Pietzsch T, Preibisch S, Rueden C, Saalfeld S, Schmid B, Tinevez JY, White DJ, Hartenstein V, Eliceiri K, Tomancak P, Cardona A. Fiji: an open-source platform for biological-image analysis. Nature Methods. 2012;9:676–682. doi: 10.1038/nmeth.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Schmid B, Schindelin J, Cardona A, Longair M, Heisenberg M. A high-level 3D visualization API for Java and ImageJ. BMC Bioinformatics. 2010;11:274. doi: 10.1186/1471-2105-11-274. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Schroeder W, Martin KM, Lorensen WE. Visualization Toolkit: An Object- Oriented Approach to 3D Graphics. Upper Saddle River: Prentice-Hall, Inc; 1998. [Google Scholar]
  87. Schubert E, Sander J, Ester M, Kriegel HP, Xu X. DBSCAN revisited, revisited: why and how you should (still) use DBSCAN. ACM Trans. Database Syst. 2017;42:1–21. doi: 10.1145/3068335. [DOI] [Google Scholar]
  88. Serra J. Image Analysis and Mathematical Morphology. Orlando, FL:  Academic Press, Inc; 1983. [Google Scholar]
  89. Shamonin DP, Bron EE, Lelieveldt BPF, Smits M, Klein S, Staring M. Fast parallel image registration on CPU and GPU for diagnostic classification of Alzheimer's disease. Frontiers in Neuroinformatics. 2014;7:00050. doi: 10.3389/fninf.2013.00050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Shuvaev SA, Lazutkin AA, Kedrov AV, Anokhin KV, Enikolopov GN, Koulakov AA. DALMATIAN: an algorithm for automatic cell detection and counting in 3D. Frontiers in Neuroanatomy. 2017;11:00117. doi: 10.3389/fnana.2017.00117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Sudlow C, Gallacher J, Allen N, Beral V, Burton P, Danesh J, Downey P, Elliott P, Green J, Landray M, Liu B, Matthews P, Ong G, Pell J, Silman A, Young A, Sprosen T, Peakman T, Collins R. UK biobank: an open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLOS Medicine. 2015;12:e1001779. doi: 10.1371/journal.pmed.1001779. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Suetterlin P, Hurley S, Mohan C, Riegman KLH, Pagani M, Caruso A, Ellegood J, Galbusera A, Crespo-Enriquez I, Michetti C, Yee Y, Ellingford R, Brock O, Delogu A, Francis-West P, Lerch JP, Scattoni ML, Gozzi A, Fernandes C, Basson MA. Altered neocortical gene expression, brain overgrowth and functional Over-Connectivity in Chd8 haploinsufficient mice. Cerebral Cortex. 2018;28:2192–2206. doi: 10.1093/cercor/bhy058. [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Susaki EA, Tainaka K, Perrin D, Kishino F, Tawara T, Watanabe TM, Yokoyama C, Onoe H, Eguchi M, Yamaguchi S, Abe T, Kiyonari H, Shimizu Y, Miyawaki A, Yokota H, Ueda HR. Whole-brain imaging with single-cell resolution using chemical cocktails and computational analysis. Cell. 2014;157:726–739. doi: 10.1016/j.cell.2014.03.042. [DOI] [PubMed] [Google Scholar]
  94. Susaki EA, Tainaka K, Perrin D, Yukinaga H, Kuno A, Ueda HR. Advanced CUBIC protocols for whole-brain and whole-body clearing and imaging. Nature Protocols. 2015;10:1709–1727. doi: 10.1038/nprot.2015.085. [DOI] [PubMed] [Google Scholar]
  95. Susaki EA, Ueda HR. Whole-body and Whole-Organ clearing and imaging techniques with Single-Cell resolution: toward Organism-Level systems biology in mammals. Cell Chemical Biology. 2016;23:137–157. doi: 10.1016/j.chembiol.2015.11.009. [DOI] [PubMed] [Google Scholar]
  96. Thompson PM, Stein JL, Medland SE, Hibar DP, Vasquez AA, Renteria ME, Toro R, Jahanshad N, Schumann G, Franke B, Wright MJ, Martin NG, Agartz I, Alda M, Alhusaini S, Almasy L, Almeida J, Alpert K, Andreasen NC, Andreassen OA, Apostolova LG, Appel K, Armstrong NJ, Aribisala B, Bastin ME, Bauer M, Bearden CE, Bergmann O, Binder EB, Blangero J, Bockholt HJ, Bøen E, Bois C, Boomsma DI, Booth T, Bowman IJ, Bralten J, Brouwer RM, Brunner HG, Brohawn DG, Buckner RL, Buitelaar J, Bulayeva K, Bustillo JR, Calhoun VD, Cannon DM, Cantor RM, Carless MA, Caseras X, Cavalleri GL, Chakravarty MM, Chang KD, Ching CR, Christoforou A, Cichon S, Clark VP, Conrod P, Coppola G, Crespo-Facorro B, Curran JE, Czisch M, Deary IJ, de Geus EJ, den Braber A, Delvecchio G, Depondt C, de Haan L, de Zubicaray GI, Dima D, Dimitrova R, Djurovic S, Dong H, Donohoe G, Duggirala R, Dyer TD, Ehrlich S, Ekman CJ, Elvsåshagen T, Emsell L, Erk S, Espeseth T, Fagerness J, Fears S, Fedko I, Fernández G, Fisher SE, Foroud T, Fox PT, Francks C, Frangou S, Frey EM, Frodl T, Frouin V, Garavan H, Giddaluru S, Glahn DC, Godlewska B, Goldstein RZ, Gollub RL, Grabe HJ, Grimm O, Gruber O, Guadalupe T, Gur RE, Gur RC, Göring HH, Hagenaars S, Hajek T, Hall GB, Hall J, Hardy J, Hartman CA, Hass J, Hatton SN, Haukvik UK, Hegenscheid K, Heinz A, Hickie IB, Ho BC, Hoehn D, Hoekstra PJ, Hollinshead M, Holmes AJ, Homuth G, Hoogman M, Hong LE, Hosten N, Hottenga JJ, Hulshoff Pol HE, Hwang KS, Jack CR, Jenkinson M, Johnston C, Jönsson EG, Kahn RS, Kasperaviciute D, Kelly S, Kim S, Kochunov P, Koenders L, Krämer B, Kwok JB, Lagopoulos J, Laje G, Landen M, Landman BA, Lauriello J, Lawrie SM, Lee PH, Le Hellard S, Lemaître H, Leonardo CD, Li CS, Liberg B, Liewald DC, Liu X, Lopez LM, Loth E, Lourdusamy A, Luciano M, Macciardi F, Machielsen MW, Macqueen GM, Malt UF, Mandl R, Manoach DS, Martinot JL, Matarin M, Mather KA, Mattheisen M, Mattingsdal M, Meyer-Lindenberg A, McDonald C, McIntosh AM, McMahon FJ, McMahon KL, Meisenzahl E, Melle I, Milaneschi Y, Mohnke S, Montgomery GW, Morris DW, Moses EK, Mueller BA, Muñoz Maniega S, Mühleisen TW, Müller-Myhsok B, Mwangi B, Nauck M, Nho K, Nichols TE, Nilsson LG, Nugent AC, Nyberg L, Olvera RL, Oosterlaan J, Ophoff RA, Pandolfo M, Papalampropoulou-Tsiridou M, Papmeyer M, Paus T, Pausova Z, Pearlson GD, Penninx BW, Peterson CP, Pfennig A, Phillips M, Pike GB, Poline JB, Potkin SG, Pütz B, Ramasamy A, Rasmussen J, Rietschel M, Rijpkema M, Risacher SL, Roffman JL, Roiz-Santiañez R, Romanczuk-Seiferth N, Rose EJ, Royle NA, Rujescu D, Ryten M, Sachdev PS, Salami A, Satterthwaite TD, Savitz J, Saykin AJ, Scanlon C, Schmaal L, Schnack HG, Schork AJ, Schulz SC, Schür R, Seidman L, Shen L, Shoemaker JM, Simmons A, Sisodiya SM, Smith C, Smoller JW, Soares JC, Sponheim SR, Sprooten E, Starr JM, Steen VM, Strakowski S, Strike L, Sussmann J, Sämann PG, Teumer A, Toga AW, Tordesillas-Gutierrez D, Trabzuni D, Trost S, Turner J, Van den Heuvel M, van der Wee NJ, van Eijk K, van Erp TG, van Haren NE, van 't Ent D, van Tol MJ, Valdés Hernández MC, Veltman DJ, Versace A, Völzke H, Walker R, Walter H, Wang L, Wardlaw JM, Weale ME, Weiner MW, Wen W, Westlye LT, Whalley HC, Whelan CD, White T, Winkler AM, Wittfeld K, Woldehawariat G, Wolf C, Zilles D, Zwiers MP, Thalamuthu A, Schofield PR, Freimer NB, Lawrence NS, Drevets W, Alzheimer’s Disease Neuroimaging Initiative, EPIGEN Consortium, IMAGEN Consortium, Saguenay Youth Study (SYS) Group The ENIGMA consortium: large-scale collaborative analyses of neuroimaging and genetic data. Brain Imaging and Behavior. 2014a;8:153–182. doi: 10.1007/s11682-013-9269-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Thompson CL, Ng L, Menon V, Martinez S, Lee CK, Glattfelder K, Sunkin SM, Henry A, Lau C, Dang C, Garcia-Lopez R, Martinez-Ferre A, Pombero A, Rubenstein JLR, Wakeman WB, Hohmann J, Dee N, Sodt AJ, Young R, Smith K, Nguyen TN, Kidney J, Kuan L, Jeromin A, Kaykas A, Miller J, Page D, Orta G, Bernard A, Riley Z, Smith S, Wohnoutka P, Hawrylycz MJ, Puelles L, Jones AR. A high-resolution spatiotemporal atlas of gene expression of the developing mouse brain. Neuron. 2014b;83:309–323. doi: 10.1016/j.neuron.2014.05.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Tustison N, Gee J. Introducing dice, Jaccard, and other label overlap measures to ITK. The Insight Journal. 2009;707:3141 [Google Scholar]
  99. Ueda HR, Ertürk A, Chung K, Gradinaru V, Chédotal A, Tomancak P, Keller PJ. Tissue clearing and its applications in neuroscience. Nature Reviews Neuroscience. 2020;21:61–79. doi: 10.1038/s41583-019-0250-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. van der Walt S, Colbert SC, Varoquaux G. The NumPy array: a structure for efficient numerical computation. Computing in Science & Engineering. 2011;13:22–30. doi: 10.1109/MCSE.2011.37. [DOI] [Google Scholar]
  101. van der Walt S, Schönberger JL, Nunez-Iglesias J, Boulogne F, Warner JD, Yager N, Gouillart E, Yu T, scikit-image contributors scikit-image: image processing in Python. PeerJ. 2014;2:e453. doi: 10.7717/peerj.453. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, Burovski E, Peterson P, Weckesser W, Bright J, van der Walt SJ, Brett M, Wilson J, Millman KJ, Mayorov N, Nelson ARJ, Jones E, Kern R, Larson E, Carey CJ, Polat İ, Feng Y, Moore EW, VanderPlas J, Laxalde D, Perktold J, Cimrman R, Henriksen I, Quintero EA, Harris CR, Archibald AM, Ribeiro AH, Pedregosa F, van Mulbregt P. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nature Methods. 2020;17:261–272. doi: 10.1038/s41592-019-0686-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Walhovd KB, Fjell AM, Westerhausen R, Nyberg L, Ebmeier KP, Lindenberger U, Bartrés-Faz D, Baaré WFC, Siebner HR, Henson R, Drevon CA, Strømstad Knudsen GP, Ljøsne IB, Penninx BWJH, Ghisletta P, Rogeberg O, Tyler L, Bertram L, Lifebrain Consortium Healthy minds 0–100 years: Optimising the use of European brain imaging cohorts (“Lifebrain”) European Psychiatry. 2018;50:47–56. doi: 10.1016/j.eurpsy.2017.12.006. [DOI] [PubMed] [Google Scholar]
  104. Wang Q, Ding S-L, Li Y, Royall J, Feng D, Lesnar P, Graddis N, Naeemi M, Facer B, Ho A, Dolbeare T, Blanchard B, Dee N, Wakeman W, Hirokawa KE, Szafer A, Sunkin SM, Oh SW, Bernard A, Phillips JW, Hawrylycz M, Koch C, Zeng H, Harris JA, Ng L. The allen mouse brain common coordinate framework: a 3D reference atlas. Cell. 2020;181:936–953. doi: 10.1016/j.cell.2020.04.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Watson C, Paxinos G, Puelles L. The Mouse Nervous System. Academic Press; 2012. [DOI] [Google Scholar]
  106. Watson AM, Watkins SC. Massive volumetric imaging of cleared tissue: the necessary tools to be successful. The International Journal of Biochemistry & Cell Biology. 2019;112:76–78. doi: 10.1016/j.biocel.2019.05.007. [DOI] [PubMed] [Google Scholar]
  107. Whitesell JD, Buckley AR, Knox JE, Kuan L, Graddis N, Pelos A, Mukora A, Wakeman W, Bohn P, Ho A, Hirokawa KE, Harris JA. Whole brain imaging reveals distinct spatial patterns of amyloid beta deposition in three mouse models of Alzheimer's disease. Journal of Comparative Neurology. 2019;527:2122–2145. doi: 10.1002/cne.24555. [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Xu J, Ma Y, Yu T, Zhu D. Quantitative assessment of optical clearing methods in various intact mouse organs. Journal of Biophotonics. 2019a;12:e201800134. doi: 10.1002/jbio.201800134. [DOI] [PubMed] [Google Scholar]
  109. Xu X, Guan Y, Gong H, Feng Z, Shi W, Li A, Ren M, Yuan J, Luo Q. Automated brain region segmentation for single cell resolution histological images based on Markov random field. Neuroinformatics. 2019b;18:181–197. doi: 10.1007/s12021-019-09432-z. [DOI] [PubMed] [Google Scholar]
  110. Yoo TS, Ackerman MJ, Lorensen WE, Schroeder W, Chalana V, Aylward S, Metaxas D, Whitaker R. Engineering and algorithm design for an image processing api: a technical report on ITK–the insight toolkit. Studies in Health Technology and Informatics. 2002;85:586–592. doi: 10.3233/978-1-60750-929-5-586. [DOI] [PubMed] [Google Scholar]

Decision letter

Editor: Joseph G Gleeson1
Reviewed by: Harold Burgess2

In the interests of transparency, eLife publishes the most substantive revision requests and the accompanying author responses.

Acceptance summary:

The paper provides MagellanMapper, a suite of tools to provide quantitative measures of brain structure to accurately labeled 3D digital atlases across mouse neural development, and demonstrates that the resulting brain parcellations are superior to a naive agglomeration of the existing 2D labels. The novel computational methods transform slice annotations in the Allen Developing Mouse Brain Atlas into digital 3D reference atlases. The response to reviewer comments was complete and persuasive.

Decision letter after peer review:

Thank you for submitting your article "Constructing and Optimizing 3D Atlases From 2D Data With Application to the Developing Mouse Brain" for consideration by eLife. Your article has been reviewed by two peer reviewers, and the evaluation has been overseen by Kate Wassum as the Senior Editor and Joseph Gleeson as Reviewing Editor. The following individual involved in review of your submission has agreed to reveal their identity: Harold Burgess (Reviewer #3).

The reviewers have discussed the reviews with one another and the Reviewing Editor has drafted this decision to help you prepare a revised submission.

We would like to draw your attention to changes in our revision policy that we have made in response to COVID-19 (https://elifesciences.org/articles/57162). Specifically, when editors judge that a submitted work as a whole belongs in eLife but that some conclusions require additional new data, as they do with your paper, we are asking that the manuscript be revised according to the guidelines below as much as possible, or for items that cannot be directly addressed due to COVID-19, either limit claims to those supported by data in hand, or to explicitly state that the relevant conclusions require additional supporting data. Our expectation is that the authors will eventually carry out the additional experiments and report on how they affect the relevant conclusions either in a preprint on bioRxiv or medRxiv, or if appropriate, as a Research Advance in eLife, either of which would be linked to the original paper.

Despite the availability of a high resolution, expertly annotated digital adult mouse brain atlas (Allen CCFv3), accurately labeled 3D digital atlases across mouse neural development are lacking. The authors have filled that gap by developing novel computational methods that transform slice annotations in the Allen Developing Mouse Brain Atlas into digital 3D reference atlases. They demonstrate that the resulting brain parcellations are superior to a naive agglomeration of the existing 2D labels, and provide MagellanMapper, a suite of tools to aid quantitative measures of brain structure. Cellular level whole-brain quantitative analysis is rapidly becoming a reality in many species and this manuscript provides a foundational resource for mouse developmental studies. The methods are sophisticated, carefully applied and thoroughly evaluated. The manuscript reports a computational approach to transforming available 2D atlases of mouse brains into the 3D volumetric datasets. By optimizing the "smoothing" steps, a better quality of such 3D atlases is produced claimed. In addition, the authors applied their method to the imaging dataset of neonatal mouse brains obtained by lightsheet microscopy, as proof of its potential utilization in research.

1) The pipeline of the method involved the "mirroring" before the "smoothing" steps. Is it possible to perform the "smoothing" of one hemisphere and then "mirror" the smoothed 3D atlas onto the other hemisphere to check for the alignment? By doing so, the other hemisphere could serve as an internal control for the quality and accuracy of the 3D atlas.

2) The authors developed the “edge-aware procedure”, employed to extend existing labels to unannotated lateral regions of the brain, taking advantage of intensity gradations in underlying microscope images. Authors should manually annotate a small part of the lateral brain region to compare accuracy and compare computationally generated labels to the partial lateral labels in P28 brain.

3) For more delicate subregions (e.g., those in the hypothalamus) without clear anatomical boundaries, this “edge-aware” adjustment step may become ineffective. What could then be done for these subregions? Also, it is important to note that the anatomical edges required the manual annotation.

4) Annotations present in the ADMBA took advantage of co-aligned ISH data (and computational approaches using co-aligned gene expression data have been used for de novo brain parcellation). Intensity differences in the light-microscope images may not provide enough contrast or access this expression data for for accurate segmentation. Could there be instances where adjacent regions do not have intensity differences, and the edge-aware procedure actually reduces the accuracy of the manual annotation? What is the evidence that contrast is sufficient to demark the boundaries?

5) It does appear that despite the care to avoid losing thin structures, there is some loss, for example for the light-green structure in the forebrain in Figure 5E. Authors should indicate if all labels were preserved, and provide information on volume changes by label size.

6) The accuracy of non-rigid registration of light-sheet images to the references is assessed only using a DSC value for whole-brain overlaps. This does not assess the precision of registration within the brain. The authors should apply some other measure to assess quality of alignment within the brain (e.g. mark internal landmarks visible in the reference and original light-sheet images, and measure the post-registration distance between them).

eLife. 2021 Feb 11;10:e61408. doi: 10.7554/eLife.61408.sa2

Author response


Despite the availability of a high resolution, expertly annotated digital adult mouse brain atlas (Allen CCFv3), accurately labeled 3D digital atlases across mouse neural development are lacking. The authors have filled that gap by developing novel computational methods that transform slice annotations in the Allen Developing Mouse Brain Atlas into digital 3D reference atlases. They demonstrate that the resulting brain parcellations are superior to a naive agglomeration of the existing 2D labels, and provide MagellanMapper, a suite of tools to aid quantitative measures of brain structure. Cellular level whole-brain quantitative analysis is rapidly becoming a reality in many species and this manuscript provides a foundational resource for mouse developmental studies. The methods are sophisticated, carefully applied and thoroughly evaluated. The manuscript reports a computational approach to transforming available 2D atlases of mouse brains into the 3D volumetric datasets. By optimizing the "smoothing" steps, a better quality of such 3D atlases is produced claimed. In addition, the authors applied their method to the imaging dataset of neonatal mouse brains obtained by lightsheet microscopy, as proof of its potential utilization in research.

1) The pipeline of the method involved the "mirroring" before the "smoothing" steps. Is it possible to perform the "smoothing" of one hemisphere and then "mirror" the smoothed 3D atlas onto the other hemisphere to check for the alignment? By doing so, the other hemisphere could serve as an internal control for the quality and accuracy of the 3D atlas.

Thank you for suggesting the use of one hemisphere as an internal control for the other hemisphere. We’ve added these analyses as Figure 4—figure supplement 3 and we reference the analysis in the main text. In our current setup, we mirror both the labels and the underlying histological image to the opposite hemisphere since the histological images have asymmetries that would not correspond with the mirrored labels. Because the two hemispheres are now identical, smoothing before versus after mirroring would also be identical. Alternatively, we could mirror the labels only (i.e. not the underlying histological image) to the opposite hemisphere before smoothing. However, the "edge-aware" smoothing step is based on the assumption that the core of each label obtained through erosion represents pixels with the highest probability of accurate placement, and for many labels this assumption is not correct for the other hemisphere due to the asymmetry.

However, we can use the asymmetric hemispheres as internal controls if we register the mirrored atlases back to their unmirrored versions. This registration allows us to compare the quality and accuracy in the hemisphere on which the atlases were not derived, tested through segmentation by 3D image registration, which is a typical use case of the atlases. We can then compare the quality and accuracy between the unsmoothed and smoothed labels.

Taking each of our original (mirrored) and smoothed (edge-aware) atlases across the ADMBA, we registered each atlas to its original (unmirrored) histological image (Figure 4—figure supplement 3A). We first tested the hemisphere from which the atlases were derived and observed a significant improvement with the smoothed atlas for within region homogeneity (median coefficient of variation 0.340 unsmoothed vs. 0.331 smoothed, p = 0.008, WSRT; Figure 4—figure supplement 3B) and distance between anatomical and label edges (median of 91 million µm unsmoothed vs. 50 million µm smoothed, p = 0.008, WSRT, Figure 4—figure supplement 2C). Repeating this on the contralateral, asymmetric hemispheres as an internal control, we observed similar evidence of improvement for both within region homogeneity (median of 0.298 unsmoothed vs. 0.291 smoothed, p = 0.008, WSRT; Figure 4—figure supplement 3D) and distance between anatomical and label edges (median of 89 million µm unsmoothed vs. 41 million µm smoothed, p = 0.008, WSRT; Figure 4—figure supplement 3E). By testing the generalizability of the 3D reconstructed atlases to new hemispheres through this internal control, we demonstrate that we can obtain similar improvements in anatomical fidelity when applied to other brains through image registration.

2) The authors developed the “edge-aware procedure”, employed to extend existing labels to unannotated lateral regions of the brain, taking advantage of intensity gradations in underlying microscope images. Authors should manually annotate a small part of the lateral brain region to compare accuracy and compare computationally generated labels to the partial lateral labels in P28 brain.

This is a great idea and one that we had not considered before. Again, we’ve added these analyses as Figure 4—figure supplement 4, and we reference the analysis in the main text. To measure the accuracy of the computationally generated labels, we manually annotated structures in the P28 lateral planes, which also allowed comparison with the partial lateral labels in the original P28 atlas. We annotated the hippocampus and Layer 1 as the most clearly demarcated structures in these planes (Figure 4—figure supplement 4D) and compared them against the equivalent composite of sub-structures in the original, partially annotated planes (Figure 4—figure supplement 4A-C) and the 3D reconstructed labels (Figure 4—figure supplement 4E-G) using the Dice Similarity Coefficient (DSC) metric. The DSC for the original labels was 0.65 (weighted average by volume from a DSC of 0.81 for the hippocampus and 0.54 for Layer 1), compared with 0.75 for the computationally generated labels (0.82 for the hippocampus and 0.71 for Layer 1), supporting the accuracy of our imputed labels. Of note, the partially labeled planes are not used in the computationally generated planes, which essentially start from scratch using the last fully labeled plane while still yielding a higher DSC than for the partial labels and the artifact reduction evident in the axial and coronal views (Figure 4—figure supplement 4A-B, F-G).

To account for the missing labels in the original planes, we masked the ground truth to the extent of all labels in each plane before calculating the DSC, which favors the original, partial labels but avoids penalizing them where they were simply unannotated. Without this masking, the original labels' DSC drops to 0.42 (hippocampus: 0.50, Layer 1: 0.37), compared with an essentially unchanged DSC of 0.74 (hippocampus: 0.82, Layer 1: 0.69) in the computationally generated labels. We also compared the partially labeled planes with the generated planes, which showed a weighted average DSC of 0.58 across all labels, or 0.70 for the same structures compared in the manual annotations (hippocampus: 0.79, Layer 1: 0.63), consistent with the discrepancy observed between each set of labels' similarity with the manual annotations and the fact that these partially labeled planes are not used in the computational label generation.

3) For more delicate subregions (e.g., those in the hypothalamus) without clear anatomical boundaries, this “edge-aware” adjustment step may become ineffective. What could then be done for these subregions? Also, it is important to note that the anatomical edges required the manual annotation.

At the heart of this issue is the challenge that manual annotation is likely superior in the sagittal plane, but leads to substantial artifacts in the axial and coronal planes. We have demonstrated that our procedure improves the axial and coronal views, however this is bound to be at the expense of the accuracy of the sagittal alignment. As the reviewers suggest, the key question becomes, whether this trade off is worthwhile, especially in these delicate subregions. We have demonstrated that the answer is “yes” qualitatively and quantitatively across the brain, and at the reviewers’ suggestion, we now also focus on two complicated subregions in greater depth. Again, we are confident the answer is “yes”, based on the minimal changes in the boundary location in the sagittal plane in the absence of clear anatomical boundaries (see Figure 4—figure supplement 5A, and also Author response image 1 and Figure 4—figure supplement 5B in point 4). In short, if anatomical data does not support moving the boundary, we designed our approach to smooth the boundary in the axial and coronal planes without moving the location in any of the planes.

Author response image 1. Preservation of regions demarcated by ISH markers.

Author response image 1.

(A) The Wnt3a ISH signal (dashed circle) shown in Thompson CL, et al. (2014, Figure 2; top row) is expressed selectively within the cortical hem in the original labels of the E13.5 atlas (lower middle row, orange structure) and remains contained in this region in the 3D reconstructed atlas (bottom row). (B) The Hoxa2 ISH signal demarcates the border between the pontine hindbrain and prepontine hindbrain in the E15.5 atlas (top row). This boundary remains well-demarcated before (lower middle row) and after (bottom row) 3D reconstruction.

The edge-aware step performs two major functions: 1) erosion of each label to its core, the annotated pixels of highest confidence, which removes the jagged border artifact, and 2) regrowing this high-confidence core to meet either anatomical boundaries if they are present, or if not, simply to meet neighboring labels in the middle where the original boundary was located. This approach allows labels to take advantage of smoothing by the edge-aware step even without nearby anatomical boundaries while maintaining their core location, shape, and spatial relationships. Because some edge artifacts persist after this step, we further smooth each label through the morphological filter step, which smooths each label in-place without using anatomical information.

To illustrate this approach in the hypothalamus, a delicate subregion not known for its clear demarcations, we show the labeled ventromedial hypothalamic nucleus (VMH; Figure 4—figure supplement 5A, top row), which contains a combination of labels near clear boundaries and those that are not (Figure 4—figure supplement 5A, upper middle row). The anatomical surface map, shown here as edges in 2D, demarcates these clear boundaries while more ambiguous boundaries remain undemarcated (Figure 4—figure supplement 5A, lower middle row). After the erosion step reduced each label to its high-confidence core, these eroded labels are each bounded on three sides by anatomical boundaries, but no boundary exists between them. During the watershed step, these labels' regrowth is guided by the anatomical boundaries on these three sides but simply meet in the middle of the two labels, where the original boundary was located, in the absence of a clear anatomical boundary. The subsequent morphological smoothing step further reduced remaining artifacts (Figure 4—figure supplement 5A, bottom row). The coronal and axial views reveal the impact of smoothing even in the absence of anatomical boundaries by removal of the jagged edge artifact while respecting the core position of each label.

4) Annotations present in the ADMBA took advantage of co-aligned ISH data (and computational approaches using co-aligned gene expression data have been used for de novo brain parcellation). Intensity differences in the light-microscope images may not provide enough contrast or access this expression data for for accurate segmentation. Could there be instances where adjacent regions do not have intensity differences, and the edge-aware procedure actually reduces the accuracy of the manual annotation? What is the evidence that contrast is sufficient to demark the boundaries?

This question builds on the concepts raised in point 3, since our simplest answer to this is that we designed the procedure to only move a boundary in the presence of clear anatomical data. While we did not utilize the ISH data directly, by preserving the original boundaries of most labels we maintain this information in the final atlas. We are very interested in integrating gene expression data into atlas boundary identification in a systematic manner and hope to pursue this direction in our future work in this field either by integrating existing ISH data or with emerging spatial transcriptomic approaches.

We are not aware of a record of when, where, and which markers in the ISH data were used to define boundaries in the initial atlas. However, from the ADMBA reference paper (Thompson CL, et al., 2014), we can see that Wnt3a was used to define boundaries as a ligand "selectively expressed in the E13.5 cortical hem," where the ISH signal is contained within the larger cortical hem region. Therefore, we examined this region in a similar manner to the ventromedial hypothalamus (Figure 4—figure supplement 5A) and show that the region expressing Wnt3a (Author response image 1A, dashed circle) remains contained within similar boundaries after atlas smoothing. The paper also presents Hoxa2 as a marker expressed predominantly in the hindbrain during early embryonic periods up through E15.5, except only minimally in the prepontine hindbrain, the region just superior to the pontine hindbrain. We examined the hindbrain in the E15.5 atlas and show preservation of the boundary between the prepontine hindbrain and the pontine hindbrain before and after smoothing, matching the boundary shown by the Hoxa2 ISH signal (Author response image 1B, dashed line; PPH = prepontine hindbrain; PH = pontine hindbrain).

We acknowledge that some adjacent regions do not have intensity differences to demarcate their boundaries. To minimize the chance of reducing the accuracy of the original manual annotation in such regions, we have used a relatively coarse anatomical edge map (generated as a surface map in 3D) to capture only clearly demarcated boundaries with strong contrast differential. Similarly to the ventromedial hypothalamus (Figure 4—figure supplement 5A), we further illustrate these boundaries in the pontine hindbrain examined (Author response image 1), an area of low contrast around many labels but also a few more clearly delineated anatomical boundaries near other labels (Figure 4—figure supplement 5B, upper rows). By using a large σ in the Laplacian of Gaussian boundary detection on the microscopy images, we limited the number of artifactual boundaries around low contrast regions. By performing the boundary detection in 3D, we also reduced detection of contrast differences limited to a single plane.

The resulting anatomical map demarcated only clear, gross boundaries for the edge-aware step, with a minimum of artifactual boundaries that might otherwise reduce the accuracy of the original labels (Figure 4—figure supplement 5B, lower middle row). Labels nearby these boundaries are influenced by and thus take advantage of them, whereas labels far away from boundaries are predominantly influenced by neighboring labels instead, meeting between them and retaining their overall original position and shape (Figure 4—figure supplement 5B, bottom row). In this way, we derive the sufficiency to demarcate boundaries from the original labels, reshaping them only when a label's boundary is nearby but not completely aligned with a clear anatomical boundary. In general, we qualitatively noticed larger artifacts in the original labels, such as the jaggedness in the orthogonal planes, than artifacts caused by the edge-aware step, which is designed to improve accuracy while respecting and maintaining the overall shape of the manual annotations. We have clarified this description in the manuscript text.

5) It does appear that despite the care to avoid losing thin structures, there is some loss, for example for the light-green structure in the forebrain in Figure 5E. Authors should indicate if all labels were preserved, and provide information on volume changes by label size.

Thank you for pointing out this omission in the text. Investigating this led to further improvements in the atlas. All labels in the original hemisphere were preserved across all of the atlases and we now state this in the text. For completeness, we note that labels positioned solely across the sagittal midline on the otherwise unlabeled hemisphere were cut off to preserve the placement of the rest of the labels and the symmetry of the brain without duplicating the underlying microscopy image at the midline as indicated in the manuscript (Materials and methods, "Atlas 3D rotation and mirroring,").

Instead of the complete loss of a structure, the loss of the light-green structure that the reviewer notes represents variation in how far a thin appendage of a larger structure (medial pallium in the P4 atlas) extends. We have examined this region in detail (see Author response image 2) and find the cause to be due to inconsistencies in the original atlas labels.

Author response image 2. Interpolation to fill gaps in and recover a thin substructure.

Author response image 2.

(A) A thin extension in the medial pallium (green structure) in the P4 atlas is missing sagittal annotations in planes 152 and 155, visible as two gaps in the axial and coronal view, as well as gaps between the extension and the rest of the structure in planes 151, 154, and 155. (B) Reconnecting this extension (left) and interpolating between the edited planes fills the gaps as seen in the axial view (middle), recovering this thin substructure in the final smoothed atlas. (C) This substructure is apparent in the anterior section of the original atlas (left) but lost in the 3D reconstructed atlas (middle), although the rest of the medial pallium is present in other planes. After interpolation and 3D reconstruction, the substructure is recovered (right).

As seen in the sequence of sagittal planes, two of six planes are missing annotations in this label, visible as two separate gaps in the axial views (Author response image 2A, 152 and 155). Of the remaining four labeled planes, three are missing connections to the main structure, instead stopping at a harsh border (Author response image 2A, 151, 154, 156). Only panel 153 shows a consistent and connected appendage.

To test the extent to which these inconsistencies in the underlying atlas led to the loss of this thin structure, we manually edited the first and last plane to connect it to the main structure (Author response image 2B, left), followed by edge interpolation between these planes to fill in the missing planes (Author response image 2B, center). After the edge-aware step with a slightly reduced filter kernel size (8 to 7) and final smoothing step on the entire atlas, the thin structure was maintained, and harsh edges adapted to the underlying contours in the microscopy image (Author response image 2B, right). Compared with the planes in Figure 5E, the structure is retained and refined with this interpolation (Author response image 2C).

While this interpolation procedure can be used to rescue this structure, we are hesitant to apply such methods across the entire atlas since it is hard to distinguish such inconsistencies from the artifacts we are trying to remove. An alternative approach to rescuing this structure would be to reduce the maximum erosion filter kernel size, but this change would also affect many other structures as the algorithm is applied in an automated fashion. Theoretically, a fully adaptive morphological filter would account for heterogeneity within structures but is particularly challenging in 3D and another focus of our future work.

To depict volume changes in each label across the ADMBA, we plotted the change in volume from the mirrored to the smoothed (edge-aware) atlases by total volume for each label (Figure 4—figure supplement 2). By adapting the morphological filter sizes during both the edge-aware and final smoothing steps, we avoid label loss and minimize changes relative to label size as seen in the smaller labels (3,132 labels are present where x < 0.2). A caveat to the label preservation is that thin structures attached to larger regions can be lost as described above. We specifically designed the skeletonization step to address this issue to maximize retention of the original footprint including thin sections during the erosion process, from which to regrow all sections.

6) The accuracy of non-rigid registration of light-sheet images to the references is assessed only using a DSC value for whole-brain overlaps. This does not assess the precision of registration within the brain. The authors should apply some other measure to assess quality of alignment within the brain (e.g. mark internal landmarks visible in the reference and original light-sheet images, and measure the post-registration distance between them).

We note that since we have not changed the intensity images, other than to mirror them for symmetry, and our image registration is between the sample and atlas intensity images, our approach does not impact the image registration. We have assessed internal alignment through image registration by marking visible landmarks in the mirrored E18.5 atlas and each of the wild-type sample brains followed by measurement of the post-registration distances. Median distance across landmarks is 1,481 µm before and 135 µm after image registration (Author response image 3). The intensity coefficient of variation metric assesses the quality of the labels, rather than the intensity images or registration, although we expect that further improvements in registration would better demonstrate improvements in the atlases. It seems clear that there is substantial scope for improving registration to 3D images. To this end, we have also wondered whether application of the edge-aware approach post-registration could augment the registration process, though this would require a substantial rewrite of such algorithms.

Author response image 3. Post-registration landmark distance for wild-type lightsheet images.

Author response image 3.

Mean distances between manually annotated atlas and wild-type brain image landmarks are plotted for each brain after registration.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Citations

    1. Young D, Yu W, Sanders SJ. 2020. E11.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    2. Young D, Yu W, Sanders SJ. 2020. E13.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    3. Young D, Yu W, Sanders SJ. 2020. E15.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    4. Young D, Yu W, Sanders SJ. 2020. E18.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    5. Young D, Yu W, Sanders SJ. 2020. P4 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    6. Young D, Yu W, Sanders SJ. 2020. P28 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]
    7. Young D, Yu W, Sanders SJ. 2020. P56 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS. [DOI]

    Supplementary Materials

    Supplementary file 1. Spreadsheet mapping colors to label names and IDs in the Allen ontology for each atlas in the ADMBA.
    elife-61408-supp1.xlsx (206.1KB, xlsx)
    Supplementary file 2. Spreadsheet of metric differences by label for WT brains before and after edge-aware atlas refinement.
    elife-61408-supp2.xlsx (135.8KB, xlsx)
    Transparent reporting form

    Data Availability Statement

    The full 3D generated atlases and wild-type brain images are being deposited with the Human Brain Project EBRAINS data platform. All data analyses are included in the manuscript and supporting files.

    The following datasets were generated:

    Young D, Yu W, Sanders SJ. 2020. E11.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. E13.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. E15.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. E18.5 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. P4 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. P28 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.

    Young D, Yu W, Sanders SJ. 2020. P56 3D Edge-Aware Refined Atlas Derived from the Allen Developing Mouse Brain Atlas. EBRAINS.


    Articles from eLife are provided here courtesy of eLife Sciences Publications, Ltd

    RESOURCES