Abstract
Identification and delineation of brain regions in histologic mouse brain sections is especially pivotal for many neurogenomics, transcriptomics, proteomics and connectomics studies, yet this process is prone to observer error and bias. Here we present a novel brain navigation system, named NeuroInfo, whose general principle is similar to that of a global positioning system (GPS) in a car. NeuroInfo automatically navigates an investigator through the complex microscopic anatomy of histologic sections of mouse brains (thereafter: “experimental mouse brain sections”). This is achieved by automatically registering a digital image of an experimental mouse brain section with a three-dimensional (3D) digital mouse brain atlas that is essentially based on the third version of the Allen Mouse Brain Common Coordinate Framework (CCF v3), retrieving graphical region delineations and annotations from the 3D digital mouse brain atlas, and superimposing this information onto the digital image of the experimental mouse brain section on a computer screen. By doing so, NeuroInfo helps in solving the long-standing problem faced by researchers investigating experimental mouse brain sections under a light microscope—that of correctly identifying the distinct brain regions contained within the experimental mouse brain sections. Specifically, NeuroInfo provides an intuitive, readily-available computer microscopy tool to enhance researchers’ ability to correctly identify specific brain regions in experimental mouse brain sections. Extensive validation studies of NeuroInfo demonstrated that this novel technology performs remarkably well in accurately delineating regions that are large and/or located in the dorsal parts of mouse brains, independent on whether the sections were imaged with fluorescence or brightfield microscopy. This novel navigation system provides a highly efficient way for registering a digital image of an experimental mouse brain section with the 3D digital mouse brain atlas in a minute and accurate delineation of the image in real-time.
Keywords: Allen Mouse Brain Atlas, Allen Common Coordinate Framework, brain region delineation and identification, mouse brain, mouse brain reference atlas, neuroanatomy, NeuroInfo, software
GRAPHICAL ABSTRACT
We developed a novel mouse brain navigation system (NeuroInfo) whose general principle is similar to that of a GPS in a car. NeuroInfo automatically navigates a researcher through the complex microscopic anatomy of histologic sections of mouse brains.
1.|. INTRODUCTION
Microscopic analysis of histologic brain sections provides important information about the organization and structure of the nervous system, especially important for neurogenomics, transcriptomics, proteomics and connectomics studies. In this context, the BRAIN Initiative (see Insel, Landis & Collins, 2013; Koroshetz et al., 2018) places emphasis on identifying specific anatomic and physiologic properties of cells within the brain. In order to do this rigorously and objectively, a systematic way of assigning cells and axonal projection targets to appropriate brain regions is required. Unfortunately, the identification of distinct brain regions in a tissue section under study (thereafter, “experimental section”) with the currently available tools (printed and digital brain atlases) is tedious, cumbersome and prone to error. The difficulties faced by researchers using brain atlases to identify brain regions in experimental sections are illustrated in Figure 1a for the mouse brain through an example from the GENSAT (Gene Expression Nervous System Atlas) project (Gerfen, Paletzki & Heintz, 2013; Pastrana, 2014). Identification of the regions of interest (green and yellow arrows in Fig. 1a) by means of commonly used approaches is hampered by many factors. First, an investigator must move back and forth across plates of a mouse brain atlas to find a section in the atlas that is most similar to the experimental section. Then, the three-dimensional (3D) boundaries of the regions of interest need to be inferred and translated from the static medium of the atlas to the experimental section visualized as an image or under a light microscope. This process must then be repeated for every experimental section under investigation. Second, finding a mouse brain atlas with scale and resolution that complements the experimental section image is complicated by the fact that the cutting angle of histologic sections may deviate from canonical planes. This issue is practically unavoidable when sectioning mouse brains. The process of matching an experimental section image to a plane in a mouse brain atlas when the experimental section is not precisely aligned with the atlas is difficult and results in highly subjective judgments. To date, no readily available tool exists for researchers to match oblique histology sections to multiple planes of a brain atlas, objectively and automatically.
To overcome the challenges of subjective judgments in atlas referencing, we developed a novel mouse brain navigation system (NeuroInfo) that 1) automatically registers images of experimental mouse brain sections with a 3D digital mouse brain atlas that is essentially based on the third version of the Allen Mouse Brain Common Coordinate Framework (CCF v3) (Allen Institute for Brain Science, 2017); 2) retrieves graphical region delineations and annotations from this 3D digital mouse brain atlas; and 3) superimposes this information onto the experimental section image on a computer screen. Data annotation is essentially achieved in two steps: section-to-atlas registration followed by atlas-based segmentation.
The term 3D digital mouse brain atlas is defined here as a collection of 3D digital files comprising anatomic mouse brain images and associated delineations, which includes the following elements: a 3D reference image embedded in a physical coordinate system, spatially-aligned 3D annotations, and a related ontology that provides the relationship among regions. The 3D reference image of the CCF v3 is based on a 3D, 10 μm isotropic, highly detailed population average of 1675 mouse brains that were imaged with serial two-photon tomography using a TissueCyte 1000 system (TissueVision, Somerville, MA, USA), aligned to a common frame. The 3D annotations of the CCF v3 include delineations of 738 regions based on multimodal references and the related ontology (Allen Institute for Brain Science, 2017).
Here we present a detailed description of the major components of NeuroInfo and their functionality, as well as the results of extensive validation studies performed on 60 experimental coronal sections from 12 different mouse brains that were generated in two different laboratories and imaged with either fluorescence or brightfield microscopy.
2.|. METHODS
2.1 |. Creation of software and implementation of section-to-atlas registration in NeuroInfo
NeuroInfo was developed as a fully integrated desktop application written in C++, using an object-oriented design approach. The software runs on Windows 7/8/10 (64-bit).
We used a specialized intensity-based 3D image registration method to establish the alignment of two-dimensional (2D) images of experimental sections (thereafter, “experimental section images”) to the 3D reference image of the CCF v3 (see Brown, 1992). Our implementation is based on the National Library of Medicine Insight Segmentation and Registration toolkit (ITK) with several significant customizations and modifications (Avants et al., 2014). In general, image registration optimizes the parameters of a transform that maps points from a “fixed” image onto corresponding points in a “moving” image. The optimization process minimizes the value of an image comparison metric that evaluates how well the images align based on intensity values observed at corresponding locations (Zitová and Flusser, 2003). In our implementation, the registration is performed in physical space—an origin and pixel/voxel spacing establishes a coordinate frame for each image—to make the transform independent of the digital resolution of the images involved. The 2D experimental section image serves as the fixed image and is treated as a 3D image that has a thickness on the order of the physical section along the sectioning direction (e.g. 80 μm). Our current implementation supports visualization of the cytoarchitecture in experimental section images with fluorescent (as shown in Fig. 2a-c) or thionin (as shown in Fig. 2d-f) Nissl stains, and these single channels were used for registration. The two-photon autofluorescence reference image of the CCF v3 was used as the moving image.
The specialized image registration process employs multiple stages of multiple-resolution coarse-to-fine refinement. Each stage of the process models an aspect of the physical process of histologic embedding, sectioning and mounting, and operates on a subset of transform parameters. The first stage is initialized with a rigid orientation transform that accounts for the plane of sectioning and the predominant orientation in which the experimental section was imaged (e.g. coronal sections processed along the rostral-caudal axis of the mouse brain with the dorsal surface appearing at the top of the image). The first stage incorporates a search at a low image resolution through the 3D reference image of the CCF v3 along the sectioning axis, using a small number of transform parameters that maintain an orthogonal virtual sectioning through the 3D reference image of the CCF v3. This stage models the position of the experimental section along the sectioning axis and the orientation of the experimental section mounted on a glass slide. The second stage optimizes the translation, oblique rotation and anisotropic scale of the experimental section image. This stage models the orientation of the tissue block within the microtome and accounts for tissue shrinkage due to histologic processing. The final stage optimizes a nonlinear transform that models deformation of a uniform 3D grid with B-splines. This stage models variability in anatomic structures between individual brains. This multiple stage approach enables exploring the very large transform parameter space in a computationally tractable manner that reflects physical aspects of the problem.
In some cases, the initial search stage does not uniquely identify the location of the experimental section image within the 3D reference image of the CCF v3. Therefore, multiple transform hypotheses can be maintained through all stages of registration. At each stage, transform hypotheses that have a very poor image comparison metric compared to the best hypotheses as well as transform hypotheses that are very similar to higher-ranking transforms, are removed. This approach enables recovery from initially incorrect linear search results.
The image comparison metric serves a critical role in image registration by informing the optimization process how well a particular set of transform parameters aligns the experimental section image with the 3D reference image of the CCF v3. Because the experimental section image was visualized with either fluorescent or thionin Nissl stains and the 3D reference image of the CCF v3 was visualized with two-photon autofluorescence, we selected the Mattes mutual information image comparison metric (Mattes et al., 2001; 2003). We modified the mutual information metric to apply different weights to different parts of the experimental section image preferentially to enforce a better anatomic match. We found that using gradient-based weights of the experimental section image improved the alignment to strong features in the mouse brain, such as white matter tracts.
2. 2 |. Implementation of atlas-based segmentation in NeuroInfo
The transform recovered via image registration enables atlas-based segmentation: the transfer of information from the CCF v3 to the experimental section image or from the experimental section image into the CCF v3, assigning image pixels to brain regions in the reference atlas. The 3D reference image and 3D structural annotations of the CCF v3 and the ontology of the Allen Reference Atlas (thereafter, “ARA ontology”) (Allen Institute for Brain Science, 2017) enable sophisticated queries that provide anatomic context to experimental section images. We developed the following capabilities enabled by section-to-atlas registration:
Creation of an atlas image that matches the experimental section image–
Image resampling is used to extract 2D atlas images (oblique angle slices) from the 3D reference image of the CCF v3 (Zitová and Flusser, 2003). The coordinates of the desired resampled atlas image are created by duplicating the physical coordinate information from the experimental section image. The forward transform recovered by registration transforms each point in the resampled atlas image into the 3D reference image space of the CCF v3, and linear interpolation of the 3D reference image of the CCF v3 determines the intensity of each pixel in the resampled atlas image. This enables extracting from the 3D reference image of the CCF v3 the best fitting oblique angle slice that corresponds to the experimental section and provides a visualization of the registration accuracy (Fig. 1b).
Identification of regions of interest in the experimental section–
The investigator identifies a point in the experimental section image using the computer mouse. The forward transform recovered by section-to-atlas registration transforms the point into the coordinate space of the 3D reference image of the CCF v3. The anatomic identity of this point is then retrieved from the 3D annotations of the CCF v3, and the ARA ontology provides the anatomic region name and its relationship to other anatomic regions. This region name is updated on the computer screen in real time (Fig. 1c-d), which enables full exploration of the entire experimental section’s anatomy (Fig. 3). Figure 4 demonstrates that the approach also works when the sectioning angle deviates from canonical planes presented in conventional mouse brain atlases.
Determination of the boundaries of a region in the experimental section image–
The investigator uses the computer mouse to select a region of interest in the experimental section image or queries a region from the ARA ontology. An annotation image that matches the experimental section image is created using the same resampling process described above, but using nearest-neighbor interpolation of the 3D annotations of the CCF v3 in the final step. The ARA ontology provides the unique identification number of the region of interest and all contained sub-regions. This information is used to build a mask for the region of interest within the resampled annotation image. The connected components algorithm identifies and traces the boundaries of each distinct region in the mask (Shapiro and Stockman, 2001) (Fig. 1d). The result is that the anatomic delineations become part of the data that can be used during analysis of the experimental section.
2. 3 |. Visualization of the CCF v3 in 3D
For visualization purposes, we built smooth anatomic 3D models for each region listed in the ARA ontology based on the 3D annotation image of the CCF v3. This was achieved using the following procedure. For each anatomic region, a 3D image mask was created from the 3D annotation image of the CCF v3 using the union of the anatomic region and all contained sub-regions determined from the ontology. A 3D triangular mesh was created using the marching cubes algorithm (Lorensen and Cline, 1987). The mesh was smoothed and decimated to reduce the number of triangles in the model (c.f. Garland and Heckbert, 1997). The mesh models were saved to disk and later recalled for 3D visualization. The 3D reference image of the CCF v3 is visualized using GPU-accelerated volume rendering with maximum intensity projection (e.g. Fig. 1e). Selected anatomic regions are visualized by loading the constructed 3D meshes and displaying them within the volume rendered 3D reference image of the CCF v3 (e.g. Fig. 1g). The location of the sectioning plane of the experimental section image is visualized by transforming a planar region that encloses the experimental image into the 3D reference image’s space using the linear components of the section-to-atlas registration transform (e.g. Fig. 1e-h).
2. 4 |. Validation of NeuroInfo
A total of 461 coronal, 80 μm-thick frozen serial sections from 78 adult male FVB/N/CD1 mouse brains (from the GENSAT project) were stained with a fluorescence Nissl stain (Neurotrace Blue; Invitrogen, Carlsbad, CA, USA) and imaged using a AxioImager Z1 microscope (Carl Zeiss Microscopy, Jena, Germany) with a 10× objective (N.A. = 0.30, a motorized stage (Ludl Electronic Products, Hawthorne, NY, USA), and a Hamamatsu Orca Flash 4 camera (Hamamatsu City; Japan) controlled by Neurolucida software (MBF Bioscience, Williston, VT, USA) to generate 2D virtual slides (thereafter: “GENSAT section images”). Each GENSAT section image contained between 80 to 200 tiles. Representative examples are shown in Fig. 2a-c.
Another set of 2D virtual slides was generated from 691 coronal 60 μm-thick serial sections of 21 brains from adult male and female C57/BL6 mice that were provided by the CHDI Foundation (New York, NY) in the framework of a research project commissioned to MBF Bioscience (thereafter: “CHDI section images”). These sections were stained with thionin and imaged with a slide scanner equipped with a 20× objective (N.A. = 0.75; BLiSS-200, MBF Bioscience, Williston, VT, USA). Again, each CHDI section image contained between 80 to 200 tiles. Representative examples are shown in Fig. 2d-f.
During pilot experiments, we determined that the registration approach depended on finding a good initial position of the experimental section image within the 3D reference image of the CCF v3. Specifically, a linear search along the sectioning axis worked well when dealing with whole experimental section images of good quality (such as those shown in Fig. 2), but it required manual initialization when dealing with images of partial or damaged experimental sections. Sections were manually inspected for histology artifacts such as large tears, folds and missing pieces. Sections with significant artifacts were omitted. A random subset of 30 GENSAT section images and 30 CHDI section images were then selected for testing. These 60 experimental section images from 12 different mouse brains were automatically registered by means of nonlinear registration with the 3D reference image of the CCF v3 using NeuroInfo. We recorded the X- and Y-scale components of the transform, Mattes mutual information metric (Mattes et al., 2001; 2003), and processing time during registration.
Then, nine GENSAT section images and six CHDI section images were randomly selected, and 14 regions were automatically identified and delineated using NeuroInfo. These 14 regions represent different cortical and subcortical regions in the rostral, caudal, middle, lateral, dorsal and ventral parts of the mouse brain (c.f. Fig.2a-f). If a region did not exist in an experimental section image, NeuroInfo did not draw any delineation upon region selection.
The same regions were manually delineated by a technician on the 15 selected experimental section images using the Stereo Investigator software (MBF Bioscience), referencing the 3D reference image of the CCF v3 and the ARA ontology. Then, the 15 experimental section images overlain with both the manual delineations (thereafter: “Dms”) and the automatic delineations (thereafter: “Das”) were checked by neuroanatomy experts (CRG, PRH), who annotated problematic or inappropriate Dms and Das. Inappropriate Dms were corrected.
After correction of inappropriate Dms, differences between Das and the corresponding Dms were quantified by means of the following metrics: 1) overlap via the Dice coefficient, calculated as
(Dice, 1945; Fischl et al., 2002; Klein et al., 2009); 2) the distance between the centroids of Da and Dm; 3) the relative number of Das in which the centroid of Da was found both within Da and Dm; 4) the difference between the areas of Da and Dm, calculated as
(Fischl et al., 2002; Klein et al., 2009); 5) the percentage of area-Dm that was covered by area-Da, calculated as
and 6) the amount of area-Da outside area-Dm expressed as a percentage of area-Dm, calculated as
2. 5 |. Statistical analysis
Mean and standard deviation of the X- and Y-scale components of the transform, the values obtained by calculating the Mattes mutual information metric and the processing time were separately calculated for the 30 GENSAT section images and the 30 CHDI section images. The Shapiro-Wilk normality test was used to determine whether the distributions of these data were consistent with a Gaussian distribution. Differences between groups were tested with two-way ANOVA (X- and Y-scale components of the transform) and the Mann Whitney test (values obtained by calculating the Mattes mutual information metric and processing time). In all analyses, an effect was considered statistically significant if its associated p value was smaller than 0.05. All calculations were performed using GraphPad Prism (version 5.00 for Windows, GraphPad Software, San Diego, CA, USA).
3.|. RESULTS
3. 1 |. Major components of NeuroInfo and their functionality
The Graphical User Interface (GUI) of NeuroInfo and its Visualization and Navigation Tool is shown in Figure 5 (the details in parentheses refer to the specific example shown in this figure). In addition to the navigation bar (“a” in Fig. 5) and toolbar (“b” in Fig. 5) for basic operations (e.g., opening images, saving and loading registration results, etc.) the major components of section registration are a Macro View window (“c” in Fig. 5) which shows an experimental section image under investigation (here a GENSAT section image), an Ontology window (“d” in Fig. 5) that displays anatomical structures in a hierarchical tree and allows users to select regions for automatic delineation (here the primary somatosensory area, barrel field [SSp-bfd]), a zoomable, full-resolution Main Image window (“e” in Fig. 5) displaying a high-power view of the selected region(s) (here SSp-bfd) in the experimental section image overlain by the region’s delineation automatically generated by NeuroInfo’s Registration Match Finding and Identification Tool (conceptually described in Figs 1 and 3), a Reference Slice window (“f” in Fig. 5) showing the oblique angle slice through the 3D reference image of the CCF v3 that best fits the experimental section image, with semitransparent overlay of the selected region(s) (here the SSp-bfd), a 3D Reference Image window (“g” in Fig. 5) showing the best fitting oblique angle slice (red line) through the 3D reference image of the CCF v3 (seen from the side), displaying the region(s) selected in the Ontology window (here SSp-bfd), and a Section Registration window (“h” in Fig. 5) that allows the user to control section registration using automated linear or nonlinear registration, or by fine-tuning section registration by adjusting translation, rotation and scaling. The NeuroInfo Atlas Calibration Tool (Fig. 6) enables calibration of the X and Y scaling to correct for tissue compression caused by histologic processing, sectioning and mounting. For doing so the user selects 1) the imaging modality (fluorescence or brightfield), and in the case of fluorescence imaging, the fluorescence registration channel (Fig. 6a), 2) the anatomic sectioning plane (Fig. 6b), 3) the sectioning direction (in the example shown in Fig. 6c ‘Rostral-most cut first’ was selected, indicated by the blue frame around the panel on the right), and 4) what is at the bottom of the experimental section image (in the example shown in Fig. 6d ‘Ventral’ was selected, indicated by the blue frame around the panel on the top left). The Atlas Calibration Tool then guides the user through the interactive registration of a single experimental section image to determine the relative scaling between the experimental section image and the 3D reference image of the CCF v3. Our validation experiments determined that this laboratory-specific scaling calibration was critical to enable successful automatic registration of images of novel experimental sections that were processed in the same manner (details are provided in Fig. 7). Once completed, the scaling calibration is used for subsequent section registration operations.
3. 2 |. Registration of experimental section images with the 3D reference image of the CCF v3
In order to match the 60 selected experimental section images as closely as possible with the 3D reference image of the CCF v3 it was necessary to adjust the scaling in X on average by 1.09 (GENSAT section images) and 1.12 (CHDI section images), and in Y on average by 1.21 (GENSAT section images) and 1.27 (CHDI section images). Both the between-groups differences (GENSAT section images vs. CHDI section images) and the within-groups differences (X-scaling vs. Y-scaling) were statistically significant (two-way ANOVA; pLab < 0.001; pDirection < 0.001) (Fig. 7a).
The mean value obtained by calculating the Mattes mutual information 001metric was 0.37 (GENSAT section images) and 0.16 (CHDI section images). This difference was statistically significant (Mann Whitney test; p < 0.001) (Fig. 7b).
The average processing time for registering the images of the experimental sections was 82 s (GENSAT section images) and 72 s (CHDI section images), with no statistically significant difference between the groups (Mann Whitney test; p = 0.819) (Fig. 7c).
3. 3 |. Identification of regions of interest in experimental section images
For the regions caudoputamen (CP), dentate gyrus (DG), primary somatosensory area (barrel field) (SSp-bfd), substantia nigra (reticular part) (SNr), anterior cingulate area (ACA), periaqueductal gray (PAG) and superior colliculus (sensory related) (SCs) (green frames in Fig. 8a; green bars in Fig. 8b-h) excellent alignments between Da and Dm were obtained, indicated by mean Dice coefficients >0.7 (Fig. 8b), average distances between the centroids of Da and Dm <200 μm (Fig. 8c), relative numbers between 83% and 100% of Das in which the centroid of Dm was found within Da (Fig. 8d), absolute mean differences between the areas of Da and Dm <0.1 (except SCs) (Fig. 8e), mean TP values >75% (except SCs) (Fig. 8f) and mean FP values <50% (Fig. 8g). Figure 8h demonstrates that among these regions were the largest ones that were investigated. Furthermore, Figure 8a shows (with the exception of SNr) that these regions are located in the dorsal parts of the mouse brain.
For the regions basolateral amygdala nucleus (BLA), substantia innominata (SI) and cortical amygdalar area (COA) (yellow frames in Fig. 8a; yellow bars in Fig. 8b-h) the alignments between Da and Dm were less precise, indicated by mean Dice coefficients between 0.5 and 0.7 (Fig. 8b). This was supported by all other metrics that returned results that were not as good as in case of the regions CP, DG, SSp-bfd, SNr, ACA, PAG and SCs (Fig. 8c-g). Figure 8h demonstrates that the regions BLA, SI and COA were among the smallest ones that were investigated. Furthermore, Figure 8a shows that these regions are located in the ventrolateral parts of the mouse brain.
For the regions pontine gray (PG), anterior commissure (olfactory limb; aco), lateral ventricle (VL) and third ventricle (V3) (red frames in Fig. 8a; red bars in Fig. 8b-h) the alignments between Da and Dm were poor, indicated by mean Dice coefficients <0.5 (Fig. 8b). The results obtained with the other metrics (except FP) were even poorer than those obtained for the regions BLA, SI and COA (Fig. 8c-g). Figure 8h demonstrates that the regions PG and aco were the smallest ones that were investigated. Figure 8a shows that these regions are located in the middle parts of the mouse brain. Furthermore, aco, VL and V3 do not represent regions that contain nerve cells. There was no significant difference between the results obtained for the GENSAT section images and the CHDI section images (Fig. 9).
4.|. DISCUSSION
The results of the validation experiments performed on 60 experimental coronal sections from 12 different mouse brains that were generated in two different laboratories and imaged with respectively fluorescence or brightfield microscopy can be summarized as follows: 1) NeuroInfo performed remarkably well in automatically delineating regions that are large and/or located in the dorsal parts of mouse brains (mean Dice coefficients >0.7), returned less precise delineations of regions that are small and/or are located in the ventrolateral parts of mouse brains (mean Dice coefficients between 0.5 and 0.7), and must be further improved for the delineation of regions that are respectively either small and located in the middle parts of mouse brains or do not contain nerve cells (mean Dice coefficients <0.5). As an example, validation results for the paraventricular nucleus of the thalamus—a small irregularly shaped structure in the midbrain—demonstrated a mean dice coefficient of 0.65, mean area overlap of 65% and mean distance between the segmented region centroids of 78 μm, which is less than 4 voxels at the resolution of the CCF v3 used for registration. Research on improving atlas-based segmentation of all anatomic regions is currently ongoing in our laboratory by implementing in NeuroInfo a novel image registration approach that uses model information from the 3D annotation image of the CCF v3 to influence global and local components of the section-to-atlas transform.
2) To our surprise, NeuroInfo was not able to correctly identify the olfactory limb of the anterior commissure with the metrics used in this study, despite the fact that it can be found easily by eye. One potential reason is that alignment information in other parts of the experimental section image dominated (e.g. strong cortical or subcortical features), compromising the local alignment of the anterior commissure. The model-based global-local registration approach currently under research is expected to improve this situation. 3) The performance of NeuroInfo in automatically delineating regions in sections of mouse brains did not depend on whether the experimental sections were imaged with fluorescence or brightfield microscopy. 4) To match the experimental section images of mouse brains provided by two different laboratories with the 3D reference image of the CCF v3 optimally, it was necessary to adjust the scaling in Y to a significantly greater extent than the scaling in X. Thus, the investigated mouse brains appeared ‘flattened’ compared to the 3D reference image of the CCF v3. The reason for this phenomenon is unknown, but may be due to cutting blade pressure on the brains along the Y axis or differences in histologic processing techniques. 5) The average processing time of registering an experimental section image of a mouse brain with the 3D reference image of the CCF v3 was 77 seconds. After registration, automatic delineation of any region contained in the experimental section image is obtained in real-time when selecting a certain region in the Ontology window of NeuroInfo’s GUI (“d” in Fig. 5). In contrast, neuroanatomists would need between several seconds (for a single region such as the caudoputamen) and a full day (e.g., for all 147 distinct regions contained in the experimental section shown in Fig. 1a) to complete the same task (see Fig. 3).
This novel automatic navigation system may fundamentally change the way researchers investigate sections of mouse brains under a light microscope, similar to the way surgical navigation systems have influenced neurosurgery over the past three decades (Mezger, Jendrewski and Bartels, 2013). Specifically, the approach presented here enhances the validity, reliability and reproducibility of studies that depend on the exact identification of distinct regions in the mouse brain, enables navigation through the complex anatomy of the mouse brain without requiring prior in-depth training in neuroanatomy, offers investigators the potential to analyze many more brain regions than is currently possible, and provides investigators with the means to directly correlate their experimental specimens with a number of reference tools such as GENSAT or the Allen Mouse Brain Connectivity Atlas (Oh et al., 2014). Likewise, the same principle enables an investigator to collate data from multiple subjects across multiple experiments by transforming measurements into a common reference space. This is a particularly valuable feature, given the emphasis on use of shared reference frameworks for collaborative initiatives such as the NIH BRAIN initiative (Insel et al., 2013).
The novel automatic navigation system presented here is currently only tested for sections from mouse brains counterstained with fluorescent or thionin Nissl stains. Future work will extend this technology to accommodate other counterstains (such as DAPI or immunohistochemical markers). Ultimately, this technology is not limited to mouse, and can be applied to histologic sections from other species with sufficient reference data available, such as rat (using the Waxholm space; Papp et al., 2014), non-human primates (Reveley et al., 2017), and ultimately humans for which the Allen Human Brain Atlas (Hawrylycz et al., 2012; Ding et al., 2016) already provides an on-line high-resolution annotated reference map.
ACKNOWLEDGEMENT
We would like to thank Dr. Amy Bernard (Allen Institute for Brain Science, Seattle, WA, USA) for facilitating the licensing of the Allen Mouse Brain CCF v3 data, and Dr. David Howland (CHDI Foundation/CHDI Management, Princeton, NJ, USA) for providing the opportunity to use the CHDI sections in this study. We also thank Alissa Wilson and Julie McMullen, Isabelle Angstman, Maddie Turneau and Aidan Sullivan (MBF Bioscience) for skillful technical assistance.
Funding information
Grant sponsor: National Institute of Mental Health; Grant number: R44 MH108053 to MBF Bioscience.
LIST OF ABBREVIATIONS (*, from the ontology of the CCF-v3)
- 2D
two-dimensional
- 3D
three-dimensional
- AAV
adeno-associated virus
- ACA*
anterior cingulate area
- aco*
anterior commissure (olfactory limb)
- ARA
Allen Reference Atlas
- BLA*
basolateral amygdala nucleus
- COA*
cortical amygdalar area
- CCF-v3
third version of the Allen Mouse Brain Common Coordinate Framework
- CP*
caudoputamen
- DG*
dentate gyrus
- Da
automatic delineation
- Dm
manual delineation
- GENSAT
Gene Expression Nervous System Atlas
- GPS
global positioning system
- GUI
graphical user interface
- ITK
National Library of Medicine Insight Segmentation and Registration toolkit
- N.A.
numerical aperture
- NIH
National Institutes of Health
- PAG*
periaqueductal gray
- PG*
pontine gray
- SCs*
superior colliculus (sensory related)
- SI*
substantia innominata
- SNr*
substantia nigra (reticular part)
- SSp-bfd*
primary somatosensory area (barrel field)
- VL*
lateral ventricle
- V3*
third ventricle.
Footnotes
CONFLICT OF INTEREST
SJT, BSE and NOC are employees of MBF Bioscience (Williston, VT, USA). CS has served as paid consultant for MBF Bioscience. JRG is the president of MBF Bioscience. QW, LN, BMH, CRG and PRH declare no conflict of interest. The Allen Institute has non-exclusively licensed rights to the CCF to MBF Bioscience to use in NeuroInfo. Due to intellectual property right restrictions, we cannot provide the NeuroInfo source code or its documentation. The NeuroInfo software is commercially available.
REFERENCES
- Allen Institute for Brain Science (2017). Technical white paper: Allen Mouse Common Coordinate Framework and Reference Atlas Available at http://help.brain-map.org/display/mouseconnectivity/Documentation?preview=/2818171/10813534/Mouse_Common_Coordinate_Framework.pdf.
- Avants BB, Tustison NJ, Stauffer M, Song G, Wu B, & Gee JC (2014). The Insight ToolKit image registration framework. Frontiers in Neuroinformatics, 8, 44. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Brown LG (1992). A survey of image registration techniques. ACM Computing Surveys, 24(4), 325–376. [Google Scholar]
- Dice LR (1945). Measures of the amount of ecologic association between species. Ecology, 26(3), 297–302. [Google Scholar]
- Ding S-L, Royall JJ, Sunkin SM, Ng L, Facer BAC, Lesnar P, Guillozet-Bongaarts A, McMurray B, Szafer A, Dolbeare TA, Stevens A, Tirrell L, Benner T, Caldejon S, Dalley RA, Dee N, Lau C, Nyhus J, Reding M, Riley ZL, Sandman D, Shen E, van der Kouwe A, Varjabedian A, Wright M, Zoller L, Dang C, Knowles JA, Koch C, Phillips JW, Sestan N, Wohnoutka P, Zielke HR, Hohmann JG, Jones AR, Bernard A, Hawrylycz MJ, Hof PR, Fischl B, & Lein ES (2016). Comprehensive cellular-resolution atlas of the adult human brain. The Journal of Comparative Neurology, 524(16), 3127–3481. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fischl B, Salat DH, Busa E, Albert M, Dieterich M, Haselgrove C, van der Kouwe A, Killiany R, Kennedy D, Klaveness S, Montillo A, Makris N, Rosen B, & Dale AM (2002). Whole brain segmentation: automated labeling of neuroanatomical structures in the human brain. Neuron, 33(3), 341–355. [DOI] [PubMed] [Google Scholar]
- Garland M, & Heckbert PS (1997). Surface simplification using quadric error metrics. In: Proceedings of the 24th annual conference on computer graphics and interactive techniques - SIGGRAPH ‘97. New York, NY: ACM Press/Addison-Wesley Publishing Co, pp. 209–216. [Google Scholar]
- Gerfen CR, Paletzki R, & Heintz N (2013). Gensat bac cre-recombinase driver lines to study the functional organization of cerebral cortical and basal ganglia circuits. Neuron, 80(6), 1368–1383. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hawrylycz MJ, Lein ES, Guillozet-Bongaarts AL, Shen EH, Ng L, Miller JA, van de Lagemaat LN, Smith KA, Ebbert A, Riley ZL, Abajian C, Beckmann CF, Bernard A, Bertagnolli D, Boe AF, Cartagena PM, Chakravarty MM, Chapin M, Chong J, Dalley RA, David Daly B, Dang C, Datta S, Dee N, Dolbeare TA, Faber V, Feng D, Fowler DR, Goldy J, Gregor BW, Haradon Z, Haynor DR, Hohmann JG, Horvath S, Howard RE, Jeromin A, Jochim JM, Kinnunen M, Lau C, Lazarz ET, Lee C, Lemon TA, Li L, Li Y, Morris JA, Overly CC, Parker PD, Parry SE, Reding M, Royall JJ, Schulkin J, Sequeira PA, Slaughterbeck CR, Smith SC, Sodt AJ, Sunkin SM, Swanson BE, Vawter MP, Williams D, Wohnoutka P, Zielke HR, Geschwind DH, Hof PR, Smith SM, Koch C, Grant SGN, & Jones AR (2012). An anatomically comprehensive atlas of the adult human brain transcriptome. Nature, 489(7416), 391–399. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Insel TR, Landis SC, & Collins FS (2013). The NIH Brain initiative. Science, 340(6133), 687–688. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Klein A, Andersson J, Ardekani BA, Ashburner J, Avants B, Chiang MC, Christensen GE, Collins DL, Gee J, Hellier P, Song JH, Jenkinson M, Lepage C, Rueckert D, Thompson P, Vercauteren T, Woods RP, Mann JJ, & Parsey RV (2009). Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. NeuroImage, 46(3), 786–802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Koroshetz W, Gordon J, Adams A, Beckel-Mitchener A, Churchill J, Farber G, Freund M, Gnadt J, Hsu NS., Langhals N, Lisanby S, Liu G, Peng GCY, Ramos K, Steinmetz M, Talley E, & White S (2018). The state of the NIH Brain initiative. The Journal of Neuroscience, 38(29), 6427–6438. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lorensen WE, & Cline HE (1987). Marching cubes: A high resolution 3D surface construction algorithm. ACM SIGGRAPH Computer Graphics, 21(4), 163–169. [Google Scholar]
- Mattes D, Haynor DR, Vesselle H, Lewellyn TK, & Eubank W (2001). Nonrigid multimodality image registration. In Sonka M & Hanson KM (Eds.) (pp. 1609–1620). Presented at the Medical Imaging 2001, San Diego, CA, USA. [Google Scholar]
- Mattes D, Haynor DR, Vesselle H, Lewllen TK, & Eubank W (2003). PET-CT image registration in the chest using free-form deformations. IEEE Transactions on Medical Imaging, 22, 120–128. [DOI] [PubMed] [Google Scholar]
- Mezger U, Jendrewski C, & Bartels M (2013). Navigation in surgery. Langenbeck’s Archives of Surgery, 398(4), 501–514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Oh SW, Harris JA, Ng L, Winslow B, Cain N, Mihalas S, Wang Q, Lau C, Kuan L, Henry AM, Mortrud MT, Ouellette B, Nguyen TN, Sorensen SA, Slaughterbeck CR, Wakeman W, Li Y, Feng D, Ho A, Nicholas E, Hirokawa KE, Bohn P, Joines KM, Peng H, Hawrylycz MJ, Phillips JW, Hohmann JG, Wohnoutka P, Gerfen CR, Koch C, Bernard A, Dang C, Jones AR, & Zeng H (2014). A mesoscale connectome of the mouse brain. Nature, 508(7495), 207–214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Papp EA, Leergaard TB, Calabrese E, Johnson GA, & Bjaalie JG (2014). Waxholm Space atlas of the Sprague Dawley rat brain. NeuroImage, 97, 374–386. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pastrana E (2014). A genetic handle on brain circuits: Neuroscience. Nature Methods, 11(2), 128–128. [DOI] [PubMed] [Google Scholar]
- Reveley C, Gruslys A, Ye FQ, Glen D, Samaha J, Russ EB, Saad Z, Seth KA, Leopold DA, & Saleem KS (2017). Three-dimensional digital template atlas of the macaque brain. Cerebral Cortex, 27(9), 4463–4477. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shapiro L & Stockman GC (2001). Computer vision London: Pearson. [Google Scholar]
- Zitová B & Flusser J (2003). Image registration methods: a survey. Image and Vision Computing, 21, 977–1000. [Google Scholar]