SUMMARY
Tissue clearing methods enable the imaging of biological specimens without sectioning. However, reliable and scalable analysis of large imaging data in 3D remains a challenge. Here, we developed a deep learning-based framework to quantify and analyze the brain vasculature, named Vessel Segmentation & Analysis Pipeline (VesSAP). Our pipeline uses a convolutional neural network (CNN) with a transfer learning approach for segmentation and achieves human-level accuracy. Using VesSAP we analyzed vascular features of whole C57BL/6J, CD1 and BALB/c mouse brains at the micrometer scale after registering them to the Allen mouse brain atlas. We reported evidence of secondary intracranial collateral vascularization in CD1 mice and found reduced vascularization of the brainstem compared to the cerebrum. Thus, VesSAP enables unbiased and scalable quantifications of the angioarchitecture of cleared mouse brains and yields biological insights into the vascular function of the brain.
INTRODUCTION
Changes in cerebrovascular structures are key indicators for a large number of diseases affecting the brain. Primary angiopathies, vascular risk factors (e.g., diabetes), traumatic brain injury, vascular occlusion and stroke all affect the function of the brain’s vascular network1–3. The hallmarks of Alzheimer’s disease including tauopathy and amyloidopathy can also lead to aberrant remodeling of blood vessels1,4, allowing capillary rarefaction to be used as a marker for vascular damages5. Therefore, quantitative analysis of the entire brain vasculature is pivotal to developing a better understanding of brain function in physiological and pathological states. However, quantifying micrometer scale changes in the cerebrovascular network of brains has been difficult for two main reasons.
First, labeling and imaging of the complete mouse brain vasculature down to the smallest blood vessels has not yet been achieved. Neither magnetic resonance imaging (MRI), micro-computed tomography (MicroCT), nor optical coherence tomography have sufficient resolution to capture capillaries in bulk tissue6–8. Fluorescent microscopy provides a higher resolution, but can typically only be applied to tissue sections up to 200 μm thickness9. Recent advances in tissue clearing could overcome this problem10, but so far there has been no systematic description of all vessels of all sizes in an entire brain in three dimensions (3D).
The second challenge relates to the automated analysis of large 3D imaging data with substantial variance in signal intensity and signal-to-noise ratio at different depths. Simple intensity and shape-based filtering approaches such as Frangi’s vesselness filters, or more advanced image processing methods with local spatial adaptation cannot reliably differentiate vessels from a background in whole brain scans11,12. Finally, imaging of a whole brain vascular network at capillary resolution results in datasets of terabyte size. Established image processing methods do not scale well to terabyte-sized image volumes as they don’t generalize well to large images and they require intensive manual fine-tuning13–15.
Here, we present VesSAP (Vessel Segmentation & Analysis Pipeline), a deep learning-based method for the automated analysis of the entire mouse brain vasculature, overcoming the above limitations. VesSAP encompasses 3 major steps: 1) staining, clearing and imaging of the mouse brain vasculature down to the capillary level by two different dyes: wheat germ agglutinin (WGA) and Evans blue (EB), 2) automatically segmenting and tracing the whole brain vasculature data via convolutional neural networks (CNN) and 3) extracting vascular features for hundreds of brain regions after registering the data to the Allen brain atlas (Fig. 1). Our deep learning-based approach for network extraction in cleared tissue is robust, despite variations in signal intensities and structures, outperforms previous filter-based methods and reaches the quality of segmentation achieved by human annotators. We applied VesSAP to the three commonly used mouse strains C57BL/6J, CD1 and BALB/c.
RESULTS
Vascular staining, DISCO clearing, and imaging
To reliably stain the entire vasculature, we used dyes WGA and EB dyes, which can be visualized in different fluorescence channels. We injected the EB dye into the living mice 12 hours before WGA perfusion, allowing its long-term circulation to mark vessels under physiological conditions16, while we perfused the WGA dye during fixation. We then performed 3DISCO clearing17 and light-sheet microscopy imaging of whole mouse brains (Fig. 2a–c, Supplementary Fig. 1, 2). WGA highlights microvessels, and EB predominantly stains major blood vessels, such as the middle cerebral artery and the circle of Willis (Fig. 2d–i, Supplementary Fig. 3). Merging the signals from both dyes yields a more complete staining of the vasculature than relying on individual dyes only (Fig. 2c,f, Supplementary Video 1). Staining from both dyes is congruent in mid-sized vessels, with signals originating from the vessel wall layer (Fig. 2j–l, Supplementary Fig. 3a–c). Using WGA, we reached a higher signal-to-noise ratio (SNR) for microvessels than bigger vessels. With EB, the SNR for small capillaries was lower but larger vessels reached a high SNR (Supplementary Fig. 4). Integrating the information from both channels allows acquisition of the entire vasculature, and results in an optimized SNR. We also compared fluorescence signal quality of the WGA staining (targeting the complete endothelial glycocalyx lining18) with a conventional vessel antibody (anti-CD31, targeting endothelial cell-cell adhesion19) and found that WGA produced a higher SNR for the blood vessels in general (Supplementary Fig. 5).
Segmentation of the volumetric images
To enable the extraction of quantitative features of the vascular structure, vessels in the acquired brain scans need to be segmented in 3D. Motivated by deep learning-based approaches in biomedical image data analysis20–28, we used a five layer CNN (Fig. 3a) to exploit the complementary signals of both dyes to derive a complete segmentation of the entire brain vasculature.
In the first step, the two input channels (WGA and EB) are concatenated. This yields a matrix in which each voxel is characterized by two features. Then, each convolutional step integrates the information from a voxel’s 3D neighborhood. We use full 3D convolutions20 without further down or up sampling and fewer trainable parameters in comparison to e.g., the 3D U-Net and V-Net29,30 in order to achieve high inference speeds. After the fourth convolution, the information from 50 features per voxel is combined with a convolutional layer with kernel size one and a sigmoidal activation to estimate the likelihood that a given voxel represents a vessel. Subsequent binarization yields the final segmentation. In both training and testing, the images are processed in sub-volumes of 50 × 100 × 100 pixels.
Deep neural networks often require large amounts of annotated data or many iterations of training. Here, we circumvented this requirement with a transfer learning approach31. In short, we first pre-trained the network on a large, synthetically generated vessel-like data set (Supplementary Fig. 6)32 and then refined on a small amount of manually annotated parts of the real brain vessel scans. This approach reduced the training iterations on manually annotated training data.
To assess the quality of the segmentation, we compared the VesSAP CNN prediction with manually labeled ground truth and alternative computational approaches (Table 1). We report voxel-wise segmentation metrics, namely accuracy, the F1-score33, Jaccard coefficient, and cl-F1, which weight the centerlines and volumes of the vessels (detailed in Methods). Compared to the ground truth, our network achieved an accuracy of 0.94 ± 0.01 and an F1-score of 0.84 ± 0.05 (for additional scores see Table 1, all values are mean ± SD). As controls, we implemented alternative state-of-the-art deep learning and classical methods. Our network outperforms classical Frangi filters11 (accuracy: 0.85 ± 0.03; F1-Score: 0.47 ± 0.18), as well as recent methods based on local spatial context via Markov random fields13,34 (accuracy: 0.85 ± 0.03; F1-score: 0.48 ± 0.04). VesSAP achieved similar performance compared to 3D U-Net and V-Net architectures, which require substantially more trainable parameters (3D U-Net: accuracy: 0.95 ± 0.01; F1-score: 0.85 ± 0.03 and V-Net: accuracy: 0.95 ± 0.02; F1-score: 0.86 ± 0.07, no statistical difference compared to VesSAP CNN, two sided t-test all, p-values > 0.3). However, VesSAP CNN outperformed the other architectures substantially in terms of speed, being ~20 and ~50 times faster in the feedforward path than V-Net and 3D U-Net, respectively. This is particularly important for our large datasets (hundreds of gigabytes). For example, VesSAP CNN segments a single brain in 4 hours, while V-Net and 3D U-Net require 3.3 days and 8 days, respectively. The superior speed of the VesSAP CNN is due to the substantially fewer trainable parameters in its architecture (e.g., our implementation of the 3D U-Net has ~178 million parameters, VesSAP CNN has ~0.059 million parameters) (Table 1). Next, we compared the segmentation accuracy of our network with human annotations. A total of 4 human experts independently annotated two volumes. We found that inter-annotator accuracy and F1-score of experts were comparable to the predicted segmentation of our network (human annotators: accuracy: 0.92 ± 0.02 and F1-score: 0.81 ± 0.06, Fig. 3b). Importantly, we extrapolate that human annotators would need more than a year to process a whole brain instead of the 4 hours required by our approach. Moreover, we observed differences in the human segmentations due to annotator bias. Thus, VesSAP CNN can segment the whole brain vasculature consistently at human-level accuracy with a substantially higher speed than currently available methods, enabling high throughput for large-scale analysis.
Table 1:
Segmentation model | CI-F1 | Accuracy | F1-Score | Jaccard | Parameters | Speed |
---|---|---|---|---|---|---|
VesSAP CNN | 0.93 ± 0.02 | 0.94 ± 0.01 | 0.84 ± 0.05 | 0.84 ± 0.04 | 0.0587 m | 1.19 s |
VesSAP CNN, trained from scratch | 0.93 ± 0.02 | 0.94 ± 0.01 | 0.85 ± 0.04 | 0.85 ± 0.04 | 0.0587 m | 1.19 s |
VesSAP CNN, synthetic training data | 0.87 ± 0.02 | 0.90 ± 0.05 | 0.72 ± 0.07 | 0.70 ± 0.05 | 0.0587 m | 1.19 s |
3D U-Net | 0.93 ± 0.02 | 0.95 ± 0.01 | 0.85 ± 0.03 | 0.85 ± 0.03 | 178.4537 m | 61.22 s |
V-Net | 0.94 ± 0.02 | 0.95 ± 0.02 | 0.86 ± 0.07 | 0.86 ± 0.07 | 88.8556 m | 26.87 s |
Frangi Vesselness | 0.84 ± 0.03 | 0.85 ± 0.03 | 0.47 ± 0.19 | - | - | 117.00 s |
Markov Random Field | 0.86 ± 0.02 | 0.85 ± 0.03 | 0.48 ± 0.04 | - | - | 24.31 s |
We show an example of a brain vasculature that was segmented by VesSAP in 3D (Fig. 3c and Supplementary Video 2, 3). Zooming into a smaller patch reveals that the connectivity of the vascular network was fully maintained (Fig. 3d, Supplementary Video 2). Comparing single slices of the imaging data with the predicted segmentation shows that vessels are accurately segmented regardless of absolute illumination or vessel diameter (Supplementary Fig. 7).
Feature extraction and atlas registration
Vessel lengths, radii and the number of bifurcation points are commonly used to describe the angioarchitecture2. Hence, we used our segmentation to quantify these features as distinct parameters to characterize the mouse brain vasculature (Fig. 4a, Supplementary Video 4). We evaluated local vessel length (length normalized to the size of the brain region of interest), local bifurcation density (sum of occurrences normalized to the size of the brain region of interest) and local vessel radius (average radius along the full length) of the blood vessels in different brain regions.
We reported the vascular features in three ways to enable a comparison with various previous studies that differed in the measures used (Supplementary Fig. 8). More specifically, first, we provided a count of segmented voxels as compared to total voxels within a specific brain region (voxel space). Second, we provided the measurements by calculating the voxel size of our imaging system and accounted for the Euclidean length (microscopic space). Third, we corrected the microscopic measurements to account for the tissue shrinkage caused by the clearing process (anatomical space) (Supplementary Tables 2–10). We calculated this shrinkage rate by measuring the same mouse brain volume using MRI prior to clearing.
Here, we use the anatomical space to report our specific biological findings as it is closest to the physiological state. For the average blood vessel length of the whole brain, we found a value of 545.74 ± 94 mm/mm³ (mean ± SD). Because our method quantifies brain regions separately, we could compare our results to the literature, which mostly reported the quantifications for either specific brain regions or extrapolations to the whole brain from regional quantifications. For example a vascular length of 922 ± 176 mm/mm³ (mean ± SD) was previously reported for cortical regions (size of 508 × 508 × 1500 μm)10. We found a similar vessel length for the same region in the mouse cortex (C57BL/6J) (913 ± 110 mm/mm³), substantiating the accuracy of our method. We performed additional comparisons to other reports (Supplementary Table 11). Moreover, we compared the measurements acquired with our algorithms to manually labeled ground truth data and found a deviation of 8.21% for the centerlines, 13.18% for the number of bifurcation points and 16.33% for the average radius. These deviations are substantially lower than the average deviation among human annotators (Methods).
We quantified and visualized the vessel radius along the entire vascular network (Fig. 4b). After extracting vascular features of the whole brain with VesSAP, we registered the volume to the Allen brain atlas (Supplementary Video 5, 6). This allowed us to map the segmented vasculature and corresponding features topographically to distinct anatomical brain regions (Fig. 4c). Each anatomical region can be further divided into sub-regions, yielding a total of 1238 anatomical structures (619 per hemisphere) for the entire mouse brain (Fig. 4d). This allows the analysis of each denoted brain region and grouping regions into clusters such as left vs. right hemisphere, gray vs. white matter or any hierarchical clusters of the Allen brain atlas ontology. For our subsequent statistical feature analysis, we grouped the labeled structures according to the 71 main anatomical clusters of the current Allen brain atlas ontology. We thus provide the whole mouse brain vascular map with extracted vessel lengths, bifurcation points and radii down to the capillary level.
VesSAP provides a reference map of the whole brain vasculature in mice
Studying the whole brain vasculature in C57BL/6J, CD1 and BALB/c strains (n = 3 mice for each strain), we found that the local vessel length and local bifurcation density differ in the same brain over different regions, while they correlate highly among different mice for the same regions (Fig. 5a,b). Furthermore, the local bifurcation density correlates highly with the local vessel length in most brain regions (Supplementary Fig. 9), and the average vessel radius is evenly distributed in different regions of the same brain (Fig. 5c). In addition, the extracted features show no statistical difference (Cohen’s d, Supplementary Table 12) between the same anatomical cluster among the strains (Supplementary Fig. 9). Finally, microvessels make up the overwhelming proportion of the total vascular composition in all brain regions (Fig. 5d). We visually inspected exemplary brain regions to validate the output of VesSAP. Both VesSAP and visual inspection revealed that the gustatory areas had a higher vascular length per volume compared to the anterodorsal nucleus (Fig. 6a–c). Visual inspection also suggested that the number of capillaries was the primary reason for regional feature variations within the same brain.
Finally, VesSAP offered insights into the neurovascular structure of the different mouse strains in our study. There are direct intracranial vascular anastomoses in the C57BL/6J, CD1 and BALB/c strains (white arrowheads, Fig. 6d–f). The anterior cerebral artery, middle cerebral artery and the posterior cerebral artery are connected at the dorsal visual cortex in CD1 (red arrowheads, Fig. 6d,e) unlike in the BALB/c strains33 (Fig. 6f).
DISCUSSION
VesSAP can generate a reference map of the adult mouse brain vasculature. These maps can potentially be used to model synthetic cerebrovascular networks35. In addition to the metrics we obtain to describe the vasculature, advanced metrics e.g., Strahler values, network connectivity, and bifurcation angles can be extracted using the data generated by VesSAP. Furthermore, the centerlines and bifurcation points can be interpreted as the edges and nodes for building a full vascular network graph, offering a means for studying local and global properties of the cerebrovascular network in the future.
The VeSAP workflow relies on staining of the blood by two different dyes. WGA binds to the glycocalyx of the endothelial lining of the blood vessels36 but may miss some segments of the large vessels18. EB is a dye with a high affinity to serum albumin37, thus, it remains in the large vessels after a short perfusion protocol. In addition, EB labeling is not affected by subsequent DISCO clearing.
Vessels have long and thin tubular shapes. In our images, the radii of capillaries (about 3 μm) are in the range of our voxel size. Therefore, a segmentation, which yields the correct diameter down to a single-pixel resolution poses a challenge as we observed a 16% deviation for the radius. This sub-pixel deviation did not pose a problem for segmenting the whole vasculature network and extracting features, because the vascular network can be defined by its centerlines and bifurcations.
The described segmentation concept is based on a transfer learning approach, where we pretrained the CNN and refined it on a small labeled dataset of 11% of the synthetic dataset and only 0.02% of one cleared brain. We consider this a major advantage compared to training from scratch. Thus, our CNN might generalize well to different types of imaging data (such as microCT angiography) or other curvilinear structures (e.g., neurons), as only a small labeled dataset is needed to adjust our pre-trained network.
Based on our vascular reference map, unknown vascular properties can be discovered and biological models can be confirmed. VesSAP showed a high number of collaterals in albino CD1 mice. Such collaterals between large vessels can substantially alter the outcome of ischemic stroke lesions: the blood deprived brain regions from the occlusion of a large vessel can be compensated by a blood supply from the collateral extensions of other large vessels33,40. Therefore, our VesSAP method can lead to the discovery of previously unknown anatomical details that could be functionally relevant.
In conclusion, VesSAP is a scalable, modular and automated machine learning-based method to analyze complex imaging data from cleared mouse brains. We foresee that our method will accelerate the applications of tissue clearing in particular for studies assessing the brain vasculature.
METHODS
Tissue preparation
Animal experiments were conducted according to institutional guidelines (Klinikum der Universität München/Ludwig Maximilian University of Munich), after approval of the ethical review board of the government of Upper Bavaria (Regierung von Oberbayern, Munich, Germany), and in accordance with the European directive 2010/63/EU for animal research. The animals were housed under a 12/12 hr light/dark cycle. For this study we injected 150 μl (2% V/V% in saline) of Evans blue dye (Sigma-Aldrich, E2129) intraperitoneally into three C57BL/6J, CD1 and BALB/c (Charles River, Strain Codes 027; 482 and 028 respectively) male, 3 months old mice (n=3 per group). After 12 hrs of postinjection time, we anaesthetized the animals with a triple combination of midazolam + medetomidine + fentanyl (i.p.; 1 ml per 100 g body mass for mice; 5 mg, 0.5 mg and 0.05 mg per kg body mass) and opened their chest for transcardial perfusion. The following media was supplied by a peristaltic pump set to deliver 8 ml/min volume: 0.25 mg wheat germ agglutinin conjugated to Alexa 594 dye (ThermoFisher Scientific, W11262) in 150 μl PBS (pH 7.2) and 15 ml PBS 1x and 15 ml 4% PFA. This short perfusion protocol was established based on preliminary experiments, where both WGA and EB staining partially washed out (data not shown) and with the goal to deliver fixative to the brain tissue using the vessels to achieve a homogenous preservation effect41.
After perfusion, the brains were extracted from the neurocranium while severing some of the segments of the circle of Willis, which is an inevitable component of most retrieval processes besides corrosion cast techniques. Next, the samples were incubated into 3DISCO clearing solutions as described17. Briefly, we immersed them in a gradient of tetrahydrofuran (Sigma-Aldrich, 186562): 50 vol%, 70 vol%, 80 vol%, 90 vol%, 100 vol% (in distilled water), and 100 vol% at 25 °C for 12 h at each concentration. At this point we modified the protocol by incubating the samples in tert-Butanol incubation for 12 hrs at 35 °C followed by immersion in dichloromethane (Sigma-Aldrich, 270997) for 12 hrs at room temperature and finally incubation with the refractive index matching solution BABB (benzyl alcohol + benzyl benzoate 1:2; Sigma-Aldrich, 24122 and W213802), for at least 24 hrs at room temperature until transparency was achieved. Each incubation step was carried out on a laboratory shaker.
Imaging of the cleared samples with light-sheet microscopy
We used a 4× objective lens (Olympus XLFLUOR 340) equipped with an immersion corrected dipping cap mounted on a LaVision UltraII microscope coupled to a white light laser module (NKT SuperK Extreme EXW-12) for imaging. The images were taken in 16 bit depth and at a nominal resolution of 1.625 μm / voxel on the XY axes. For 12× imaging we used a LaVision objective (12x NA 0.53 MI PLAN with an immersion corrected dipping cap). The brain structures were visualized by the Alexa 594 (using a 580/25 excitation and a 625/30 emission filter) and Evans blue fluorescent dyes (using a 640/40 excitation and a 690/50 emission filter) in a sequential order. We maximized the signal to noise ratio (SNR) for each dye independently to avoid saturation of differently sized vessels when only a single channel is used. We achieved this by independently optimizing the excitation power so that the strongest signal in the major vessels does not exceed the dynamic range of the camera. In z-dimension we took the sectional images in 3 μm steps using left and right sided illumination. Our measured resolution was 2.83 μm × 2.83 μm × 4.99 μm for X, Y and Z, respectively (Supplementary Fig. 2). To reduce defocus, which derives from the Gaussian shape of the beam, we used a 12 step sequential shifting of the focal position of the light-sheet per plane and side. The thinnest point of the light-sheet was 5 μm.
Imaging of the cleared samples with confocal microscopy
Additionally, the cleared specimens were imaged with an inverted laser-scanning confocal microscope (Zeiss, LSM 880) for further analysis. Before imaging, samples were mounted by placing them onto the glass surface of a 35 mm glass-bottom petri dish (MatTek, P35G-0–14-C) immersed in BABB. A 40x oil-immersion objective lens was used (Zeiss, ECPlan-NeoFluar 40x/1.30 Oil DIC M27, 1.3 NA, WD = 0.21 mm). The images were acquired by the settings for Alexa 594 (using a 561 excitation and a 585–733 emission range) and Evans blue fluorescent dyes (using a 633 excitation and a 638–755 emission range) in a sequential order.
Magnetic resonance imaging
We used a nanoScan PET/MR device (3 Tesla, Mediso Medical Imaging Systems) equipped with a head coil for murine heads to acquire anatomical scans in the T1 modality.
Reconstruction of the datasets from the tiling volumes
We stitched the acquired volumes using TeraStitcher’s automatic global optimization function (v1.10.3). We produced volumetric intensity images of the whole brain considering each channel separately. To improve the alignment to the Allen brain atlas we downscaled our dataset in the XY plane to achieve pseudo-uniform voxel spacing matching the Z plane.
Deep learning network architecture
We rely on a deep 3D convolutional neural network (CNN) for segmentation of our blood vessel dataset. The networks general architecture consists of 5 convolutional layers, 4 with ReLu (rectified linear unit) followed by one convolutional layer with a sigmoid activation (Fig. 3a). The input layer is designed to take n images as an input. In the implemented case, the input to the first layer of the network are n=2 images of the same brain, which have been stained differently (Fig. 3a). To specifically account for the general class imbalance (much more tissue background than vessel signals) in our dataset, and the potential of high false positive rates associated with the class imbalance, we chose the generalized soft-Dice as the loss function to our network. At all levels we used full 3D convolutional kernels (Fig. 3a).
The networks training is driven by an Adam optimizer with a learning rate of 1e−5 and exponential decay rate of 0.9 for the 1st moment and 0.99 for the 2nd moment42. A prediction or segmentation with a trained model takes volumetric images of arbitrary size as input and outputs a probabilistic segmentation map of identical size. To deal with volumes of arbitrary size and extension, we processed them in smaller sub-volumes of 100 ×100 × 50 pixels size. The algorithms have been implemented using the Tensorflow framework and KERAS43. They are trained and tested on two NVIDIA Quadro P5000 GPUs and on machines with 64GB and 512GB RAM respectively.
Transfer learning
Typically, supervised learning tasks in biomedical imaging are aggravated by a scarce availability of labeled training data. Our transfer learning approach aims to circumvent this problem by pre-training our models on synthetically generated data and refining them on a small set of real images44. Specifically, our approach pre-trains the VesSAP CNN on 3D volumes of vascular image data, synthetically generated together with the corresponding training labels, using the approach by Schneider and colleagues45. The pre-training is carried out on a dataset of 20 volumes of a size of 325 × 304 × 600 pixels for 38 epochs. During pre-training we applied a learning rate of 1e−4. Afterwards, the pre-trained model was fine-tuned by retraining on a real microscopic dataset consisting of eleven volumes with a size of 500 × 500 × 50 pixels. The image volumes were manually annotated by a commission including the expert who had previously prepared the samples and operated the microscope. The labels were verified and further refined in consensus by two additional human raters. The data we used in this fine-tuning step amounts to 11 % of the volume of the synthetic datasets and only a fraction of 0.02% of the voxel volume of a single whole brain. For the fine-tuning step, we used a learning rate of 1e−5. The final model was obtained after training on the real dataset for 6 epochs. This is a substantially shorter training compared to training from scratch, where we train the same VesSAP CNN architecture for 72 epochs until we reach the best F1-Score on the validation set. The labeled dataset consists of 17 volumes of 500 × 500 × 50 pixels from 5 mice brains. Three of those brains are from the CD1 and two from the C57BL/6J strain. The volumes are chosen from regions throughout the whole brain, to represent the variability in the vascular dataset, both in terms of vessel shapes and illumination. To ensure independence, volumes of the training set and test/validation set are chosen from independent brains. All datasets include brains from the two strains. Our training dataset consists of eleven volumes, the validation dataset of three volumes and the test dataset of three volumes, too. We cross-test on our test and validation dataset by rotating these. The volumes are processed during training and inference in 25 small sub-volumes of 100 × 100 × 50 pixels.
We observed an average F1-Score of 0.84 ± 0.02 (mean ± SD), an average accuracy of 0.94 ± 0.01 (mean ± SD), and an average Jaccard coefficient of 0.84 ± 0.04 (mean ± SD) on our test datasets (Fig. 3b). We tested for statistical significance among the top three learning methods (VesSAP CNN, V-Net and 3D U-Net) using a paired t-test. We found that the differences in F1-Scores are not statistically significant (all p-values > 0.3, rejecting the hypothesis of different distributions).
Since the F1-Score, accuracy and Jaccard coefficient are all voxel wise volumetric scores and can fall short in evaluating connectedness of components we developed the cl-F1 score. cl-F1 is calculated on the intersection of centerlines and vessel volumes and not on volumes only, like the traditional F1-score is46. To determine this score we first calculate the intersection of the centerline of our prediction with the labeled volume and then calculated the intersection of the labeled volume’s centerline with the predicted volume. Next, we treat the first intersection as recall, as it is susceptible to false negatives, and the second as precision, as it is susceptible to false positives, and input this into the traditional F1-score formulation below:
(I.) |
We report an average cl-F1 score of 0.93 ± 0.02 (mean ± SD) on the test set.
All scores are given as mean and standard deviation. Our model reached the best model selection point on the validation dataset after 6 epochs of training.
Comparison to 3D U-Net and V-Net
To compare our proposed architecture to different segmentation architectures, we implemented the V-Net and 3D U-Net, both more complex CNNs with substantially more trainable parameters, which further include down- and up-sampling. While experiments show that 3D U-Net and V-Net reach marginally higher performance scores, the differences are not statistically significant (two sided t-test, p > 0.3). Their amount of parameters make them a factor of 51 and 23 slower during the inference stage. For the segmentation of one of our large whole brain datasets this translates to a time of 4 hours versus 8 days for 3D U-Net and 3.8 days for V-Net. This difference is also prevalent in the number of trainable parameters. VesSAP CNN has 0.058 million parameters, whereas 3D U-Net consists of more than 178 million and V-Net of more than 88 million. Furthermore, the light VesSAP CNN already reaches human-level performance. We therefore consider the problem of vessel segmentation as solved by the VesSAP CNN for our data. It should be mentioned that the segmentation network is a modular building block of the overall VesSAP pipeline, and can be chosen by each user according to his or her own preferences and, importantly, according to the computational power available.
Pre-processing of segmentation
The pre-processing factors into the overall success of the training and segmentation. The intensity distribution among the brains and among brain regions differs substantially. To account for the intensity distributions, two preprocessing strategies have been applied successively.
- High-cut filter: In this step the intensities x above a certain threshold c, which is defined by a global percentile, where each volume was set to that threshold. Next, they were normalized by f(x).
(II.) - Normalization of intensities: The original intensities were normalized to the range of 0 to 1, where x is the pixel intensity and X are all intensities of the volume.
(III.)
Inter-annotator experiment for the segmentation
To compare VesSAP’s segmentation to a human level annotation we implemented an inter-annotator experiment. For this experiment we determined a gold standard label (ground truth) for two volumes of 500 × 500 × 50 pixels from a commission of three experts, including the expert who imaged our data and is therefore most familiar with the images. Next, we gave the two volumes to 4 other experts to label the complete vasculature. The experts spend multiple hours to label each patch within the ImageJ and ITK-snap environment and were allowed to use their favored approaches to generate what they considered to be the most accurate labeling. Finally, we calculated the accuracy and dice scores for the different annotators, compared to the gold standard and compared them to the scores of our model (Table 1).
Feature extraction
In order to quantify the anatomy of the mouse brain vasculature we extracted descriptive features based on our segmentation. First, we calculate the features in voxel space. Later we registered them to the Allen brain atlas.
As features we extracted the centerlines, the bifurcation points and the radius of the segmented blood vessels. We consider those features to be independent to the elongation of the light-sheet scans and the connectedness of the vessels due to staining, imaging and/or segmentation artefacts.
Our centerline extraction is based on a 3D thinning algorithm47. Before extracting the centerlines we applied two cycles of binary erosion and dilation to remove false negative pixels within the volume of segmented vessels as those would induce false centerlines. Based on the centerlines we extracted bifurcation points. A bifurcation is the branching point on a centerline where a larger vessel splits into two or more small vessels (Fig. 4a). In a network analysis context they are significant as they represent the nodes of a vascular network48. Furthermore, bifurcation points have significance in a biological context. In neurodegenerative diseases, capillaries are known to degenerate49, thereby substantially reducing the number of bifurcation points in an affected brain region compared to a healthy brain. We implemented an algorithm to detect the bifurcation points by taking the centerlines as an input and calculating the surrounding centerline pixels for every point on that centerline to determine whether a point is a centerline. The radius of a blood vessel is a key feature to describe vascular networks. The radius yields information about the flow and hierarchy of the vessel network. The proposed algorithm calculates the Euclidean distance transform for every segmented pixel v to the closest background pixel bclosest. Next, the distance transform matrix was multiplied with the 3D centerline mask equaling the minimum radius of the vessel around the centerline.
(IV.) |
Feature quantification
Here we describe in detail how we calculated the features between the three different spaces:
Voxel space to microscopic space
To quantify our vessel length in SI units instead of voxels we calculated their Euclidean length, which depends on the direction of the connection of skeleton pixels (Supplementary Fig. 9). To calculate the Euclidean length of our centerlines, we carried out a connected component analysis, which transformed each pixel of the skeleton into an element of an undirected weighted graph, where zero weight means no connection and non-zero weights denote the Euclidean distance between two voxels (considering 26 connectivity). Thus, we obtain a large and sparse adjacency matrix. An element-wise summation of such a matrix provides the total Euclidean length of the vascular network along the extracted skeleton.
As measuring connected components is computationally very expensive, we calculated this Euclidean length of the centerlines for twelve representative volumes of 500 × 500 × 50 pixels, and divided by the number of skeleton pixels. We calculate an average Euclidean length εCl of 1.3234 ± 0.0063 (mean ± SD) voxels per centerline element. This corresponds to a length of 3.9701±0.0188 (mean ± SD) μm in cleared tissue. Since the standard deviation of this measurement is low at less than 0.5% of the length, we apply this correction factor to the whole brain centerline measurements. This correction does not apply to the bifurcation points and our radius statistics, as bifurcations are independent of length and the radius extraction returns a Euclidean distance by default. Euclidean length, depending on the direction of the connection of skeleton pixels, the length of a skeleton pixel is different (Supplementary Fig. 9).
Microscopic space to anatomical space
To account for the tissue shrinkage (Supplementary Fig. 9), which is inherent to DISCO clearing, we carried out an experiment to measure the degree of shrinkage. Before clearing, we imaged three live BALB/c mice brains with magnetic resonance imaging and calculated their brain’s average volumes, through precise manual segmentation by an expert. Next, we cleared three BALB/c brains, processed them using VesSAP and measured the total brain volumes using the atlas alignment. We report an average volume of 423.84 ± 2.04 mm3 for the live mice and 255.62 ± 6.57 mm3 for the cleared tissue. This accumulates to a total volume shrinkage of 39.69 %. We applied this as a correction factor for the volumetric information (e.g. brain regions).
Similar to previous studies, the shrinkage is uniform in all three dimensions. This is important, when considering shrinkage in one dimension as needed to account for the shrinkage in the centerlines and the radius. The one dimensional correction factor КL then corresponds to the cube root of the volumetric correction factor КV.
Accounting for those factors we calculate the vessel length per volume (Z) in cleared (Zcleared) and real tissue (Zreal) in Equation V., where NV,vox is the number of total voxels in the reference Volume and NCl,vox is the number of centerline voxels in the image Volume:
(V.) |
Similarly, we calculate the bifurcation density (B) in cleared and real tissue in Equation VI., where NBif,vox is the number of bifurcations in the reference Volume:
(VI.) |
Please note, the voxel spacing of 3 μm has to be considered when reporting the features in SI-units.
Inter-annotator experiment for the features
To estimate the error in VesSAP’s feature quantification, we extracted the features on a labeled test set of 5 volumes of 500 × 500 × 50 pixels. When comparing to the gold standard label, we calculate errors (disagreements) of 8.21% for the centerlines, 13.18% for the number of bifurcation points and 16.33% for the average radius. To compare VesSAP’s extracted features to a human level annotation we implemented an inter-annotator experiment. For this experiment we had 4 annotators, who labeled the vessels and radius in 2 volumes of 500 × 500 × 50 pixels using the ImageJ and ITK-snap. Finally, we calculated the agreement of the extracted features between all annotators and compared to the gold standard label.
We calculated this for each of the volumes and find an average error (disagreement) of 34.62% for the radius, 25.20% for the bifurcation count and 12.55% for the centerline length.
The agreement between the VesSAP output and the gold standard is higher than the average agreement between the annotators and the gold standard. This difference underlines the quality and reproducibility of VesSAP’s feature extraction.
Registration to the reference atlas
We used the average template, the annotation file and the latest ontology file (Ontology ID: 1) of the current Allen mouse brain atlas CCFv3 201710. Then we scaled the template and the annotation file up from 10 to 3 μm3 to match our reconstructed brain scans and multiplied the left side of the (still symmetrical) annotation file with −1 so that the labels could be later assigned to the corresponding hemispheres. Next, the average template and the 3D vascular datasets were downsampled to 10% of their original size in each dimension to achieve a reasonably fast alignment with the elastix toolbox50 (v4.9.0). For the sake of the integrity of the extracted features, we aligned the template to each of the brain scans individually using a two-step rigid and deformable (B-Spline, optimizer: AdaptiveStochasticGradientDescent, metric: AdvancedMattesMutualInformation, grid spacing in physical units: 90, in the VesSAP repository we host the log and parameter files for each brain scan) registration and applied the transformation parameters to the full resolution annotation volume (3 μm resolution). Subsequently, we created masks for the anatomical clusters based on the current Allen brain atlas ontology.
Statistical analysis of features
Data collection and analysis were not performed blind to the strains. Data distribution was assumed to be normal, though this was not formally tested. All data values of the descriptive statistics are given as mean ± SEM unless stated otherwise. Data were analyzed with standardized effect size indices (Cohen’s d)51 to investigate differences of vessel length, number of bifurcation points and radii between brain areas across the three mouse strains (n=3 per strain). Descriptive statistics were evaluated across brain regions in the pooled (n=9) dataset.
Data visualization
All volumetric datasets were rendered using Imaris, Vision4D and ITK Snap.
CODE AND DATA AVAILABILITY
VesSAP codes and data are publicly hosted at http://DISCOtechnologies.org/VesSAP, and include the imaging protocol, data (original scans, registered atlas data), trained algorithms, training data and a reference set of features describing the vascular network in all brain regions. Additionally, the source code is hosted on GitHub (https://github.com/vessap/vessap) and on the executable platform CodeOcean (https://doi.org/10.24433/CO.1402016.v1). Implementation of external libraries is available on request.
Supplementary Material
ACKNOWLEDGMENTS
This work was supported by the Vascular Dementia Research Foundation, Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy within the framework of the Munich Cluster for Systems Neurology (EXC 2145 SyNergy - ID 390857198), ERA-Net Neuron (01EW1501A to A.E.), Fritz Thyssen Stiftung (A.E., Ref. 10.17.1.019MN), DFG research grant (A.E., Ref. ER 810/2-1), Helmholtz ICEMED Alliance (A.E.), NIH (A.E. Ref. AG057575), and the German Federal Ministry of Education and Research via the Software Campus initiative (O.S.). S.S. is supported by the Translational Brain Imaging Training Network (TRABIT) under the European Union’s ‘Horizon 2020’ research & innovation program (Grant agreement ID: 765148). Furthermore, NVIDIA supported this work via the GPU Grant Program. M.I.T is member of Graduate School of Systemic Neurosciences (GSN), Ludwig Maximilian University of Munich. We thank R. Cai, C. Pan, F. Voigt, I. Ezhov, A. Sekuboyina, F. Hellal, R. Malik, U. Schillinger and T. Wang for technical advice, and C. Heisen for critical reading of the manuscript.
Footnotes
CONFLICT OF INTEREST STATEMENT
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
- 1.Bennett RE et al. Tau induces blood vessel abnormalities and angiogenesis-related gene expression in P301L transgenic mice and human Alzheimer’s disease. Proc Natl Acad Sci U S A 115, E1289–E1298, doi: 10.1073/pnas.1710329115 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Obenaus A et al. Traumatic brain injury results in acute rarefication of the vascular network. Sci Rep 7, 239, doi: 10.1038/s41598-017-00161-4 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Völgyi K et al. Chronic Cerebral Hypoperfusion Induced Synaptic Proteome Changes in the rat Cerebral Cortex. Molecular Neurobiology 55, 4253–4266, doi: 10.1007/s12035-017-0641-0 (2018). [DOI] [PubMed] [Google Scholar]
- 4.Klohs J et al. Contrast-enhanced magnetic resonance microangiography reveals remodeling of the cerebral microvasculature in transgenic ArcAbeta mice. J Neurosci 32, 1705–1713, doi: 10.1523/JNEUROSCI.5626-11.2012 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Edwards-Richards A et al. Capillary rarefaction: an early marker of microvascular disease in young hemodialysis patients. Clin Kidney J 7, 569–574, doi: 10.1093/ckj/sfu106 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Calabrese E, Badea A, Cofer G, Qi Y & Johnson GA A Diffusion MRI Tractography Connectome of the Mouse Brain and Comparison with Neuronal Tracer Data. Cereb Cortex 25, 4628–4637, doi: 10.1093/cercor/bhv121 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Dyer EL et al. Quantifying mesoscale neuroanatomy using x-ray microtomography. eNeuro 4, doi: 10.1523/ENEURO.0195-17.2017 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Li T, Liu CJ & Akkin T Contrast-enhanced serial optical coherence scanner with deep learning network reveals vasculature and white matter organization of mouse brain. Neurophotonics 6, 035004, doi: 10.1117/1.NPh.6.3.035004 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Tsai PS et al. Correlations of neuronal and microvascular densities in murine cortex revealed by direct counting and colocalization of nuclei and vessels. Journal of Neuroscience 29, 14553–14570 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Lugo-Hernandez E et al. 3D visualization and quantification of microvessels in the whole ischemic mouse brain using solvent-based clearing and light sheet microscopy. J Cereb Blood Flow Metab 37, 3355–3367, doi: 10.1177/0271678X17698970 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Frangi AF, Niessen WJ, Vincken KL & Viergever MA in International conference on medical image computing and computer-assisted intervention. 130–137 (Springer; ). [Google Scholar]
- 12.Sato Y et al. Three-dimensional multi-scale line filter for segmentation and visualization of curvilinear structures in medical images. Medical image analysis 2, 143–168 (1998). [DOI] [PubMed] [Google Scholar]
- 13.Di Giovanna AP et al. Whole-Brain Vasculature Reconstruction at the Single Capillary Level. Sci Rep 8, 12573, doi: 10.1038/s41598-018-30533-3 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Xiong B et al. Precise Cerebral Vascular Atlas in Stereotaxic Coordinates of Whole Mouse Brain. Front Neuroanat 11, 128, doi: 10.3389/fnana.2017.00128 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Clark TA et al. Artery targeted photothrombosis widens the vascular penumbra, instigates peri-infarct neovascularization and models forelimb impairments. Scientific Reports 9, 2323 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Yao J, Maslov K, Hu S & Wang LV Evans blue dye-enhanced capillary-resolution photoacoustic microscopy in vivo. J Biomed Opt 14, 054049, doi: 10.1117/1.3251044 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Ertürk A et al. Three-dimensional imaging of solvent-cleared organs using 3DISCO. Nature Protocols 7, 1983–1995, doi: 10.1038/nprot.2012.119 (2012). [DOI] [PubMed] [Google Scholar]
- 18.Reitsma S et al. Endothelial glycocalyx structure in the intact carotid artery: a two-photon laser scanning microscopy study. Journal of vascular research 48, 297–306 (2011). [DOI] [PubMed] [Google Scholar]
- 19.Schimmenti LA, Yan HC, Madri JA & Albelda SM Platelet endothelial cell adhesion molecule, PECAM-1, modulates cell migration. Journal of cellular physiology 153, 417–428 (1992). [DOI] [PubMed] [Google Scholar]
- 20.Tetteh G et al. DeepVesselNet: Vessel Segmentation, Centerline Prediction, and Bifurcation Detection in 3-D Angiographic Volumes. arXiv:1803.09340 [cs] (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Weigert M et al. Content-aware image restoration: pushing the limits of fluorescence microscopy. Nature Methods 15, 1090–1097, doi: 10.1038/s41592-018-0216-7 (2018). [DOI] [PubMed] [Google Scholar]
- 22.Wang H et al. Deep learning enables cross-modality super-resolution in fluorescence microscopy. Nat Methods 16, 103–110, doi: 10.1038/s41592-018-0239-0 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Falk T et al. U-Net: deep learning for cell counting, detection, and morphometry. Nat Methods 16, 67–70, doi: 10.1038/s41592-018-0261-2 (2019). [DOI] [PubMed] [Google Scholar]
- 24.Caicedo JC et al. Data-analysis strategies for image-based cell profiling. Nat Methods 14, 849–863, doi: 10.1038/nmeth.4397 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Dorkenwald S et al. Automated synaptic connectivity inference for volume electron microscopy. Nat Methods 14, 435–442, doi: 10.1038/nmeth.4206 (2017). [DOI] [PubMed] [Google Scholar]
- 26.Liu S, Zhang D, Song Y, Peng H & Cai W in Machine Learning in Medical Imaging. (eds Wang Qian, Shi Yinghuan, Suk Heung-Il, & Suzuki Kenji) 185–193 (Springer International Publishing; ). [Google Scholar]
- 27.Long J, Shelhamer E & Darrell T Fully convolutional networks for semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition, 3431–3440 (2015). [DOI] [PubMed] [Google Scholar]
- 28.Livne M et al. A U-net deep learning framework for high performance vessel segmentation in patients with cerebrovascular disease. Frontiers in neuroscience 13, doi: 10.3389/fnins.2019.00097 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Çiçek Ö, Abdulkadir A, Lienkamp SS, Brox T & Ronneberger O 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. International Conference on Medical Image Computing and Computer-Assisted Intervention, 424–432, doi: 10.1007/978-3-319-46723-8_49 (2016). [DOI] [Google Scholar]
- 30.Milletari F, Navab N & Ahmadi SA V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. 2016 Fourth International Conference on 3D Vision (3DV), 565–571, doi: 10.1109/3DV.2016.79 (2016). [DOI] [Google Scholar]
- 31.Litjens G et al. A survey on deep learning in medical image analysis. Medical Image Analysis 42, 60–88, doi: 10.1016/j.media.2017.07.005 (2017). [DOI] [PubMed] [Google Scholar]
- 32.Schneider M, Reichold J, Weber B, Szekely G & Hirsch S Tissue metabolism driven arterial tree generation. Med Image Anal 16, 1397–1414, doi: 10.1016/j.media.2012.04.009 (2012). [DOI] [PubMed] [Google Scholar]
- 33.Chalothorn D, Clayton JA, Zhang H, Pomp D & Faber JE Collateral density, remodeling, and VEGF-A expression differ widely between mouse strains. Physiological Genomics 30, 179–191, doi: 10.1152/physiolgenomics.00047.2007 (2007). [DOI] [PubMed] [Google Scholar]
- 34.Li SZ in Computer Vision — ECCV ′94. (ed Eklundh Jan-Olof) 361–370 (Springer; Berlin Heidelberg: ). [Google Scholar]
- 35.Menti E, Bonaldi L, Ballerini L, Ruggeri A & Trucco E in International Workshop on Simulation and Synthesis in Medical Imaging. 167–176 (Springer; ). [Google Scholar]
- 36.Kataoka H et al. Fluorescent imaging of endothelial glycocalyx layer with wheat germ agglutinin using intravital microscopy. Microsc Res Tech 79, 31–37, doi: 10.1002/jemt.22602 (2016). [DOI] [PubMed] [Google Scholar]
- 37.Steinwall O & Klatzo I Selective vulnerability of the blood-brain barrier in chemically induced lesions. Journal of neuropathology and experimental neurology 25, 542–559 (1966). [DOI] [PubMed] [Google Scholar]
- 38.Pan C et al. Shrinkage-mediated imaging of entire organs and organisms using uDISCO. Nat Methods, doi: 10.1038/nmeth.3964 (2016). [DOI] [PubMed] [Google Scholar]
- 39.Cai R et al. Panoptic imaging of transparent mice reveals whole-body neuronal projections and skull–meninges connections. Nature neuroscience 22, 317–327 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Zhang H, Prabhakar P, Sealock R & Faber JE Wide genetic variation in the native pial collateral circulation is a major determinant of variation in severity of stroke. J Cereb Blood Flow Metab 30, 923–934, doi: 10.1038/jcbfm.2010.10 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Breckwoldt MO et al. Correlated magnetic resonance imaging and ultramicroscopy (MR-UM) is a tool kit to assess the dynamics of glioma angiogenesis. Elife 5, e11712, doi: 10.7554/eLife.11712 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Kingma DP & Ba J Adam: A method for stochastic optimization. arXiv.org, doi: arXiv:1412.6980v9 (2017). [Google Scholar]
- 43.Abadi M et al. in 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16). 265–283. [PMC free article] [PubMed] [Google Scholar]
- 44.Hoo-Chang S et al. Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning. IEEE transactions on medical imaging 35, 1285–1298, doi: 10.1109/TMI.2016.2528162 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Schneider M, Hirsch S, Weber B, Székely G & Menze BH Joint 3-D vessel segmentation and centerline extraction using oblique Hough forests with steerable filters. Medical Image Analysis 19, 220–249, doi: 10.1016/j.media.2014.09.007 (2015). [DOI] [PubMed] [Google Scholar]
- 46.Paetzold JC et al. in Medical Imaging Meets NeurIPS 2019 workshop. (2019).
- 47.Lee TC, Kashyap RL & Chu CN Building Skeleton Models via 3-D Medial Surface Axis Thinning Algorithms. CVGIP: Graphical Models and Image Processing 56, 462–478, doi: 10.1006/cgip.1994.1042 (1994). [DOI] [Google Scholar]
- 48.Rempfler M et al. Reconstructing cerebrovascular networks under local physiological constraints by integer programming. Medical Image Analysis 25, 86–94, doi: 10.1016/j.media.2015.03.008 (2015). [DOI] [PubMed] [Google Scholar]
- 49.Marchesi VT Alzheimer’s dementia begins as a disease of small blood vessels, damaged by oxidative-induced inflammation and dysregulated amyloid metabolism: implications for early detection and therapy. The FASEB Journal 25, 5–13, doi: 10.1096/fj.11-0102ufm (2011). [DOI] [PubMed] [Google Scholar]
- 50.Klein S, Staring M, Murphy K, Viergever MA & Pluim JP Elastix: a toolbox for intensity-based medical image registration. IEEE transactions on medical imaging 29, 196–205 (2009). [DOI] [PubMed] [Google Scholar]
- 51.Cohen J The effect size index: d. Statistical power analysis for the behavioral sciences 2, 284–288 (1988). [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
VesSAP codes and data are publicly hosted at http://DISCOtechnologies.org/VesSAP, and include the imaging protocol, data (original scans, registered atlas data), trained algorithms, training data and a reference set of features describing the vascular network in all brain regions. Additionally, the source code is hosted on GitHub (https://github.com/vessap/vessap) and on the executable platform CodeOcean (https://doi.org/10.24433/CO.1402016.v1). Implementation of external libraries is available on request.