Abstract
Nuclei detection is often a critical initial step in the development of computer aided diagnosis and prognosis schemes in the context of digital pathology images. While over the last few years, a number of nuclei detection methods have been proposed, most of these approaches make idealistic assumptions about the staining quality of the tissue. In this paper, we present a new Multi-Pass Adaptive Voting (MPAV) for nuclei detection which is specifically geared towards images with poor quality staining and noise on account of tissue preparation artifacts. The MPAV utilizes the symmetric property of nuclear boundary and adaptively selects gradient from edge fragments to perform voting for a potential nucleus location. The MPAV was evaluated in three cohorts with different staining methods: Hematoxylin & Eosin, CD31 & Hematoxylin, and Ki-67 and where most of the nuclei were unevenly and imprecisely stained. Across a total of 47 images and nearly 17,700 manually labeled nuclei serving as the ground truth, MPAV was able to achieve a superior performance, with an area under the precision-recall curve (AUC) of 0.73. Additionally, MPAV also outperformed three state-of-the-art nuclei detection methods, a single pass voting method, a multi-pass voting method, and a deep learning based method.
Nuclei detection is often a critical initial step in the development of computer aided diagnosis and automated tissue grading schemes in the context of digital pathology images1,2,3,4,5,6,7,8,9,10,11,12. However, nuclei detection is a challenging task in images with poor staining and noise. In breast cancer diagnosis, the Nottingham Histologic Score system is highly correlated to the appearance of cancer nuclei13. In melanomas, the quantitative assessment of melanocytes within the epidermis is an important cue for disease presence14,15. It is clear therefore that features pertaining to nuclear shape and spatial distribution of cell nuclei have important diagnostic value12. However, manually identifying the location and extent of melanocyte invasion or breast cancer nuclei can be subjective and time consuming. With recent interest in developing automated and computerized cancer diagnosis and grading schemes7,16,17,18, there is a need for improved methods for nuclei detection and counting.
With recent advent of digital pathology, a number of groups have recently proposed approaches for nuclei detection1,14,15,19,20,21,22,23,24,25,26. Most of these approaches make assumptions about the staining quality of the tissue and the resulting image quality and consequently may yield sub-optimal detection performance if the tissue preparation and staining is less than ideal. Due to the uneven absorption of the staining dyes by the tissue, variations in staining procedures adopted in various labs, variations in the exposure time for stain absorption by the tissue, differences in the quality of biochemical tissue staining and resulting appearance of the slides can be substantial. This can adversely affect the appearance of the tissue and the associated histologic primitives such as nuclei. For instance, over-staining of hematoxylin can cause adjacent nuclei to appear to be clumped together, over-staining with eosin can cause the nuclei to blend in with the stroma and result in under-emphasis of the nuclear boundaries. Uneven and imprecise staining can also cause hollow cores within the centers of individual nuclei. An example is shown in Fig. 1(b). Unlike the image shown in Fig. 1(a) where all nuclei are uniformly stained and have a consistent appearance, most nuclei in Fig. 1(b) appear to have an ill-defined nuclear contour, with a hollow interior. While a number of gradient based approaches for nuclear segmentation have been proposed, e.g., active contour19 and level set schemes21, these approaches typically require model initialization. However, because gradient based approaches rely on the presence of strong edges for the model to latch onto, they will invariably yield worse performance on poorly stained images where the nuclei are either not uniformly stained or there are cavities within the nuclei. While deep learning (DL) based approaches13 have the ability to learn discriminative features, the heterogeneous appearance of nuclei in different cohorts and often the availability of only a small number of representative training samples may hinder these supervised learning based approaches. Thus, it is necessary to develop improved, efficient methods to identify individual nuclei on pathology images.
Apart from the gradient based approaches, another class of nuclear detection approaches is the voting family of methods. By utilizing the inherent symmetric property of individual nuclei as a cue, voting based methods aggregate the votes provided by each pixel on the nuclear contour. Since the nuclear contour is generally symmetric, the center region is expected to have a higher value than other regions. While the hollow interior and ill-defined nuclear contour may affect other methods, the voting based methods22,24,27 are typically able to handle the limitations of other families of nuclei detection methods due to their ability to ignore the heterogeneous appearance of nuclei and infer the center of nuclei from broken edge fragments.
In this paper, we present a Multi-Pass Adaptive Voting (MPAV) method that will adaptively select/modify the gradient information on nuclear contours. Unlike existing voting based methods, pixels that will join in the voting procedure will be adaptively selected and refined. This could potentially help alleviate the issue of incorrect voting, resulting in more accurate detection results. In MPAV, we assume that the shape of the objects of interest, i.e., nuclei, is convex. This is generally speaking a valid assumption since nuclei typically have elliptical shapes, which are convex.
Previous Related Work and Novel Contributions
Table 1 enumerates some recent techniques for nuclei detection. Most approaches typically tend to use image derived cues, such as color/intensity25,28,29,30,31, edges19,21,24,32,33,34, texture35, self learned features13,36, and symmetry22,24,27,37.
Table 1. Typical existing methods for nulcei detection.
Category | Nuclei detection methods |
---|---|
Color-based | blue ratio28 |
color clustering29 | |
local adaptive thresholding25 | |
LoG filter31 | |
Edge-based | adaptive H-minima transform31 |
watershed19,21,24 | |
gradient & mophological operation33 | |
circular Hough transform34 | |
Texture-based | diffused gradient vector field35 |
Deep learning-based | convolutional autoencoder neural network36 |
stacked sparse autoencoder13 | |
Voting-based | multiple passes voting27 |
single pass voting22,24 | |
region-based voting37 | |
multi-pass adaptive voting |
The color and texture-based methods require consistent color/texture appearance for the individual nuclei in order to work optimally. The method presented in ref. 31 applied the Laplacian of Gaussian (LoG) filter to detect the initial seed points representing nuclei. However, due to the uneven distribution of nuclear stain, the response of LoG filter may not reflect the true nuclear center. Filipczuk et al. applied circular Hough transform to detect the nuclear center34, however the circular Hough transform assumes that the shape of the underlying region of interest can be represented by a parametric function, i.e., circle or ellipse. In poorly stained tissue images, the circular Hough transform is likely to fail due to the great variations in appearance of nuclear edges and the presence of clusters of edge fragments.
Recently, there has been substantial interest in developing and employing DL based methods for nuclei detection in histology images13,36. The DL methods are supervised classification methods that typically employ multiple layers of neural networks for object detection and recognition. They can be easily extended and employed for multiple different classification tasks. Recently a number of DL based approaches have been proposed for image analysis and classification applications in digital pathology13,36. For instance, Xu et al. proposed a stacked sparse autoencoder (SSAE) to detect nuclei in breast cancer tissue images. They showed that the DL scheme was able to outperform hand-crafted features on multi-site/stain histology images. However, DL methods required a large number of dedicated training samples since the learning process requires a large number of parameters to be learned. These approaches therefore tend to be heavily biased and sensitive to the choice of the training set.
The key idea behind voting based techniques is to cluster circular symmetries along the radial line/inverse gradient direction on an object’s contour in order to infer the center of the object of interest. An illustrative example is shown in Fig. 2(a,b). Figure 2(a) shows a synthetic phantom nucleus with foreground color as grey, and the background color in white. A few sample pixels/points on the nuclei contour with their inverse gradient directions are shown as blue arrows in Fig. 2. Figure 2(b) illustrates the voting procedure with three selected pixels on the contour. Note that for each pixel, a dotted triangle is used to represent an active voting area. The region where three voting areas converge can be thought of as a region with a high likelihood of containing a nuclear center.
Several effective symmetric voting-based techniques have been developed employing variants of the same principal. Parvin et al.27 proposed a multi-pass voting (MPV) method to calculate the centroid of overlapping nuclei. Qi et al.22 proposed a single pass voting (SPV) technique followed by a mean-shift procedure to calculate the seed points of overlapping nuclei. In order to further improve the efficiency of the approach, Xu et al.24 proposed a technique based on an elliptic descriptor and improved single pass voting for nuclei via a seed point detection scheme. This initial nuclear detection step was followed by a marker-controlled watershed algorithm to segment nuclei in H&E stained histology images. In practice, the MPV procedure tends to yield more accurate results compared to the SPV procedure in terms of nuclei detection. The SPV procedure may help improve overall efficiency of nuclear detection24, however, it needs an additional mean-shift clustering step to identify the local maxima in the voting map. This additional clustering step requires estimating additional parameters and increases overall model complexity.
Since existing voting-based techniques typically utilize edge features, nuclei with hollow interiors could result in incorrect voting and hence in generation of a spurious detection result. One example is shown in Fig. 2(c), where we can see a color image, its corresponding edge map and one of the nuclei, denoted as A. Nucleus A has a hollow interior so that it has two contours, an inner and an outer contour, which results in two edge fragments in the edge map (see second row of Fig. 2(c)). For the outer nuclear contour, the inverse gradients are pointing inwards, whereas for the inner nuclear contour, the inverse gradients are pointing outwards. As one may expect, the inverse gradient obtained from the inner contour minimally contributes towards identifying the nuclear centroid (because the active voting area appears to be outside the nucleus, while the nuclear center should be within the nucleus). Another synthetic example of a nucleus with a hollow interior is shown in Fig. 2(c), and a few inverse gradient directions are drawn on the inner contour. In most cases, those inverse gradients from the inner contour will lead to a spurious result in regions of clustered nuclei. In Fig. 2(e), three synthetic nuclei with hollow regions are shown. It is clear that due to the vicinity of these three nuclei, the highlighted red circle region has received a large number of votes and thus could lead to a potential false positive detection. In section, we will show that in real histopathologic images, existing voting-based techniques tend to generate many false positive detection results.
In this paper, we present a Multi-Pass Adaptive Voting (MPAV) method. The MPAV is a voting based technique which adaptively selects and refines the gradient information from the image to infer the location of nuclear centroids. The schematic for the MPAV is illustrated in Fig. 3. The MPAV consists of three modules: gradient field generation, refinement of the gradient field, and multi-pass voting. Given a color image, a gradient field is generated by using image smoothing and edge detection. In the second module, the gradient field is refined, gradients whose direction leads away from the center of the nuclei are removed or corrected. The refined gradient field is then utilized in a multi-pass voting module to guide each edge pixels for generating the nuclear voting map. Finally, a global threshold is applied on the voting map to obtain candidate nuclear centroids. The details of each module are discussed in the next section and the notations and symbols used in this paper are summarized in Table 2.
Table 2. Notation and symbols used in this paper.
Symbols | Description |
---|---|
I | image |
(x, y) | pixel location/coordinate in image |
σGau | the parameter of Gaussian filtering in initial processing |
TCanny | the threshold of Canny edge detection in initial processing |
rmin, rmax | radial range of active voting area |
Q | total number of edge fragments in the edge map |
α | inverse gradient direction |
δ | angular range of active voting area |
A(x, y; rmin, rmax, α, δ) | active voting area |
(u, v) | center of 2-D Gaussian Kernel |
K(u, v; σ, A) | 2-D Gaussian Kernel |
inverse gradient map | |
E | pixels set that belongs to nuclear edges |
Hi | convex hull for the ith edge fragment |
refined inverse gradient map with strategy 1 (GS1) | |
refined inverse gradient map with strategy 2 (GS2) | |
N | variable for iteration count for voting |
Vi | vote image at iteration i |
VN | final vote image at iteration N |
TD | global threshold applied on voting map/probablity map |
precision rate | |
recall rate | |
F-measure | |
MSPV | original single pass voting method24 |
single pass voting method in conjuntion with GS1 | |
single pass voting method in conjuntion with GS2 | |
MMPV | original multiple pass voting method27 |
MSSAE | stacked sparse autoencoder (SSAE) method13 |
multi-pass adaptive voting method with GS1 | |
multi-pass adaptive voting method with GS2 |
Multi-Pass Adaptive Voting
For a pixel located at (x, y), an active voting area A(x, y; rmin, rmax, α, δ) is generated, where rmin and rmax indicate the radial range of the active voting area. rmin is usually set to 1 pixel without loss of reliability (an active voting area can now be represented as A(x, y; rmax, α, δ)). Let α represents the inverse gradient direction at pixel (x, y); δ indicates the angular range. An active voting area is a fan-like area illustrated in Fig. 4. Within the voting area, a kernel, normally a 2-D Gaussian Kernel K(u, v; σ, A)22,24,27, is placed at (u, v) to diffuse the votes. That is, pixels close to (u, v) will have received a higher number of votes and vice versa.
Gradient Field Generation
An unsupervised sparse non-negative matrix factorization (SNMF) based color deconvolution technique38 is first applied to separate different stains. We denote the single stain image as I and its 2-D image coordinates as (x, y). In order to reduce image noise, a smooth image Is is obtained by convolving I with a Gaussian kernel , see Fig. 5(a,b) for examples.
The inverse gradient map is generated using the following equations:
where and represent the gradients along the X and Y axes, respectively.
In this approach, an edge detection method based on Canny operator39 is applied to Is to obtain nuclear edges. An example of edge detection is shown in Fig. 5(c), where white regions represent pixels that belong to nuclear edges. From here on, we only consider pixels that belong to nuclear edges. We denote the pixels set that belongs to a nuclear edge as E. An updated gradient map is then generated as follows:
Refinement of the Gradient Field
In the last step, we obtained an edge map that reflects which pixels will join in the voting procedure (shown in Fig. 5(c)). Since in the scenario that there are staining inconsistencies (see Fig. 2(c)), the inverse gradients on the inner edge fragment may not contribute to the voting of nuclear centroid and may introduce spurious votes in the voting map, we refine the inverse gradient of pixels that will join in the voting procedure. In other words, we do not blindly use all pixels and their corresponding gradient information during the voting step. Instead, we adaptively select the valid pixels and only use their associated gradient information during the voting stage. The steps corresponding to the gradient field refinement are described below.
Convex Hull Center Computation
For an edge fragment, Ei,i∈{1,…,Q}, where Q is the total number of edge fragments in the edge map, we compute the corresponding convex hull, denoted as Hi, and identify the center of the convex hull, denoted as Ci. The convex hull region corresponding to the edge fragment in Fig. 2(c) is shown in Fig. 6(a). The red circle reflects the center of Ci. The white region represents the convex hull region for the outer edge shown in Fig. 2(c), whereas the original pixels belonging to the edge fragment, i.e., pixels p(x, y) ∈ Ei, are marked with rectangles and stars.
Angle Difference Computation
For a pixel in the current edge fragment, p(x, y) ∈ Ei, we compute a vector, , formed by the current pixel, p, and the center of Ci. We then compute the angle difference θ(x, y) between the vector and inverse gradient direction as follows:
Two illustrative examples of the procedure for angle difference computation are shown in Fig. 7(a,b), respectively, where a hollow circle represents a pixel p(x, y) ∈ Ei and a solid red circle represents the center of Ci. The angle difference θ(x, y) between the vector and inverse gradient D(x, y) (shown via the solid blue arrow) is also shown in Fig. 7(a,b).
Identifying Valid Voting Pixels
We employ the following heuristic to determine whether the gradient of a pixel is appropriate for participating in the voting strategy,
where Lp(x,y) denotes the label for pixel p(x, y). The rationale behind Eq. 4 is that if the inverse gradient points towards the hull centroid, it has a positive contribution towards the symmetry voting. On the other hand, if the inverse gradient points outwards from the hull centroid, it contributes negatively and may result in a false positive detection result. Illustrative examples are shown in Fig. 7. Figure 7(a) shows an example where the pixel is determined to be an invalid pixel, while Fig. 7(b) shows an example where the pixel is determined to be a valid pixel. In Fig. 6(a), all the invalid pixels are marked with black rectangles whereas all valid pixels are marked with red stars.
Invalid Voting Gradients Refinement
We present two gradients refinement strategies, Strategy 1 (GS1) and Strategy 2 (GS2), to refine the inverse gradient filed (we denote the MPAV with GS1 as , the MPAV with GS2 as , and the original MPV method as MMPV). GS1 simply ignores the pixels which are labeled as ‘invalid’. Mathematically, the refined inverse gradient map is then calculated as follows:
GS2 involves setting the inverse gradient directions of ‘invalid’ pixels to the reverse direction. Mathematically, the refined inverse gradient map is calculated as follows:
Figure 6(b,c) show the refinement result calculated by using Eqs 5 and 6, respectively. It is clear that for GS1 shown in Eq. 5, the inverse gradients of all invalid pixels are set to 0. For the GS2 shown in Eq. 6, the the directions of inverse gradients for all invalid pixels are reversed. GS1 ignores pixels with ‘invalid’ gradients and reduces the number of pixels that used in the voting procedure, whereas GS2 tries to refine the gradient direction of pixels with ‘invalid’ gradient and attempts to infer a more accurate result. We do not suggest that either of GS1 or GS2 is necessarily better than the another, rather they represent two alternative strategies which the user might want to invoke for a specific application.
Multi-Pass Voting
The steps of MPV are summarized as follows.
Initialize voting parameters
Radial maximum range rmax, maximum angular range δmax, and iteration number N are the voting parameters that need to be initialized. The angular ranges set is computed as follows,
The initial voting directions α for all pixels are based on the inverse gradient map G′ (note that we have two different refinement strategies which correspond to two different refined inverse gradient maps and , respectively). That is,
where and represent the gradients along the X and Y axes, respectively.
Update the vote image
For each pixel (x, y) ∈ E and G′(x, y) ≠ 0, compute the vote image Vi, for the ith iteration, using Eq. 8. In Eq. 8, K(u, v; σ, A(x, y; rmax, αi, δi)) represents the kernel placed at location (u, v) within an active voting region A(x, y; rmax, αi, δi). Gradient map G′ can then be refined gradient map or .
Update voting direction
For each pixel (x, y) ∈ E and G′(x, y) ≠ 0, we employ the following equation to update the voting direction,
where
Intuitively, Eq. 9 involves updating the voting direction for pixel (x, y) with the pixel located at (u*, v*), which is the pixel that has the maximum vote value within the active voting region A, calculated in Eq. 10. In this way, we can perform the next pass/iteration of voting with a more precise voting direction, since the voting from the first iteration may be affected by the presence of noisy edge fragments that do not belong to the nuclei.
Iteration parameter adjustment
Using a smaller angular value, i.e., δi+1, set i = i + 1, and repeat steps 2.3.2 to 2.3.4, until i = N. The reason to reduce the voting area angle is to reduce the size of voting area. The idea is that by doing this we can identify a more precise nuclear centroid candidate region.
After N iterations, a final vote image VN is obtained. The final vote image VN contains the voting information from all pixels that join in the voting procedure. Figure 8 shows the voting image for the MMPV, and and , respectively. The bright regions in Fig. 8 correspond to a high vote values and therefore indicate potential nuclear centers. On the other hand, black regions indicate fewer votes and background location. It is clear that all voting images for each of MMPV, and have high values near the ground truth nuclear centers (the manually labeled nuclei centers are shown in Fig. 9(d)). However, it should be noted that in the voting image obtained by MMPV, there are many regions with a high vote value. Compared to the vote image obtained by MMPV, and appear to yield more specific and accurate detection results.
Finally, a global threshold TD is applied to the voting image VN to obtain nuclear centers. Specifically, we use the automated Ostu’s threshold method40 to determine a global threshold TD. We then apply this threshold on VN to obtain a binary image. The centroids of the connected components are then identified as the final nuclear centers. Figure 9(a–c) show the final nuclear centers, highlighted with blue crosses, obtained from each of MMPV, , and , respectively. The manually labeled nuclear centers are shown in Fig. 9(d).
Experimental Setup
Image Dataset
We evaluated the and on three different histopathologic image datasets. Table 3 describes the properties of these datasets. Dataset A is a publicly available histopathologic image dataset (H&E-stained digital images of breast cancer tissue slides)41. Dataset B consists of 10 breast cancer images stained with CD31 antibody and hematoxylin13. Dataset C consists of 5 skin cancer images stained with Ki-6742. Dataset D consists of 21 breast cancer tissue images stained with H&E. Note that the dataset A and D contain images with noisy background and nuclei with heterogeneous appearance due to the uneven staining. The nuclei locations were manually labeled by two experts, one a breast cancer pathologist with over 10 years of experience and the second an image computing scientist with over 3 years of experience in working with breast pathology. Figure 10 illustrates example images from A, B, C, and D, respectively.
Table 3. Image datasets used for evaluation.
Evaluation Metrics
The main objective of the evaluation was to determine if the nuclear centers detected by the MPAV technique are concordant with the manually labeled nuclear centers. We calculate the center of each segmented region obtained by the automated technique. A nucleus was identified as having been correctly detected if its center is within a range of 15 pixels, i.e., about 3.6825 μm, of the manually labeled nuclei center location.
Denoting as the total number of manually labeled nuclei centers, as the total number of detected nuclei centers, as the number of true-positives, (i.e., correctly detected objects compared to the manually labeled ones), as the number of false-positives. (i.e., falsely detected objects compared to the manually labeled ones).
The performance is evaluated with respect to recall (), precision (), and F-measure () which are defined as follows:
In order to evaluate detection performance, all the nuclei detection techniques (see comparative methods below in Section IV-C) were evaluated via the precision-versus-recall rate curve (PRC). In a PRC, the horizontal and vertical axes represents and , respectively. The closer the PR curve to the upper left corner, the better the corresponding detection method. Area under the PRC curve (AUC) is also calculated. The minimum and maximum value of a AUC is 0 and 1, corresponding in turn to the worst and best possible detection results. Note that the PRC for each of MSPV, , and , MMPV, , and is calculated by tuning a global threshold TD on all VN for all the images in dataset. Assuming the value range in VN is [0, 1], we use a set of value varying from 0 to 1 with a resolution of 0.05 for the TD. The detection performance of applied a TD corresponds to a point on the PRC curve.
Experimental Design and Comparative Strategy
Performance comparison with existing voting based techniques
We evaluated the performance of the original SPV technique24 (denoted as MSPV) and its enhanced version by adaptively choosing and refining the gradient information using , and , respectively. We also evaluated the performance of MMPV and the enhanced versions , and , respectively. For all the voting based methods, we set rmin = 1, rmax = 40, σ = 4; for SPV methods, ; for MMPV, , and , we set the voting iteration number N to 3, (More details on parameters selection in initial processing and the parameters selection of voting area please refer to the supplementary material A and B).
Performance comparison with DL based technique
We also compared the detection performance of , and with a DL based technique, SSAE (denoted as MSSAE), developed by Xu et al.13. The MSSAE is a 4 layers deep neural network. The input patch size used in MSSAE was 34 × 34 pixels, therefore, the input layer receives 342 × 3 input units. The first and second hidden layers have 400 and 255 hidden units, respectively. The output layer is a soft-max function which classifies the input patches as nucleus or non-nucleus. The MSSAE was trained with a different set, other than datasets A, B, and C, of 37 H&E images of 2200 × 2200 pixels at 40x magnification13.
Results and Discussions
Performance comparison with existing voting based techniques
Qualitative results
Figure 11 presents three examples illustrating MMPV, , and . The manually labeled nuclei centers are shown in the last column, whereas automatically calculated nuclei centers are shown in the first three columns (all nuclei centers are marked with blue crosses) for MMPV, , and . It is clear that MMPV tends to generate false positives errors in regions where nuclei are clustered and image noise is present. This is because neighboring pixels from different nuclear edge fragments generate spurious regions with high vote value. and are able to generate more accurate results with fewer false positives (automatically generated nuclear centers which are not concordant with the manually labeled nuclear centers).
Quantitative results
The detection performance in terms of PRC for datasets A, B, C, and D are shown in Fig. 12. The first row shows the performance of MSPV24 compared to the and techniques (Fig. 12(a–c)). The second row shows the performance of MMPV27 compared to and (Fig. 12(d–f)). Figure 12(g) shows the overall detection performance comparison for methods MMPV, MSPV, , , , and on datasets A, B, C and D.
The nuclei detection results for dataset A and D are shown in Fig. 12(a,e) and (d,h), respectively. and achieve better performance in such noisy images since they adaptively refine the gradient information for voting. In Fig. 12(b,f), it may be observed that the improvements obtained by the gradient refinement strategies for dataset B are less substantial. In Fig. 12(c,g), the detection performances are almost identical. This is because the images in datasets B and C contain less noise that may affect the voting procedure. The refinement procedure thus has little effect on the voting result. In dataset A and D, there are many clustered nuclei, and hence, and yield substantial gains in performance over MMPV, MSPV, , and . Figure 12(g) shows the detection performance comparison of MMPV27, MSPV24, , , and the and for datasets A, B, and C, comprising a total of 11844 annotated nuclei. It is worth nothing that for , where the gradients direction are reversed (in Eq. 6), provides the best performance.
In Table 4, we present the quantitative result of MMPV, MSPV, , , , and in terms of the best and its associated and (fifth, thrid and fourth column, respectively). The area under the PRC curve for all plots, shown in Fig. 12, are also calculated and presented in Table 4.
Table 4. Performance comparison of MSPV, , and , MMPV, , and , in terms of , , , and AUC, evaluated evaluated on datasets A, B, and C.
Dataset | Techniques | (%) | (%) | (%) | AUC |
---|---|---|---|---|---|
Dataset A | MSPV | 67.59 | 67.44 | 67.51 | 0.72 |
70.26 | 71.37 | 70.81 | 0.75 | ||
70.91 | 81.33 | 75.76 | 0.79 | ||
MMPV | 65.27 | 77.86 | 71.01 | 0.72 | |
77.64 | 71.53 | 74.46 | 0.77 | ||
79.94 | 78.46 | 79.19 | 0.82 | ||
Dataset B | MSPV | 80.51 | 68.34 | 73.93 | 0.77 |
75.81 | 74.89 | 75.35 | 0.79 | ||
79.54 | 72.04 | 75.60 | 0.79 | ||
MMPV | 75.37 | 75.36 | 75.36 | 0.78 | |
78.57 | 75.27 | 76.88 | 0.80 | ||
79.29 | 73.80 | 76.44 | 0.80 | ||
Dataset C | MSPV | 72.00 | 76.00 | 73.95 | 0.74 |
73.00 | 78.00 | 75.42 | 0.75 | ||
73.00 | 79.00 | 75.88 | 0.75 | ||
MMPV | 76.43 | 74.99 | 75.70 | 0.77 | |
72.00 | 82.00 | 76.68 | 0.77 | ||
74.00 | 81.00 | 77.34 | 0.77 | ||
Dataset D | MSPV | 60.58 | 52.70 | 56.37 | 0.56 |
64.94 | 55.44 | 59.81 | 0.61 | ||
66.63 | 51.31 | 57.98 | 0.62 | ||
MMPV | 60.40 | 59.76 | 60.08 | 0.60 | |
66.01 | 57.85 | 61.66 | 0.64 | ||
67.65 | 61.15 | 64.24 | 0.67 | ||
All | MSPV | 64.33 | 66.93 | 65.60 | 0.66 |
74.33 | 62.58 | 67.95 | 0.68 | ||
75.45 | 62.43 | 68.32 | 0.70 | ||
MMPV | 75.62 | 61.53 | 67.85 | 0.67 | |
72.64 | 65.09 | 68.65 | 0.71 | ||
72.02 | 69.16 | 70.56 | 0.73 |
The highest performance for each dataset for each metric is shown in bold.
Performance comparison with DL based technique M SSAE
Qualitative results
Figure 13 presents three visual examples for the performance comparison of methods MSSAE, , and on datasets A, B, and C. The manually labeled nuclei centers are shown in the last column, whereas automatically calculated nuclei centers for methods MSSAE, , and are shown in the first three columns (all nuclear centers are marked with blue crosses). An example image from dataset A is shown in the first row, MSSAE appears to produce a number of false positive detection results for this poorly stained image. Additionally MSSAE is prone to treat the boundary point as the nuclei center due to the heterogeneous intensity distribution within the nuclei and nuclei boundaries. In the second row, one may observe that the MPAV appears to produce more false positive errors for nuclei that have a thin and long shape (see Fig. 13(f,g)).
Quantitative results
The detection performance in terms of PRC for all 4 datasets are shown in Fig. 14. As shown in Fig. 14(a), the nuclei detection result for MSSAE on dataset A is poor, since the MSSAE is learned from dataset with well stained nuclei (and we also trained MSSAE with dataset A and then tested on dataset A, the result is even worse, possibly on account of too few training instances). In Fig. 14(c), MSSAE marginally outperforms and . However, and , provide consistently good results on all four datasets. and may generate spurious result if the nuclear size is relatively large, since the voting is based on a predefined range of active voting area, i.e., rmax and rmin. However, the number of abnormally large nuclei is small in most of the images, most of the nuclei generally fall within a fixed size range. and are unsupervised methods that do not require training, whereas the MSSAE requires a large number of samples to learn the underlying patterns, and the heterogeneous appearance of nuclei in different stain conditions may hence affect the learned features and final detection performance.
Table 5 presents corresponding quantitative result of MSSAE, , and in terms of the best and its associated and for datasets A, B, C, and D. The area under the PRC curve, for the curves shown in Fig. 14 are presented in the last column of Table 5.
Table 5. Performance comparison of MSSAE, , and , in terms of , , , and AUC, evaluated on datasets A, B, and C.
Dataset | Techniques | (%) | (%) | (%) | AUC |
---|---|---|---|---|---|
Dataset A | MSSAE | 33.60 | 47.53 | 39.37 | 0.36 |
77.64 | 71.53 | 74.46 | 0.77 | ||
79.94 | 78.46 | 79.19 | 0.82 | ||
Dataset B | MSSAE | 75.07 | 82.64 | 78.67 | 0.81 |
78.57 | 75.27 | 76.88 | 0.80 | ||
79.29 | 73.80 | 76.44 | 0.80 | ||
Dataset C | MSSAE | 75.63 | 78.75 | 77.16 | 0.81 |
72.00 | 82.00 | 76.68 | 0.77 | ||
74.00 | 81.00 | 77.34 | 0.77 | ||
Dataset D | MSSAE | 80.58 | 39.99 | 53.45 | 0.57 |
66.01 | 57.85 | 61.66 | 0.64 | ||
67.65 | 61.15 | 64.24 | 0.67 | ||
All | MSSAE | 58.21 | 70.60 | 63.81 | 0.62 |
72.64 | 65.09 | 68.65 | 0.71 | ||
72.02 | 69.16 | 70.56 | 0.73 |
The highest performance for each dataset for each metric is shown in bold.
Statistics evaluation of detection results
We calculated the AUC for every image in the datasets and plotted the AUCs as a point for each image, as shown in Fig. 15. Figure 15(a–d) show the AUC cloud plots for dataset A–D respectively. A higher AUC corresponds to a better performance. From these cloud plots, one may observe the detection performance for each image within a dataset. The two-sample t-test p-values between each technique are calculated based on the AUCs obtain from different techniques and presented in Table 6. In this work, we used 0.05 as the significance level for the p-valudes to determine statistical significance. One may observe that in dataset A and D, the is significantly different than other techniques. Related to Figs 12 and 14, the is actually provide “significantly” better results than MSSAE, MSPV, and MMPV techniques in the noisy images. As in the cases of dataset B and C, the technique is not significantly different than the MSSAE techniques. Note that in Figs 12 and 14, we calculated the PRC for all the images within each image dataset, whereas in Fig. 15, we generated the PRC and calculated the AUC for each image whithin the dataset spearately.
Table 6. Statistics evaluation of MSSAE, MSPV, MMPV, , and , in terms of AUC, evaluated on datasets A, B, C, and D.
Dataset | Techniques | MSSAE | MSPV | MMPV | ||
---|---|---|---|---|---|---|
Dataset A | MSSAE | 1 | 5.83 × 10−14 | 2.50 × 10−11 | 2.77 × 10−13 | 2.15 × 10−14 |
MSPV | 5.83 × 10−14 | 1 | 4.00 × 10−2 | 5.51 × 10−5 | 1.14 × 10−6 | |
MMPV | 2.50 × 10−11 | 4.00 × 10−2 | 1 | 7.00 × 10−2 | 1.00 × 10−2 | |
2.77 × 10−13 | 5.51 × 10−5 | 7.00 × 10−2 | 1 | 3.90 × 10−1 | ||
2.15 × 10−14 | 1.14 × 10−6 | 1.00 × 10−2 | 3.90 × 10−1 | 1 | ||
Dataset B | MSSAE | 1 | 8.00 × 10−2 | 6.20 × 10−1 | 7.90 × 10−1 | 5.60 × 10−1 |
MSPV | 8.00 × 10−2 | 1 | 2.88 × 10−2 | 1.60 × 10−3 | 1.81 × 10−4 | |
MMPV | 6.20 × 10−1 | 2.88 × 10−2 | 1 | 1.80 × 10−1 | 4.08 × 10−2 | |
7.90 × 10−1 | 1.60 × 10−3 | 1.80 × 10−1 | 1 | 5.40 × 10−1 | ||
5.60 × 10−1 | 1.81 × 10−4 | 4.08 × 10−2 | 5.40 × 10−1 | 1 | ||
Dataset C | MSSAE | 1 | 8.40 × 10−3 | 1.90 × 10−1 | 2.90 × 10−1 | 2.50 × 10−1 |
MSPV | 8.40 × 10−3 | 1 | 4.81 × 10−2 | 3.20 × 10−2 | ||
MMPV | 1.90 × 10−1 | 4.81 × 10−2 | 1 | 7.70 × 10−2 | 8.30 × 10−1 | |
2.90 × 10−1 | 3.20 × 10−2 | 7.70 × 10−2 | 1 | 9.30 × 10−1 | ||
2.50 × 10−1 | 3.20 × 10−2 | 8.30 × 10−1 | 9.30 × 10−1 | 1 | ||
Dataset D | MSSAE | 1 | 4.20 × 10−3 | 8.70 × 10−1 | 9.00 × 10−2 | 2.00 × 10−2 |
MSPV | 4.20 × 10−3 | 1 | 3.00 × 10−3 | 6.81 × 10−6 | 5.96 × 10−7 | |
MMPV | 8.70 × 10−1 | 3.00 × 10−3 | 1 | 1.40 × 10−1 | 3.00 × 10−2 | |
9.00 × 10−2 | 6.81 × 10−6 | 1.40 × 10−1 | 1 | 4.50 × 10−1 | ||
2.00 × 10−2 | 5.96 × 10−7 | 3.00 × 10−2 | 4.50 × 10−1 | 1 |
A two sample t-test is performed between each technique, the p-values are shown in the table. The p-values that are smaller than 0.05 are shown in bold, which is considered to be statistically significant in this work.
Concluding Remarks
In this paper, a Multi-Pass Adaptive Voting technique was presented for automated nuclei detection on histopathological images. While most previous nuclear detection approaches tend to work well on well stained images, these approaches tend to fail on unevenly and imprecisely stained images. In practice, due to the variations of staining and slide preparation methods, a number of images tend to stain poorly. The MPAV is able to adaptively ignore or refine and identify the pixels that may lead to inaccurate nuclei centers in the voting procedure. To show the effectiveness of the MPAV, we compared it with existing voting based methods as well as a state-of-the-art deep learning method. The main contribution of the presented MPAV is to adaptively utilize the gradient information of pixels on edge fragments to generate more accurate detection results by exploiting the symmetry of nuclei. Both qualitative and quantitative evaluation results show that the MPAV appears to be able to address many of the limitations of existing voting based techniques for nuclei detection in unevenly and imprecisely stained histology images. While compared to the DL-based method, MPAV provides a consistently superior detection performance whereas the DL method requires more training samples for poorly stained images. Note that the MPAV approach aims solely to detect nuclei and not to explicitly segment them, though the result of detection could serve as the initialization for other (e.g., watershed24) nuclear segmentation approaches. However, our approach does have its limitations. Firstly, due to the noise and extraneous tissue components in the image, there may exist some noisy edge pixels outside the true nuclei on the edge segments. In these cases, the estimated center of the convex hull may fall outside the true nuclei, which leads to an inaccurate determination of which pixels are either valid or invalid for the purposes of voting. In our experiments, the number of such cases was small. Secondly, while our new gradient refinement strategy is able to improve the detection performance in noisy histopathological images, for relatively good quality images with well stained nuclei, like datasets B and C shown in Fig. 10(b,c), there is marginal or no improvement (detection result shown in Fig. 12(b,e,c,f)). However, the MPAV technique only has a marginal increment in terms of computational cost compared to the original MPV method (A Matlab implementation of MPAV algorithm is provided in the supplementary material C). In future work, we aim to integrate the MPAV with segmentation methods such as level sets and active contour schemes to explicitly extract nuclear boundaries.
Additional Information
How to cite this article: Lu, C. et al. Multi-Pass Adaptive Voting for Nuclei Detection in Histopathlogical Images. Sci. Rep. 6, 33985; doi: 10.1038/srep33985 (2016).
Supplementary Material
Acknowledgments
Research reported in this publication was supported by the National Natural Science Foundation of China (Grant No. 61401263, No. 61401265, No. 61273259, No. 61502290, No. 61501287, No. 61573232), Industrial Research Project of Science and Technology in Shaanxi Province (2015GY016), Natural Science Basic Research Plan in Shaanxi Province of China (Program No. 2015JQ6228, 2016JQ6056), Fundamental Research Funds for the Central Universities of China (Grant No. GK201402037, GK201503061), Six Major Talents Summit of Jiangsu Province (No. 2013-XXRJ-019), Natural Science Foundation of Jiangsu Province of China (No. BK20141482), Jiangsu Innovation & Entrepreneurship Group Talents Plan (No. JS201526), and Scientific Research Foundation for the Returned Overseas Chinese Scholars, State Education Ministry, National Cancer Institute of the National Institutes of Health under award numbers 1U24CA199374-01, R21CA167811-01, R21CA179327-01, R21CA195152-01; the National Institute of Diabetes and Digestive and Kidney Diseases under award number R01DK098503-02, the DOD Prostate Cancer Synergistic Idea Development Award (PC120857); the DOD Lung Cancer Idea Development New Investigator Award (LC130463), the DOD Prostate Cancer Idea Development Award; the Ohio Third Frontier Technology development Grant, the CTSC Coulter Annual Pilot Grant, the Case Comprehensive Cancer Center Pilot Grant, VelaSano Grant from the Cleveland Clinic, the Wallace H. Coulter Foundation Program in the Department of Biomedical Engineering at Case Western Reserve University. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Footnotes
Dr. Madabhushi is an equity holder in Elucid Bioimaging and in Inspirata Inc. He is also a scientific advisory consultant for Inspirata Inc and also sits on its scientific advisory board. He is also an equity holder in Inspirata Inc. Additionally his technology has been licensed to Elucid Bioimaging and Inspirata Inc. He is also involved in a NIH U24 grant with PathCore Inc. Dr. Lu, Mr. Xu, Dr. Xu, Dr. Gilmore, and Dr. Mandal declare no competing financial interests.
Author Contributions C.L. developed the method, M.M. and A.M. conceived the experiments, J.X. conducted the experiments, H.X. and H.G. provided the annotations, C.L. and H.X. analyzed the results. All authors reviewed the manuscript.
References
- Sahirzeeshan A., Veltri R., Epstein J. I., Christudass C. & Madabhushi A. Selective invocation of shape priors for deformable segmentation and morphologic classification of prostate cancer tissue microarrays. Computerized medical imaging and graphics 14, 176–181 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ginsburg S., Lee G., Ali S. & Madabhushi A. Feature Importance in Nonlinear Embeddings (FINE): Applications in Digital Pathology. IEEE Trans. Med. Imaging 35, 76–88 (2015). [DOI] [PubMed] [Google Scholar]
- Sridhar A., Doyle S. & Madabhushi A. Content-based image retrieval of digitized histopathology in boosted spectrally embedded spaces. J Pathol Inform 6, 41–49 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Basavanhally A., Viswanath S. & Madabhushi A. Predicting classifier performance with limited training data: applications to computer-aided diagnosis in breast and prostate cancer. PLoS One 10, 1–18 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee G., Ali S., Veltri R., Epstein J. I., Christudass C. & Madabhushi A. Cell orientation entropy (COrE): predicting biochemical recurrence from prostate cancer tissue microarrays. Med Image Comput Comput Assist Interv. 16, 396–403 (2013). [DOI] [PubMed] [Google Scholar]
- Lewis J. S., Ali S., Luo J., Thorstad W. L. & Madabhushi A. A quantitative histomorphometric classifier (QuHbIC) identifies aggressive versus indolent p16-positive oropharyngeal squamous cell carcinoma. Am J Surg Pathol. 16, 128–137 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Basavanhally A. et al. “Multi-field-of-view framework for distinguishing tumor grade in ER+ breast cancer from entire histopathology slides”. IEEE Transactions on Biomedical Engineering 60, 2089–2099 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ghaznavi F., Evans A., Madabhushi A. & Feldman M. Digital imaging in pathology: whole-slide imaging and beyond. Annu Rev Pathol. 24, 331–359 (2013). [DOI] [PubMed] [Google Scholar]
- Madabhushi A., Agner S., Basavanhally A., Doyle S. & Lee G. Computer-aided prognosis: predicting patient and disease outcome via quantitative fusion of multi-scale, multi-modal data. Comput Med Imaging Graph. 35, 506–514 (2011). [DOI] [PubMed] [Google Scholar]
- Madabhushi A. et al. Integrated diagnostics: a conceptual framework with examples. Clin Chem Lab Med. 48, 989–998 (2010). [DOI] [PubMed] [Google Scholar]
- Bhargava R. & Madabhushi A. A Review of Emerging Themes in Image Informatics and Molecular Analysis for Digital Pathology. Annual Review of Biomedical Engineering, 18, 1–10 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu C. & Mandal M. Automated analysis and diagnosis of skin melanoma on whole slide histopathological images. Pattern Recognition 48, 2738–2750 (2015). [Google Scholar]
- Xu J. et al. Stacked sparse autoencoder (SSAE) for nuclei detection on breast cancer histopathology images. IEEE Transactions on Medical Imaging 35, 119–130 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lu C., Mahmood M., Jha N. & Mandal M. Detection of melanocytes in skin histopathological images using radial line scanning. Pattern Recognition 46, 509–518 (2013). [Google Scholar]
- Lu C., Mahmood M., Jha N. & Mandal M. Automated segmentation of the melanocytes in skin histopathological images. IEEE Journal of Biomedical and Health Informatics 17, 284–296 (2013). [DOI] [PubMed] [Google Scholar]
- Lee G. et al. Supervised Multi-View Canonical Correlation Analysis (sMVCCA): Integrating histologic and proteomic features for predicting recurrent prostate cancer. IEEE transactions on medical imaging, 34, 284–97 (2014). [DOI] [PubMed] [Google Scholar]
- Lee G. et al. Co-occurring gland angularity in localized subgraphs: predicting biochemical recurrence in intermediate-risk prostate cancer patients. PloS one 9, 1–14 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ali S., Lewis J. S. & Madabhushi A. Spatially aware cell cluster(spaccl) graphs: Predicting outcome in oropharyngeal p16+ tumors, in International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), ser. Lecture Notes in Computer Science, Springer Berlin Heidelberg. 1em plus 0.5em minus 0.4em Springer Berlin Heidelberg 8149, 412–419 (2013). [DOI] [PubMed]
- Ali S. & Madabhushi A. An integrated region-, boundary, shape-based active contour for multiple object overlap resolution in histological imagery. IEEE transactions on medical imaging 31, 1448–60 (2012). [DOI] [PubMed] [Google Scholar]
- Ali S. & Madabhushi A. Graphical processing unit implementation of an integrated shape-based active contour: Application to digital pathology. J Pathol Inform 2, 1–13 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fatakdawala H. et al. Expectation–maximization-driven geodesic active contour with overlap resolution (emagacor): Application to lymphocyte segmentation on breast cancer histopathology. IEEE Transactions on Biomedical Engineering, 57, 1676–1689 (2010). [DOI] [PubMed] [Google Scholar]
- Qi X., Xing F., Foran D. & Yang L. Robust segmentation of overlapping cells in histopathology specimens using parallel seed detection and repulsive level set. IEEE Trans Biomed Eng, 59, 754–765 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Veta M. et al. Automatic Nuclei Segmentation in H & E Stained Breast Cancer Histopathology Images. PLOS one, 8, 1–13 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xu H., Lu C. & Mandal M. An efficient technique for nuclei cluster segmentation based on ellipse descriptor analysis and improved seed detection algorithm. IEEE Journal of Biomedical and Health Informatics, 18, 1729–1741 (2014). [DOI] [PubMed] [Google Scholar]
- Lu C., Mahmood M., Jha N. & Mandal M. A robust automatic nuclei segmentation technique for quantitative histopathological image analysis. Analytical and Quantitative Cytology and Histology, 34, 296–308 (2012). [PubMed] [Google Scholar]
- Lu C. & Mandal M. Toward automatic mitotic cell detection and segmentation in multispectral histopathological images. IEEE Journal of Biomedical and Health Informatics, 18, 594–605 (2014). [DOI] [PubMed] [Google Scholar]
- Parvin B. et al. Iterative voting for inference of structural saliency and characterization of subcellular events. IEEE Trans. on Image Processing, 16, 615–623 (2007). [DOI] [PubMed] [Google Scholar]
- Chang H. et al. Invariant delineation of nuclear architecture in glioblastoma multiforme for clinical and molecular association. IEEE Transactions on Medical Imaging 32, 670–682 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cataldo S., Ficarra E., Acquaviva A. & Macii E. Automated segmentation of tissue images for computerized ihc analysis. Computer methods and programs in biomedicine 100, 1–15 (2010). [DOI] [PubMed] [Google Scholar]
- Ruifrok A. & Johnston D. Quantification of histochemical staining by color deconvolution. Analytical and quantitative cytology and histology/the International Academy of Cytology [and] American Society of Cytology, 23, 291–299 (2001). [PubMed] [Google Scholar]
- Al-Kofahi Y., Lassoued W., Lee W. & Roysam B. Improved automatic detection and segmentation of cell nuclei in histopathology images. IEEE Transactions on Biomedical Engineering, 57, 841–852 (2010). [DOI] [PubMed] [Google Scholar]
- Jung C. & Kim C. Segmenting Clustered Nuclei Using H-minima Transform-Based Marker Extraction and Contour Parameterization. IEEE Transactions on Biomedical Engineering, 57, 2600–2604 (2010). [DOI] [PubMed] [Google Scholar]
- Esmaeilsabzali H., Sakaki K., Dechev N., Burke R. & Park E. Machine vision-based localization of nucleic and cytoplasmic injection sites on low-contrast adherent cells. Medical & biological engineering & computing, 50, 11–21 (2012). [DOI] [PubMed] [Google Scholar]
- Filipczuk P., Fevens T., Krzyzak A. & Monczak R. Computer-aided breast cancer diagnosis based on the analysis of cytological images of fine needle biopsies. IEEE Transactions on Medical Imaging, 32, 2169–2178 (2013). [DOI] [PubMed] [Google Scholar]
- Liu T. et al. An automated method for cell detection in zebrafish. Neuroinformatics, 6, 5–21 (2008). [DOI] [PubMed] [Google Scholar]
- Cruz-Roa A., Ovalle J., Madabhushi A. & Osorio F. A deep learning architecture for image representation, visual interpretability and automated basal-cell carcinoma cancer detection. Medical Image Computing and Computer-Assisted Intervention-MICCAI 2013. 16, 403–410 (2013). [DOI] [PubMed] [Google Scholar]
- Xing F., Su H., Neltner J. & Yang L., Automatic ki-67 counting using robust cell detection and online dictionary learning. IEEE Transactions on Biomedical Engineering, 61, 859–870 (2014). [DOI] [PubMed] [Google Scholar]
- Xu J. et al. Sparse non-negative matrix factorization (snmf) based color unmixing for breast histopathological image analysis, Computerized Medical Imaging and Graphics 40, 20–29 (2015). [DOI] [PubMed] [Google Scholar]
- Canny J. A computational approach to edge detection. IEEE Trans Pattern Anal Mach Intell. 8, 679–698 (1986). [PubMed] [Google Scholar]
- Otsu N. A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man and Cybernetics 9, 62–66 (1979). [Google Scholar]
- Ludovic R. et al. Mitos&atypia-detection of mitosis and evaluation of nuclear atypia score in breast cancer histological images, http://mitos-atypia-14.grand-challenge.org/home/ (2014).
- Nielsen P., Spaun E., Riber-Hansen R. & Torben S. Automated quantification of MART1-verified Ki-67 indices: useful diagnostic aid in melanocytic lesions. Human pathology 45, 1153–61 (2014). [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.