Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Jul 15.
Published in final edited form as: Conf Proc IEEE Eng Med Biol Soc. 2018 Jul;2018:3410–3413. doi: 10.1109/EMBC.2018.8512961

Clumped Nuclei Segmentation with Adjacent Point Match and Local Shape-Based Intensity Analysis in Fluorescence Microscopy Images

Xiaoyuan Guo 1, Hanyi Yu 1, Blair Rossetti 1, George Teodoro 2, Daniel Brat 3, Jun Kong 4
PMCID: PMC7363396  NIHMSID: NIHMS1584838  PMID: 30441120

Abstract

Highly clumped nuclei captured in fluorescence microscopy images are commonly observed in a wide spectrum of tissue-related biomedical investigations. To ensure the quality of downstream biomedical analyses, it is essential to accurately segment clustered nuclei. However, this presents a technical challenge as fluorescence intensity alone is often insufficient for recovering the true nuclei boundaries. In this paper, we propose an segmentation algorithm that identifies point pair connection candidates and evaluates adjacent point connections with a formulated ellipse fitting quality indicator. After connection relationships are determined, we recover the resulting dividing paths by following points with specific eigenvalues from the image Hessian in a constrained searching space. We validate our algorithm with 560 image patches from two classes of tumor regions of seven brain tumor patients. Both qualitative and quantitative experimental results suggest that our algorithm is promising for dividing overlapped nuclei in fluorescence microscopy images widely used in various biomedical research.

I. INTRODUCTION

Fluorescence microscopy images are commonly used in clinical and biomedical research to help visualize cellular components, such as membranes and nuclei. Analysis of fluorescence microscopy images often requires an accurate identification of individual cell nuclei. However, clumped nuclei due to a high cell density often make it challenging to achieve an accurate nuclei segmentation. Although a large number of clumped nuclei segmentation methods have been proposed for fluorescence microscopy image analysis, they are subject to salient defects [1], [2]. For example, shape-based methods are sensitive to clumped nuclei shape variance which often leads to under-segmentation [2], [3], [4], [5]. Similarly, under-segmentation occurs in concavity-based methods that heavily rely on the clumped nuclei contour concavity [5], [6], [7]. By contrast, existing watershed-based methods usually yield over-segmentation due to the variance of image intensity [8], [9]. A 3D method is developed to segment clumped nuclei for improved segmentation accuracy [8]. However, its computational cost is expensive.

We propose a new algorithm that fully exploits the boundary shape of nuclei clumps to identify connecting point pairs by local high curvature voting and point pair screening. This is followed by a novel method that connects point pairs with dividing curves derived from local shape-based intensity analysis. Our method shows significant improvements over an existing marker-controlled watershed method [9] on a set of 560 fluorescence microscopy image patches.

II. METHOD

Our algorithm utilizes the contour shape of clustered nuclei to generate candidate point pairs and local shape based intensity analysis to separate overlapped nuclei with dividing curves. The work-flow consists of four main steps: (1) Voting for high curvature candidate points; (2) Screening close point pairs by a curvature-based distance metric for adjacent point pairs and a comprehensive metric taking into account curvature, distance, and matched normal vector angle for non-adjacent point pairs; (3) Assessing adjacent point pair connections via an ellipse fitting quality indicator; and (4) Identifying dividing curves by local shape based intensity analysis (see Fig. 1).

Fig. 1:

Fig. 1:

The overall method flowchart for our proposed clumped nuclei segmentation is demonstrated. Nuclei are shown in blue with additional nuclear components shown as red and green interior regions.

A. High Curvature Candidate Point Voting

The blue-fluorescent DAPI nucleic acid stain preferentially binds to DNA within nuclei. Thus, we focus on the blue image channel for overall nuclear contour detection. We obtain the boundaries of such clustered nuclei using a global adaptive threshold. We propose a new method to find high curvature points on clump contours based on neighboring high curvature point votes. Curvature value of each point on the boundary is computed by k=xyyx(x2+y2)3/2, where x and y are coordinates of points on the denoised contour convolved with a smoothing Gaussian filter [2]. Different from high curvature point detection methods that select points with curvature values higher than a specified threshold [2], our method can avoid redundant candidate point detection with local high curvature point votes. As the following segmentation analysis depends on these initial candidate points, limiting the number of redundant candidate points can substantially reduce the computational cost and improve segmentation accuracy. In our data, it is quite common to have a long set of high curvature points along a concave contour segment defined by the curvature sign, as demonstrated in Fig. 1(I). For those high curvature points in spatial proximity, it is unreasonable to mark all these points as candidates. Thus, we propose to vote for the optimal candidate point representation for a group of adjacent high curvature points by:

s*=Ck(s)sdsCk(s)ds (1)

where s ∈ [0, 1] represents the normalized boundary arc length on a cell concave contour segment ∂C; κ(s) is the curvature value at s; s* is the normalized arc length of the optimal point representation for a segment of adjacent concavity points engaged in local voting. The bottom image in Fig. 1(I) demonstrates the final high curvature point representation (in yellow circle) voted by all local high curvature points detected in the top image in Fig. 1(I). With this procedure, we detect all high curvature point representations on each contour.

B. Close Point Pair Screening

Let us denote P = (pi, |i = 1, 2, …,N as a set of high curvature candidate points detected from Subsection II A. As demonstrated in Fig. 1(II), we next identify all pairs of points (pi, pj) in spatial proximity by two circular searching regions of radius r1 and r2 (r1 < r2) for under-segmentation reduction. Specifically, r1 is used to detect the pairs of both adjacent and non-adjacent points, whereas the ring area defined by r2 − r1 is for detecting pairs of non-adjacent points in a neighboring area. In our analysis, they are set to 45 pixels and 70 pixels, respectively. In this way, we can alleviate under-segmentation suffered by numerous shape-based methods. The spatial proximity in our study is measured by the Euclidean distance. In addition, unlike the “bottleneck” points defined as close points presenting opposite gradient directions in previous research [3], [4], we relax this constraint and classify close points as either adjacent or non-adjacent point pairs. Plotted in the left panel in Fig. 1(II), the search regions for close point pairs are represented by the red circle with radius r1 and the green ring with inner radius r1 and outer radius r2, respectively. Here we use the searching region in the red circle of radius r1 for detecting pairs of both adjacent and non-adjacent points in spatial proximity, whereas we only use the green ring for detecting pairs of non-adjacent points in a neighboring area. In the left plot of Fig. 1(II), points 9 and 11 in the red circle region are close adjacent points of point 10, while point 5 in the green area is also close but not adjacent to point 10. Although these two types of close point pairs are processed in different ways in our proposed method, they both are important for an accurate segmentation.

(1). Analysis for Pairs of Adjacent Points in Proximity:

Let us denote C+ = {(pi, qi)|i = 1, 2, … , N1} as the set of neighboring adjacent point pairs given the searching circular area of radius r1. For each such pair demonstrated in the right panel of Fig. 1(II), we evaluate the associated Walking Energy to determine if we merge them (i.e. discard the undesired point and keep the point with the higher curvature value). Walking Energy is defined as Ep,q=p(s)q(s)k(s)ds, where κ(s) is the curvature of the corresponding curve segment s ∈ [p(s), q(s)]. Walking Energy represents the amount of ”effort” required for a walk from a point p(s) to its close adjacent point q(s). In our formulation, we assume it consumes more energy to walk along a convex than non-convex contour. Demonstrated in the right panel of Fig. 1(II), point 6 has two close point neighbors, i.e. point 5 and 7. The true walking route from point 6 to 7 is illustrated in a red curve, while that from point 6 and 5 is shown in green. Those pairs of neighboring adjacent points requiring low Walking Energy are merged. For each such pair of points under investigation, the point with higher curvature value is retained after points get merged. Adjacent points associated with a high Walking Energy would be retained, as shown in the right panel of Fig. 1(II) where the Walking Energy for points 6 and 7 is sufficiently high to keep them separated. For any pair of such adjacent points presenting a high Walking Energy, it is highly likely that such points should be connected to separate overlapped nuclei. Therefore, we further evaluate the boundary segmentation quality in such cases by an ellipse fitting method discussed in Subsection II-C.

(2). Analysis for Pairs of Non-Adjacent Points in Proximity:

The resulting set of non-adjacent point pairs in close proximity can be represented as C1={(pi,qi)|i=1,2,,N2} when we use the circular searching area of radius r1. To alleviate under-segmentation, we next expand the searching space to a ring searching region with radius between r1 and r2. To assess if such pairs of non-adjacent points (pi, qi) result from intersection of overlapped nuclei, we formulate and use the following measure to evaluate such point pairs:

V(pi,qi)=αθ(pi,qi)D(pi,qi)+β(k(pi)+k(qi)) (2)

where θ is the angle between two normal vectors of the nuclear contour at pi and qi; D is the Euclidean distance between (pi, qi); α and β are weights set as 100 and 0.34, respectively. If V (pi, qi) is larger than a specified threshold (e.g. 200 in our analysis), and (pi, qi) are non-adjacent points in a ring searching space between the radius of r1 and r2, such point pair is added to C2={(pi,qi)|i=1,2,,N3}. After this evaluation analysis, the finalized set of non-adjacent point pairs in close proximity is C=C1C2. For each pair in C, we next link the associated points to partition the whole contour into multiple subsets. Demonstrated in Fig. 1(III), a whole nuclei cluster can be partitioned into multiple subsets by connecting points (2, 15), and (5, 10). As each candidate subset can produce new possible adjacent point pairs, the method for pairs of adjacent points in proximity discussed above is applied to each candidate subset.

C. Point Pair Segregation Assessment by Ellipse Fitting

We proceed with evaluating the segregation quality Qi,i+1 of each adjacent point pair (pi, pi+1) by ellipse fitting. For each sub-contour segment under evaluation, an optimal ellipse, shown in red in Fig. 1(IV), is fit to a candidate nucleus region with a closed sub-contour consisting of a contour arc and a dividing line between (pi, pi+1). An ellipse fitting quality measurement is defined as:

Qi,i+1=μSi,i+1++vψi,i+1(Δxi,i+1+Δyi,i+1)+γ1ΔLi,i+1+γ2ηi,i+1 (3)

where Si,i+1+ is the ratio of overlapped area between the candidate nucleus and fitting ellipse to their union; ψi,i+1 is the fitting angle formed by the line segment connecting the ellipse center and pi and that connecting the ellipse center and pi+1; Δxi,i+1 and Δyi,i+1 are the centroid coordinate difference between the fitting ellipse and the evaluated nucleus region; ΔLi,i+1 is the perimeter difference between the fitting ellipse and the evaluated nucleus contour; ηi,i+1 is the ratio of major to minor axis length derived from the fitting ellipse. μ, ν, γ1 and γ2 are weights set to 10.70, 10.70, 0.67 and 3.40 in our experiment.

With the formulated ellipse fitting quality measurement, we connect those neighboring adjacent point pairs when the associated Q is larger than a threshold (e.g. 0.7 in our analysis). This ellipse fitting evaluation process is carried out in an iterative manner. Each time, we connect the pair with the largest Q. In the post-processing step, we prune intersected connections and dividing lines that form a sharp angle to avoid over-segmentation. Finally, we obtain pairs of points to be connected in a set C*={(pi*,qi*)|i=1,2,,NC}.

D. Identification of Dividing Curves

Given identified point pairs for connection, a vast majority of nuclei segmentation methods connect each such pair with a straight separating line. However, it is our observation that a better division is given by a curve following the image gradient information, as shown in Fig. 1(V). Therefore, we next aim to recover a dividing curve for each such pair with local shape-based intensity analysis. Denoting I(x, y) as the fluorescence microscopy image, we can represent it by Taylor series expansion as:

I(x,y)I(x0,y0)+I(x0,y0)[xx0yy0]+12[xx0yy0]HI(x0,y0)[xx0yy0] (4)

where HI(x0, y0) is the Hessian matrix representing the second derivative values of I at the specific point (x0, y0). In order to have a better dividing path than a straight line connecting two contour points (p, q), we start off from one end point p and look for the next adjacent pixel with its local neighboring intensity change almost zero in one direction and drastically increased in its orthogonal direction. As a natural dividing curve passes along pixels with local intensity minimum, we can detect such adjacent pixels by computing two eigenvalues of Hessian matrix HI satisfying such properties: 0 ≈ λ1λ2. The adjacent pixel with a near zero λ1 and the maximum λ2 is considered as the next pixel on the dividing curve.

To ensure that the final optimized connection path can merge to the other end point q smoothly, we search forward in a targeting region with a constrained deviation angle (i.e. ±45° in our experiment) from the vector connecting the current dividing point to q at each step. Within such a sector-shaped searching space, we can force the resulting optimal connection path to remain in intensity “valley” with large gradient variance and to converge smoothly to the other end point q. This is demonstrated by the connection between point 5 and 9 in Fig. 1(V) where our method is able to recover a dividing curve along a local minimum intensity route for overlapped nuclei segmentation.

III. EXPERIMENTAL RESULTS

To validate the performance of the proposed algorithm, we tested our method with a dataset of 112 images (1024 × 1376 pixels) from two brain tumor regions of seven patients. As such images contain a large number of nuclei that are not overlapped, five image patches with overlapped nuclei clusters are randomly selected from each image. This results in 560 image patches for method evaluation.

We visually inspect the resulting nuclei segmentation results and compare our method with the marker-controlled watershed method [9]. Representative results of four overlapped nuclei clusters in our dataset are presented in Fig. 2. We present the instances of clumped nuclei from original images in the first row, and ground-truth annotations on the second. Nuclei segmentation results from marker-controlled watershed segmentation and our proposed method are presented in the third and forth rows, respectively. It is noticed that our method is able to produce segmentation results similar to human annotations. In particular, those dividing curves recovered from local shape based intensity analysis can help improve segmentation results. By contrast, watershed method fails to identify and separate some overlapped nuclei clusters in the third row. To quantitatively evaluate the performance of our proposed method, five metrics including Jaccard index, Precision, Recall, F1 score, and Hausdroff distance are used. The resulting quantitative evaluation results are presented in Table I. For each patient, both marker-controlled watershed method and our proposed method are tested on 40 images of Bulk Tumor (BT) and another 40 of Tumor Margin (TM). There are four rows of results associated with each patient in Table I. The top two lines present quantitative results of watershed based method on BT and TM images, while the bottom two demonstrate results from our method on BT and TM images. Note that our method presents superior performance as measured by most of the metrics. Specifically, Jaccard index values from our proposed method are much higher than those of watered based method while Watershed method has better Precision as it tends to miss a large number of nuclei segregation events. One limitation of our method is that it could fail for cases where overlapped nuclei result in internal holes, as it uses contours of clustered nuclei. This will be improved in future work.

Fig. 2:

Fig. 2:

Comparisons of segmentation results over four representative overlapped nuclei regions are presented with original images, ground truth, results of marker-controlled watershed method, and outcomes of our proposed method from top to bottom rows, respectively.

TABLE I:

Quantitative method evaluation and comparison with 560 image patches from BT and TM tissue regions: (mean,std).

Patient Jaccard Precision Recall F1 score HD (pix.)
1 0.44,0.19 0.96,0.08 0.46,0.21 0.58,0.22 34.3,19.1
0.38,0.18 0.97,0.07 0.40,0.20 0.53,0.20 34.6,18.7
0.75,0.20 0.86,0.18 0.86,0.17 0.84,0.17 22.0,25.6
0.81,0.14 0.93,0.13 0.87,0.11 0.89,0.11 15.0,19.6
2 0.60,0.16 0.96,0.11 0.62,0.16 0.73,0.16 20.2,15.5
0.64,0.16 0.98,0.07 0.66,0.16 0.77,0.13 24.3,17.9
0.80,0.15 0.90,0.13 0.89,0.12 0.88,0.11 13.8,18.4
0.83,0.15 0.91,0.12 0.91,0.12 0.90,0.11 13.6,20.1
3 0.49,0.18 0.94,0.11 0.52,0.21 0.63,0.19 32.2,23.5
0.54,0.17 0.97,0.10 0.55,0.17 0.68,0.15 31.7,20.3
0.70,0.23 0.89,0.14 0.79,0.25 0.80,0.21 21.7,24.3
0.77,0.17 0.89,0.11 0.85,0.16 0.86,0.13 13.8,18.2
4 0.44,0.18 0.94,0.13 0.46,0.20 0.59,0.20 33.6,22.0
0.47,0.15 0.97,0.09 0.48,0.16 0.62,0.16 34.5,17.9
0.75,0.18 0.80,0.19 0.93,0.11 0.84,0.14 27.1,29.3
0.80,0.19 0.87,0.15 0.92,0.16 0.88,0.15 22.0,29.0
5 0.57,0.16 0.95,0.10 0.59,0.17 0.71,0.15 28.9,19.3
0.49,0.16 0.97,0.10 0.50,0.16 0.64,0.17 36.4,19.6
0.80,0.17 0.85,0.16 0.93,0.13 0.88,0.14 21.4,30.8
0.86,0.12 0.90,0.10 0.95,0.09 0.92,0.08 15.5,20.4
6 0.51,0.19 0.93,0.13 0.55,0.21 0.65,0.19 26.6,19.5
0.67,0.20 0.93,0.17 0.71,0.20 0.78,0.19 22.2,24.7
0.71,0.21 0.84,0.18 0.84,0.20 0.81,0.17 23.8,25.4
0.82,0.14 0.88,0.13 0.93,0.09 0.90,0.10 13.1,18.1
7 0.58,0.20 0.95,0.11 0.61,0.21 0.71,0.19 33.3,25.0
0.44,0.20 0.96,0.08 0.45,0.21 0.58,0.22 33.0,18.7
0.78,0.18 0.86,0.14 0.91,0.16 0.86,0.14 19.2,30.3
0.78,0.17 0.86,0.15 0.90,0.14 0.86,0.14 18.5,26.1

IV. CONCLUSIONS

This paper presents a novel segmentation method for clumped nuclei in fluorescence microscopy images. Our analysis first generates precise candidate point representations based on a high curvature point voting method, followed by detection of connecting point pairs based on spatial proximity, shape convexity, and curvature information via close point pair screening. An ellipse model is proposed to fit to each resulting area associated with each point pair candidate. We define a fitting quality as a function of fitting angle, fitting area, fitting ellipse center shift, fitting perimeter change, and fitting area elongation. Only point pairs presenting good fitting quality are connected. Instead of separating overlapped nuclei by a straight line, we recover dividing curves by local shape based intensity analysis in a sector-shaped searching space. We validate our algorithm with 560 image patches from two classes of tumor regions associated with seven brain tumor patients. Both qualitative and quantitative validation results suggest that our algorithm is promising for dividing overlapped nuclei in fluorescence microscopy images widely used in various biomedical research.

Acknowledgments

Funded by NIH K25CA181503, and CNPq. The studies involving human subjects were approved by the Emory University IRB.

Contributor Information

George Teodoro, University of Brasília, Dept. of Computer Science, Brasília, DF, Brazil.

Daniel Brat, Northwestern University, Dept. of Pathology, Chicago, IL 60611.

Jun Kong, Emory University, Dept. of Biomedical Informatics, Atlanta, GA 30322.

REFERENCES

  • [1].Reljin N,Slavkovic-Ilic M, Tapia C,et al. ,Multifractal-based nuclei segmentation in fish images. Biomedical microdevices, 19(3):2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Wen Q, Chang H, and Parvin B, A Delaunay triangulation approach for segmenting clumps of nuclei. IEEE International Symposium on Biomedical Imaging: From Nano to Macro, pp. 9–12, 2009. [Google Scholar]
  • [3].Wang H, Zhang H and Ray N, September. Clump splitting via bottleneck detection. IEEE International Conference on Image Processing, pp. 61–64, 2011. [Google Scholar]
  • [4].Liao M, Zhao YQ, Li XH, Dai PS, Xu XW, Zhang JK and Zou BJ, Automatic segmentation for cell images based on bottleneck detection and ellipse fitting. Neurocomputing, 173, pp.615–622, 2016. [Google Scholar]
  • [5].Fouad Shereen, et al. ”Morphological Separation of Clustered Nuclei in Histological Images” International Conference Image Analysis and Recognition. Springer, Cham, 2016. [Google Scholar]
  • [6].Zhang C, Sun C, Su R and Pham TD, Segmentation of clustered nuclei based on curvature weighting. Conference on Image and Vision Computing New Zealand, pp. 49–54, 2012. [Google Scholar]
  • [7].Zhang C, Sun C, and Pham Tuan D.. ”Segmentation of clustered nuclei based on concave curve expansion.” Journal of microscopy 251.1 (2013): 57–67. [DOI] [PubMed] [Google Scholar]
  • [8].Atta-Fosu Thomas,et al. ”3D Clumped Cell Segmentation Using Curvature Based Seeded Watershed.”Journal of imaging 2.4(2016):31. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Bartell LR, Bonassar LJ and Cohen I,A watershed-based algorithm to segment and classify cells in fluorescence microscopy images. arXiv:1706.00815, 2017. [Google Scholar]

RESOURCES