Abstract
We present a fully automatic algorithm to identify fluid-filled regions and seven retinal layers on spectral domain optical coherence tomography images of eyes with diabetic macular edema (DME). To achieve this, we developed a kernel regression (KR)-based classification method to estimate fluid and retinal layer positions. We then used these classification estimates as a guide to more accurately segment the retinal layer boundaries using our previously described graph theory and dynamic programming (GTDP) framework. We validated our algorithm on 110 B-scans from ten patients with severe DME pathology, showing an overall mean Dice coefficient of 0.78 when comparing our KR + GTDP algorithm to an expert grader. This is comparable to the inter-observer Dice coefficient of 0.79. The entire data set is available online, including our automatic and manual segmentation results. To the best of our knowledge, this is the first validated, fully-automated, seven-layer and fluid segmentation method which has been applied to real-world images containing severe DME.
OCIS codes: (100.0100) Image processing, (170.4500) Optical coherence tomography, (170.4470) Ophthalmology
1. Introduction
Diabetic retinopathy is the leading cause of blindness among working-aged adults in the United States [1]. Among those affected, approximately 21 million people develop diabetic macular edema (DME) [2]. DME results from the breakdown of the blood-retinal barrier as retinal vascular endothelial and mural cells become damaged from chronic hyperglycemia [3]. Retinal hypoxia results in increased production of vascular endothelial growth factors (VEGF) and other signaling cascades [3–5]. These further progress DME through mechanisms such as cytotoxic damage to retinal fluid transport cells [6–9]. Ultimately, the imbalance between vascular leakage and fluid transport leads to retinal edema and vision loss.
While clinical trials of several treatment modalities have been conducted for therapies for DME [10–19], the significant proportion of subjects that fail to respond to any single therapy suggests that DME pathophysiology is multifactorial, and unfortunately no consensus yet exists for determining which patients are likely to respond to specific therapies. This may be due to the absence of a standard method for stratifying patients based on disease mechanisms [20, 21]. In recent years, the additional depth-resolved dimension of data provided by optical coherence tomography (OCT) imaging has prompted groups to correlate morphological patterns of the retina on OCT with DME and vision outcomes [22–26]. For clinical trials, the most commonly used quantitative imaging biomarker of DME severity is currently central subfield thickness, which does not capture details such as edema volume or changes in specific retinal layers. We expect that these parameters will provide valuable prognostic information to guide treatment decisions. However, despite the development of automated retinal layer segmentation methods applicable to normal eyes or eyes with limited pathological deformation or loss of retinal structures [27–42], most DME studies require either manual [22, 24–26, 43] or semi-automatic [44] evaluation of OCT images. Currently few automated algorithms exist to quantify morphological or pathological features on images with DME [43, 45–51].
Furthermore, while averaging-based denoising methods enhance image quality and improve the performance of automated image segmentation algorithms [52], they also result in drastic increases in image acquisition times, which may not be tolerated by ill or uncooperative patients, especially when capturing dense volumetric scans. It is therefore desirable to develop automatic OCT segmentation algorithms which can be applied to lower quality images captured using fast-image scanning protocols, as these are often required in real-world, clinical settings.
Finally, while a few state-of-the-art algorithms provide combined fluid and layer segmentation methods [47, 48, 50], to the best of our knowledge, no fully-automated algorithm has been validated to identify all retinal layers and fluid-filled regions. Furthermore, there is limited rigorous analysis of layer-specific changes in retinal thickness following therapy. As such, the selection of personalized therapeutic strategies for patients remains subjective. As a step towards this goal, we present a fully automated algorithm based on kernel regression (KR) [53] and our previously described graph theory and dynamic programming (GTDP) framework [46, 54, 55] to identify fluid-filled regions and seven retinal layers on spectral domain (SD)-OCT images of eyes with DME. In the first phase of developing our algorithm, we created KR-based classifiers to estimate fluid and retinal layer positions. We then used these classification estimates in the second phase to guide GTDP segmentation for a more accurate result.
In Section 2, we provide an introduction to KR, and in Section 3 we present our new general purpose KR-based classification method. We then combine our KR-based classification method with GTDP segmentation in Section 4 and describe our implementation on OCT images of eyes with DME in Sections 5 and 6. Finally, we validate the performance of our algorithm in Sections 7 and 8.
2. Kernel regression review
KR is a non-parametric method for deriving local estimates of a function using a kernel that weighs the relative importance of nearby points [53]. While traditional KR-based applications include image denoising and interpolation [53], deblurring [56], and object detection [57], we use KR to improve image classification and segmentation. In this section, we provide an overview on the second order iterative Gaussian steering KR method for image denoising as introduced and described in detail in [53].
2.1 Problem formulation
Looking at a small patch of pixels on a noisy image the intensity of a pixel can be described by
(1) |
where is the pixel located at row and column on , is an unspecified (non-parametric) regression function, and is an independent and identically distributed zero-mean noise value at . Using this formulation, we would like to obtain a local estimate of the regression function to generate a denoised image can approximate using the second order Taylor series expansion about a point near following
(2) |
We next solve for the unknowns and in Eq. (2) using the weighted linear least squares formulation [53, 58] following
(3) |
and “diag” defines a diagonal matrix and is a linearly normalized kernel function used to weigh the observations such that . Since Eq. (3) is a linear least squares formulation, the solution can be given by Eq. (4), where the first element in is the denoised image
(4) |
2.2 Adaptive iterative Gaussian steering kernel
In order to perform KR as described in Section 2.1, an appropriate kernel must be selected. In this section, we describe the popular adaptive iterative Gaussian steering kernel [53]. To summarize, this kernel can be described as a Gaussian function that is elongated , rotated , and scaled based on the local data to reduce edge blurring.
To compute the Gaussian steering kernel for a pixel we first smooth the image using a linear Gaussian filter with standard deviations of and along the and dimensions, and then we compute the image gradients and We then calculate the truncated singular value decomposition following Eq. (5) [53, 59, 60].
(5) |
Second, we compute the steering parameters and as described in Eq. (6), where and are parameters that prevent undefined or zero values for and We then denoise following Eq. (4) using a linearly normalized version of the kernel defined Eq. (7), where is the global smoothing parameter.
(6) |
(7) |
Third, we iteratively improve the denoised image ( ) by recalculating the steering parameters and kernel using the denoised image gradients ( ) [53].
An example image and its Gaussian steering KR denoised result are shown in Fig. 1(a) and 1(f), respectively. Figure 1(c) and 1(e) are the iterative Gaussian steering kernels for Fig. 1(b) and 1(d), respectively. The kernel in Fig. 1(c) is wider and more isotropic since its corresponding image patch does not contain strong edges, while the kernel in Fig. 1(e) is narrower and oriented along the image edge. In Fig. 1(f), the regions external to the retina (black) were not denoised with the exception of a padded boundary surrounding the retina. The thickness of this boundary is half the kernel size to encapsulate all pixels required for feature vector computation in Sections 3.2 and 5.8. Parameter values used to generate Fig. 1(f) were chosen empirically using the training data set and include a kernel size of 73.7 × 147.4 µm (11 × 11 pixels, = 121), 3 iterations for , , , , , and
3. Kernel regression-based classification
In this section, we integrate KR with supervised classification [61, 62] to classify noisy images. The following subsections describe our novel, general purpose KR-based classification method. Specific application of this method for segmenting retinal layers is discussed in Sections 5 and 6.
3.1 Steering kernel regression estimation
Since pixels within an image are individually classified, the presence of noise worsens classification results. Thus, the first step is to use the KR method described in Eq. (4) to denoise the image and obtain: 1) direct estimates for the noise-free first and second derivatives ( and ), and 2) adaptive directional kernels for each pixel.
3.2 Computing features
To classify a pixel , the pixel needs associated descriptors called features. Each feature is a scalar value used to describe the pixel, and features comprise a feature vector Basic features include pixel intensity, gradient, and location. Other feature types include textural features such as Laws’ Texture Energy Measures [63, 64].
To mitigate the effects of noise on feature computation, we perform weighted averaging. Since the steering kernels in Section 2.2 adapt to the shape of the underlying structure, we use the kernels from Section 3.1 to filter the features following
(8) |
where is the feature for pixel and is the (now denoised) feature for pixel We also use the kernels themselves as features (e.g. kernel height, width, and area) to reveal more indistinct information. Specific feature examples are defined in Section 5.8.
3.3 Defining the true (training) classes
In supervised classification, a set of classes (categories) is pre-defined for a given classifier [62]. Given a training data set, we manually identify the true class for each training pixel such that Specific examples of these classes are given in Section 5.2.
3.4 Defining the classifier function
For an image we define a classifier which estimates the class for a pixel following
(9) |
Provided that the training data set is composed of total training pixels, we first combine the feature vectors for all pixels into the feature vector set where is the feature vector for the training pixel. For each class , we then create a feature vector subset where and is the number of training pixels with as the true class. From this definition, it can be seen that .
We next define the classifier function used to estimate the class. To take into consideration the typical spread of feature values, we use a weighted negative Gaussian classifier function following Eq. (10), where is the mean feature vector across the training feature vector subset is the element-wise reciprocal of the variance feature vector for is a set of constant weights used to alter the relative importance of features, and is the Hadamard product, an element-wise multiplication operation for vectors and matrices. Selection of these weights is described in Section 3.5.
(10) |
3.5 Weighted sequential forward feature selection
Since not all features in may be useful for predicting the true class , we reduce the dimension of by defining a subset of length containing the most relevant features [65–67]. We achieve this by performing weighted sequential forward feature selection (wSFFS), a variation of the classic SFFS method [66] that simultaneously selects features and their corresponding weights. In the classic SFFS method, begins as a null set and features are added to it sequentially [66] by minimizing a predefined criterion function (e.g. misclassification rate).
A limitation of the SFFS method is that it weighs each feature equally. As a result, we use the wSFFS method to select features along with their optimal weight. Given a set of possible weight values we create a new set of weighted feature vectors We then perform SFFS on to find the most relevant weighted feature. After sequentially adding weighted features, the result is a subset of features and corresponding weights where the weight We subsequently use this new set of reduced features and weights in Eq. (10) to classify the pixel in place of the full, non-weighted feature vector
Finally, to reduce classifier overfitting, we use K-fold cross-validation [68] during feature selection. We also increase the classification accuracy by using K-means clustering [69] to divide into clusters. We then perform wSFFS on each of the clusters to generate sets of selected feature vectors and weights. When classifying a pixel, the set that minimizes the criterion function is used to determine the final class.
4. Classification-based GTDP segmentation
To improve the KR-based classification method from Section 3, we apply our GTDP segmentation framework introduced in [54] to more accurately isolate class boundaries. This section provides an overview of our GTDP method and describes how the classification results can be used to both improve and simplify GTDP segmentation.
4.1 GTDP layer segmentation review
We represent an image as a graph, where each pixel is a node on the graph and adjacent nodes are connected by edges. To create a path preference, we assign positive weights to the edges based on features such as image intensity or gradient, where ultimately the lowest weighted path is preferred. For example, to segment the darkest path across an image, we assign weights based on image intensity, where the weight of the edge connecting nodes and following
(11) |
where is the intensity of at node and is a positive number specifying the minimum weight of an edge since a weight equal to zero indicates two unconnected nodes. We also specify that for all node pairs, as this generates an undirected graph with half as many weight values.
To segment a horizontal layered structure, we add an additional node to either side of the graph to automatically initialize the start and end nodes of our path. The additional node on the left is connected to all nodes in the leftmost column of the graph, and the right additional node has edges connecting it to all nodes in the rightmost column. These additional edges are assigned a weight of We then apply Dijkstra’s algorithm [70] to find the shortest path from the start node to the end node, which corresponds to our desired segmentation. For more details referring to our GTDP segmentation algorithm, refer to [54, 55].
4.2 Classification-based search region limitation
For scenarios in which multiple similar features are to be segmented within a given image (e.g. retinal layer boundaries), the graph may be isolated to a smaller region containing only one of the features. This may be done using a priori knowledge of the feature to segment. For example, to segment the inner plexiform layer (IPL) / inner nuclear layer (INL) boundary, we can set all edge weights in the outer half of the retina to zero. This prevents the algorithm from segmenting outer retinal features such as the outer segment (OS) / retinal pigmented epithelium (RPE) boundary. Implementation examples of these types of search region limitation can be found in [45, 46, 54, 55, 71].
To simplify and reduce the complexity of search region limitation, we can modify the edge weights of the graph to emphasize a particular region. For example, let us extend our example in Section 4.1 to segment the darkest path across an image. Given an image and its classification result where is the estimated class of , we would like to segment the darkest path across the portion of the image where To do this, we compute the edge weights for following
(12) |
where is the class at node and represents the number of nodes not equal to From Eq. (12), it can be seen that edges connecting nodes with class are given lower weights than edges connecting nodes assigned to other classes. The advantage of this method is that far fewer a priori rules are required to isolate the desired region, as the classification method automatically generates estimates of region locations. Furthermore, hard limits are not set on the search region, as portions of the graph are not excluded.
5. Training the DME classifier
In this Section, we describe an implementation of our KR-based classification method to generate a DME classifier to identify fluid-filled regions and seven retinal layers on SD-OCT images of eyes with DME. An outline of our method is shown on the left side of Fig. 2 . All parameter values were either selected empirically based on the training data set or were extensions from our previous segmentation algorithms [45, 54, 55].
5.1 DME training data set
To learn our DME classifier, we obtained training data separate from our validation data set. We used the Duke Enterprise Data Unified Content Explorer search engine to retrospectively identify patients within the Duke Eye Center Medical Retina practice with a billing code for DME (ICD-9 362.07) associated with their visit. An ophthalmologist then identified six patients imaged in clinic using the standard Spectralis (Heidelberg Engineering, Heidelberg, Germany) 61-line volume scan protocol with severe DME pathology and varying image quality. Averaging of the B-scans was determined by the photographer, and ranged from 9 to 21 raw images per averaged B-scan. The volumetric scans were 61 B-scans × 768 A-scans with an axial resolution µm/pixel, lateral resolution ranging from 11.07 – 11.59 µm/pixel, and azimuthal resolution ranging from 118 – 128 µm/pixel.
5.2 Target classes
To generate the target classes for classifier training, we manually segmented fluid-filled regions and semi-automatically segmented all eight retinal layer boundaries following the definitions in Fig. 3 . This was done for 12 B-scans within the training data set (two from each volume). The B-scans selected consisted of six images near the fovea (B-scan 31 for all volumes) and six peripheral images (B-scans 1, 6, 11, 16, 21, and 26, one for each of the six volumes). We then used the manual segmentations to assign the true class for each pixel, with a total of eight possible classes defined in Table 1 and the classified result shown in Fig. 4(a) .
Table 1. DME classes.
Class | Description | Color |
---|---|---|
Fluid | Blue | |
NFL | Light Blue | |
GCL-IPL | Cyan | |
INL | Green | |
OPL | Yellow | |
ONL-ISM | Orange | |
ISE | Red | |
OS-RPE | Dark Red |
5.3 Pre-processing the image
Prior to training the classifier, we perform image pre-processing steps for all B-scans within a volume following our validated DME algorithm described in [45]. To summarize, we first remove Spectralis registration boundaries by replacing them with the mirror image of the valid image regions. To standardize the images following [45] and [55], we resize the image to a lateral and axial resolution of µm/pixel and µm/pixel, respectively, and then linearly normalize the image to range from 0 to 1.
5.4 Flattening the retina
Next, we flatten the curved retina to obtain a shorter path for GTDP segmentation. Following [45], we achieve this by first heavily filtering each image using a 40.2 µm × 80.4 µm Gaussian with µm × 147.4 µm. We then threshold the filtered image at 0.5 to extract the inner and outer hyper-reflective complexes. We refine the hyper-reflective mask by 1) opening any gaps in the clusters using a 33.5 µm × 40.2 µm rectangular structuring element, 2) removing all clusters less than 0.018 mm2 in size, and 3) closing any remaining gaps using the same structuring element. Using GTDP, we segment the boundaries of the hyper-reflective complexes as described in [55] which obtains estimates at the middle of the ONL and RPE.
Unlike our previous algorithms which flattened the retina on B-scans individually [45, 54, 55], in this algorithm we perform the above steps for all images in the volume to generate 2D map estimates ( ) of the OPL and RPE surfaces. Using the map of the RPE, we linearly interpolate any missing A-scans and B-scans. We then search for outliers by computing the absolute second derivative in the vertical direction and isolating all values greater than three standard deviations above the mean. Outlier clusters within each B-scan are then linearly interpolated, and the map of the RPE is smoothed using a 500 µm Gaussian filter with µm. Finally, for each B-scan, we circularly shift the A-scans until the RPE estimate is flat. Since the ONL boundary no longer aligns with the flattened RPE and image, we also flatten the ONL using the same flattening transformation.
5.5 Segmenting the total retina
To isolate the retina from the vitreous and choroid, we implement our GTDP layer segmentation framework described in Section 4.1 to segment the ILM using the dark-to-light gradient weights computed following
(13) |
where is the maximum of the dark-to-light vertical (filter ) and diagonal gradients of at node after filtering the image using a Gaussian (20.1 µm × 67 µm, µm × 26.8 µm), and is a linear normalization of to range from to While using Dijkstra’s algorithm to segment the ILM, we limit the search region to span from the top of the image down to the RPE estimate generated in Section 5.4. To ensure that hyper-reflective spots or detachments internal to the ILM are not segmented, we re-segment the boundary using the weights in Eq. (14) where is the Euclidean distance from node to node We also limit the search region to span from the initial ILM estimate to 67 µm below the ILM.
(14) |
To segment BM, we compute weights following Eq. (15) where is the light-to-dark equivalent of and is the intensity of the filtered image at node . We also set the top boundary of the search region to the mean location between the ONL estimate and 100.5 µm below the ILM, and we set the bottom boundary to the bottom of the image.
(15) |
5.6 Locating the fovea
To compute the features in Section 5.8 and gauge the location of the optic nerve and merged retinal layers, we first identify the location of the fovea on the volumetric macular scan. Upon segmenting the ILM and BM across all B-scans, we locate the fovea based on the 2D total retinal thickness (TRT) map computed by subtracting the ILM from BM. To do this, we first filter the TRT map ( ) using a Gaussian (120 µm × 89.8 µm and σ = 18.95 µm × 85.8 µm). We then narrow our search region to the central 3 mm of the volumetric macular scan and find all local extrema with a depth greater than 13.4 µm. If the extremum with the lowest or highest TRT resides closest to the center of the volume, then we assign that location as the fovea. Otherwise, we create a summed voxel projection (SVP) by summing pixels ranging from the ILM to 33.5 µm below the ILM across all B-scans. As with the TRT, we smooth the SVP, locate the extrema, and computed distances from the fovea. Looking at the five extrema closest to the center of the volume, we set the extremum with the lowest intensity as the fovea.
5.7 Kernel regression-based denoising
We next perform iterative Gaussian steering KR on each training image as described in Section 2 to 1) extract the normalized kernels for every pixel in the retina, and 2) recover defined in Eq. (4) which contains the denoised image and its first and second order derivatives. Parameter values are listed in Section 2.2, and since only the pixels within the retina need to be classified, we do not denoise pixels within the vitreous and choroid.
5.8 Computing DME training features
To learn the DME classifier, we computed feature vectors for the 12 training images manually segmented in Section 5.2. We generated a total of 22 features ( ) for each vector, and only pixels within the retina were processed. The nine KR-based features are defined in Table 2 , the four position-related features are listed in Table 3 , and the nine Laws’ Texture Energy Measure features [63, 64] are listed in Table 4 . We then enhance the texture features in Table 4 using Eq. (8) and further enhance all features except using adaptive histogram equalization [72]. Finally, we normalize all features to match the range of the training features. Figure 5 shows Laws’ feature computed with and without KR-based denoising. As can be seen from Fig. 5, utilizing KR reduced the amount of noise present within the feature.
Table 2. Kernel regression-based DME training features.
Definition | Description |
---|---|
KR denoised pixel | |
| |
KR denoised vertical first derivative | |
| |
KR denoised horizontal first derivative | |
| |
KR denoised vertical second derivative | |
| |
KR denoised diagonal second derivative | |
| |
KR denoised horizontal second derivative | |
| |
Horizontal kernel full width at half max | |
| |
Vertical kernel full width at half max | |
| |
Kernel half max area |
Table 3. Position-based DME training features.
Definition | Description |
---|---|
Horizontal distance from the fovea, where is the foveal A-scan | |
| |
Vertical distance from the fovea, where and are the vertical positions of the ILM and BM for column determined in Section 5.5. | |
| |
Radial distance from the fovea, where and are the lateral and azimuthal resolutions, and is the B-scan number of the image within the volumetric scan | |
| |
Retinal thickness |
Table 4. Laws’ texture training features.
5.9 Selecting relevant features and weights
To select the most relevant features and weights for fluid and layer classification, we implemented our wSFFS method described in Section 3.5. We used ten-fold cross-validation [73] during this process and adjusted selection parameters until we achieved the desired classification result. During feature selection, we set the maximum number of features in to 10, ranged from 0.1 to 1 in 0.1 increments, used the classifier function defined in Eq. (10), and allowed a maximum of three clusters.
Next, we defined three types of DME classifiers: 1) a layer classifier ( ) for segmenting layer boundaries, 2) a fluid classifier ( ) for identifying fluid-filled regions, and 3) a layer + fluid classifier ( ) used to remove false positive fluid detected by the fluid classifier. For each classifier type, we defined the criterion function following Eqs. (16) and (17), where is the set of reduced feature vectors for all training data, is the true class for and is the class that is being analyzed. For the fluid classifier, so only the fluid class is evaluated by the criterion function. For the layer classifier, so only the layers are analyzed, and for the layer + fluid classifier so all classes are analyzed. The resulting feature and weights are shown in Table 5 and Fig. 6 , where is a 3-cluster classifier, and and are 1-cluster classifiers.
Table 5. Selected DME classifier features and weights.
Layer
|
Fluid
|
Layer + Fluid
|
|||||
---|---|---|---|---|---|---|---|
Feature | Weight | Feature | Weight | Feature | Weight | ||
v11 | 1.5 | v1 | 2.7 | v11 | 1.4 | ||
v12 | 0.9 | v3 | 1.9 | v13 | 0.8 | ||
v10 | 0.8 | v11 | 1.8 | v1 | 0.5 | ||
v6 | 0.8 | v13 | 1.0 | v14 | 0.5 | ||
v13 | 0.6 | v17 | 1.0 | v10 | 0.3 | ||
v14 | 0.5 | v9 | 0.6 | v12 | 0.3 | ||
v1 | 0.4 | v9 | 0.3 | ||||
v15 | 0.3 | v15 | 0.2 | ||||
v7 | 0.3 | v7 | 0.2 |
(16) |
(17) |
5.10 Merged DME classifier
Given three classification images , , and output by the layer, fluid, and layer + fluid classifiers developed in Section 5.9, we next merge the results to create a single classification result. We achieve this by first setting all regions in and to zero if it is not classified as fluid. Next, we remove all fluid clusters in that do not overlap with the fluid in . We also examine the shape of the fluid clusters and remove any that are abnormally wide and thin. More specifically, if a cluster is more than 536 µm wide and less than 33.5 µm tall, then we consider the cluster as a false positive. Finally, we generate the merged classification image by setting and subsequently assigning all pixels in to (fluid) where . Figure 4 shows an example training image comparing the true classes (Fig. 4(a)) to the automatically generated classes output by our algorithm (Fig. 4(b)).
6. Automatic classification-based GTDP segmentation algorithm for DME
We next leveraged the classification results from Section 5 to create a fully automatic GTDP segmentation algorithm to identify fluid-filled regions and seven retinal layers on SD-OCT images with DME. An outline of the algorithm is shown on the right side of Fig. 2 and explained in detail in the following subsections.
6.1 Pre-processing, flattening, and segmenting the retina
We pre-process all images in a volumetric scan, flatten the entire retina based on estimates of the RPE, and use GTDP to segment the ILM and BM following Sections 0 – 5.5.
6.2 Classification of fluid and retinal layers
To classify the images, we locate the fovea and perform second order iterative Gaussian KR as described in Sections 5.6 and 5.7, respectively. We then compute the 12 features listed in Table 5 for all pixels residing within the retina following Section 5.8. Finally, we classify the feature vectors using the three classifiers from Section 5.9, and merge the results following Section 5.10 to generate the final image containing the fluid and layer class estimates.
6.3 GTDP segmentation of intraretinal layers
To segment the six remaining intraretinal layer boundaries, we begin by creating a binary mask of the OS/RPE boundary based on the layer classification image To generate the binary image we set all pixels from the bottom of the classified ISE layer to the bottom of the classified RPE layer to 1. We then segment the OS/RPE boundary using the weights in Eq. (18) and a search region extending from the ILM to BM, where is the value of at node and Furthermore, for Eq. (18) and for all subsequent weight calculations, we compute using the denoised image generated in Section 6.2.
(18) |
Next, we segment the ISM/ISE by setting all pixels in ranging from the bottom of the classified ONL-ISM to the bottom of the classified ISE. We then segment the boundary using Eq. (18) where and the search region extends from halfway between the classified ILM and OS/RPE to13.4 µm above the OS/RPE.
The third intraretinal boundary we segment is the NFL/GCL. Since the ILM is oftentimes significantly brighter and creates a pseudo light-dark transition within the NFL, we take the pixel-wise square root of and use the resulting image to compute the light-to-dark gradient image of the NFL/GCL boundary, generating . We then compute the graph weights following
(19) |
where is the light-to-dark gradient of at node and values in ranging from the middle of the classified NFL to the bottom of the GCL-IPL are set to 1. We then set the search region to range from the ILM to the ISM/ISE. However, since the thickness of the NFL varies more than all other layers and is oftentimes very faint on the temporal side, we provide additional search region steps as follows: 1) the bottom boundary for A-scans less than 1 mm from the fovea is set to the ILM plus 40.2 µm times the distance from the fovea (in mm), 2) the top boundary for A-scans between 1.5 and 3 mm nasal to the fovea is set to the ILM plus 6.7 µm times the distance from the fovea, and 3) the bottom boundary for A-scans greater than 1 mm temporal to the fovea is set to a third of the distance between the ILM and ISM/ISE.
Next, we segment the OPL/ONL using Eq. (20) where and the binary image is set to 1 between the top of the classified OPL and the middle of the classified ONL. We then set the top of the search region to either the NFL/GCL plus the NFL thickness, or 67 µm below the NGL/GCL, whichever is closer to the NFL/GCL. Finally, the bottom of the search region is set to the ISM/ISE.
(20) |
To segment the INL/OPL, we set values in extending from the top of the classified INL to the bottom of the classified OPL to 1. We also morphologically dilate the edges of the fluid in using a disk structuring element with a radius of 2 pixels. The dilated edges are also set to 1 in . We then use Eq. (18) with and to compute graph weights, and we limit the search region extent from 13.4 µm below the NFL to the OPL/ONL.
With the IPL/INL being the last boundary to segment, we set values in ranging from the bottom of the classified GCL-IPL to the bottom of the INL to 1. To discourage the IPL/INL from running through fluid regions, we set fluid regions to a negative value in We then compute graph weights following Eq. (20) with and Finally, we limit the search region to range from the a third of the distance between the NFL/GCL and INL/OPL, to the INL/OPL.
7. Algorithm validation
7.1 DME validation data set
We obtained our validation data set by identifying ten patients with DME that were not included in the training data set. The method for selecting these data sets is described in Section 5.1, with the difference that the images had to be of adequate quality (i.e. layer and fluid boundaries needed to be visible). The image acquisition parameters were consistent with the training data set, and lateral and azimuthal resolutions ranged from 10.94 – 11.98 µm/pixel and 118 – 128 µm/pixel, respectively. We made the entire data set available online, including the training and validation data sets and their corresponding automatic and manual segmentation results. This data set can be found at www.duke.edu/~sf59/Chiu_BOE_2014_dataset.htm.
7.2 Automatic versus manual segmentation
Two ophthalmologists manually segmented all fluid-filled regions and eight retinal layer boundaries using custom software (DOCTRAP V50.9) for 110 images (11 per patient). The images selected consisted of the foveal scan and scans ± 2, ± 5, ± 10, ± 15, and ± 20. The graders worked in a dimly lit room alone and free of other distractions. A tablet computer was used to allow the grader to utilize a stylus to mark or draw the segmentation as if writing on a flat surface. While marking the images, the ophthalmologists were able to use either a freeform (drawing) or connect-the-dots (piecewise cubic-interpolated clicking) method. They also had the options to hide/show other markings, erase/add/correct markings, and pin/unpin adjacent layer boundaries. The graders first segmented the retinal layers using predominantly the connect-the-dots function with freehand segmentation reserved for particularly complex contours such as those encountered when retinal layers are severely disrupted by edema. In cases where boundaries between layers were completely obscured or lost, the graders made their best estimate. The retinal layer segmentation could then be hidden to allow segmentation of edema without influence by retinal layer segmentation. The graders segmented retinal edema using the freehand function, and all edema was segmented whether it was cystic or non-cystic in nature. Cysts were segmented individually, whereas non-cystic edema was segmented according to the contour of thickening and the perceived location of the edema with respect to the retinal layers. The expert graders (MJA and PSM) are both fellowship-trained medical retina specialists on the faculty at the Duke Eye Center. They each have over five years of clinical experience working with diabetic subjects and reading OCT images. Both have prior experience with manual OCT grading and have each published two manuscripts involving OCT analysis and grading. Outside consultation and discussion between graders was prohibited.
To validate our algorithm, we compared our fully automatic KR + GTDP segmentation to our pre-determined primary grader (MJA) within a 6-mm lateral radius from the fovea. We then measured the Dice coefficient [74, 75] for retinal layer and fluid classification. We also measured the mean layer thickness difference between automatic and manual segmentation for each B-scan, and computed the absolute mean error across all 110 B-scans. To realize the improvement of our new KR + GTDP segmentation algorithm compared to existing methods that do not consider the presence of fluid, we also evaluated the performance of our GTDP algorithm developed for normal eyes on the DME data set. This Normal GTDP algorithm is a modified version of our 2010 algorithm described in [54], where enhancements had been made following our 2010 publication to improve segmentation results and address issues such as registration artifacts seen on Spectralis SD-OCT images. Finally, we computed the mean Dice coefficient and thickness difference for each layer for a given patient and performed the Wilcoxon matched-pairs test across all patients for each automated algorithm compared to the inter-manual results, as well as the Normal and DME algorithms compared to each other.
8. Experimental results
Figure 7 (right) shows the qualitative results of our automatic KR + GTDP segmentation algorithm for varying amounts of edema, including cases where there was no fluid present (Fig. 7(f)). This is in comparison to the automatic GTDP results using the normal algorithm (Fig. 7, middle). Table 6 and Table 7 report quantitative results comparing automatic and manual segmentation, with an overall mean Dice coefficient and thickness difference of 0.78 ± 0.10 and 5.22 ± 5.81 µm, respectively, for our proposed KR + GTDP algorithm. This is comparable to the inter-manual results which yielded an overall mean Dice coefficient and thickness difference of 0.79 ± 0.09 and 4.30 ± 2.38 µm, respectively. On the other hand, our popular GTDP algorithm developed for eyes without pathology had mean overall values of 0.70 ± 0.19 and 11.39 ± 19.11 µm. Table 8 shows the resulting p-values when comparing a) the mean patient Dice and thickness difference values between the automated and gold standard methods, to b) the inter-manual (Columns 1 and 2) or Normal GTDP (Column 3) results. An insignificant result against the inter-manual results indicates that the automated algorithm performed comparably to a second expert grader, while a significant result against the Normal algorithm indicates an improvement in the segmentation algorithm. Finally, Fig. 8 is an example comparison of the segmentations completed by the algorithm (Fig. 8(b)) and by the two graders (Fig. 8(c) and 8(d)), and Fig. 9 shows example errors in automatic segmentation (Fig. 9, bottom) when compared to a manual grader (Fig. 9, middle). The performance of the algorithms in MATLAB (MathWorks, Natick, MA) were 1.8 seconds per image on average for the Normal GTDP algorithm and 11.4 seconds for the KR + GTDP algorithm. This is in comparison to the 5.5 minutes required per image for manual segmentation.
Table 6. Dice coefficients for DME segmentation.
Class |
Inter-Manual
|
Normal GTDP
Automatic vs Manual
|
DME KR + GTDP
Automatic vs Manual
|
|||||
---|---|---|---|---|---|---|---|---|
Mean | SD | Mean | SD | Mean | SD | |||
Fluid | 0.58 | 0.32 | N/A | N/A | 0.53 | 0.34 | ||
NFL | 0.86 | 0.06 | 0.78 | 0.19 | 0.86 | 0.05 | ||
GCL-IPL | 0.89 | 0.05 | 0.74 | 0.25 | 0.88 | 0.06 | ||
INL | 0.77 | 0.07 | 0.63 | 0.22 | 0.73 | 0.08 | ||
OPL | 0.72 | 0.08 | 0.63 | 0.19 | 0.73 | 0.09 | ||
ONL-ISM | 0.87 | 0.06 | 0.86 | 0.09 | 0.86 | 0.08 | ||
ISE | 0.85 | 0.05 | 0.85 | 0.06 | 0.86 | 0.05 | ||
OS-RPE | 0.82 | 0.05 | 0.79 | 0.06 | 0.80 | 0.06 |
Table 7. Layer thickness differences for DME segmentation in µm.
Class |
Inter-Manual
|
Normal GTDP
Automatic vs Manual
|
DME KR + GTDP
Automatic vs Manual
|
|||||
---|---|---|---|---|---|---|---|---|
Mean | SD | Mean | SD | Mean | SD | |||
Fluid | N/A | N/A | N/A | N/A | N/A | N/A | ||
NFL | 4.02 | 3.30 | 25.83 | 50.42 | 3.68 | 4.13 | ||
GCL-IPL | 4.91 | 6.74 | 11.90 | 20.33 | 4.84 | 5.12 | ||
INL | 6.04 | 8.48 | 9.91 | 20.40 | 7.90 | 11.92 | ||
OPL | 4.21 | 3.74 | 6.50 | 5.90 | 6.35 | 6.11 | ||
ONL-ISM | 4.93 | 5.25 | 19.58 | 32.29 | 6.80 | 8.37 | ||
ISE | 3.03 | 2.11 | 2.53 | 2.00 | 2.88 | 2.28 | ||
OS-RPE | 2.99 | 2.30 | 3.51 | 2.39 | 3.61 | 2.47 |
Table 8. Wilcoxon matched-pairs test p-values.
Class |
Normal GTDP vs Manual
|
DME KR + GTDP vs Manual
|
DME KR + GTDP vs Normal GTDP
|
|||||
---|---|---|---|---|---|---|---|---|
Dice | Thickness | Dice | Thickness | Dice | Thickness | |||
Fluid | N/A | N/A | 0.432 | N/A | N/A | N/A | ||
NFL | 0.014* | 0.014* | 0.695 | 0.557 | 0.049* | 0.027* | ||
GCL-IPL | 0.002* | 0.004* | 0.432 | 0.557 | 0.010* | 0.006* | ||
INL | 0.006* | 0.557 | 0.027* | 0.275 | 0.014* | > 0.999 | ||
OPL | 0.002* | 0.037* | 0.322 | 0.084 | 0.004* | 0.922 | ||
ONL-ISM | 0.557 | 0.010* | 0.193 | 0.049* | 0.275 | 0.004* | ||
ISE | 0.695 | 0.084 | 0.322 | 0.275 | 0.770 | 0.232 | ||
OS-RPE | 0.106 | 0.275 | 0.846 | 0.065 | 0.193 | 0.695 |
* indicates statistical significance
9. Discussion
Despite having two qualified graders to manually mark the validation data sets, their results did not overlap with high accuracy (Table 6). This is likely due to the lower image quality and severe DME pathology present within the data set, as we deliberately chose to use real-world, clinical images in our study. While not necessarily a representative example for all graded images, Fig. 8 shows the challenge faced by both expert and automatic graders when confronted with an image from clinical practice that is affected by severe DME. The massive retinal thickening and punctate inner retinal hyper-reflectivity results in shadowing and loss of detail in the outer retinal layers. While some cysts are well-delineated hypo-reflective spaces in Fig. 8(a), others are minimally hypo-reflective compared to the surrounding tissue and create a judgment call for the grader. In this case, one grader perceived some non-cystic edema (Fig. 8(d)) while the other identified primarily cystic edema (Fig. 8(c)). The subjective nature of complex image segmentation by humans is thus one of the major obstacles to generating robust imaging-based biomarkers and is the primary unmet scientific and clinical need that is addressable using automated algorithms.
Despite the difficulty in managing such a challenging data set, Table 6 indicates that our fully automatic KR + GTDP algorithm generated results with an error close to the inter-observer variability (Dice coefficient of 0.78 versus 0.79). Furthermore, this was a significant improvement from previous algorithms developed by us, as our Normal GTDP method was tested on images without the significant pathology as seen on our DME data set. KR + GTDP classification of fluid-filled regions yielded the lowest Dice coefficient with errors similar to those shown in Fig. 9. In Fig. 9(a), false positives were detected near areas containing hyper-reflective deposits. Since the training data set did not include images with these hyper-reflective spots, the algorithm perceived these areas as “normal” and its surrounding regions as hypo-reflective, thus resulting in false positive classifications. Future versions of the algorithm can include the deposits as an additional class. Figure 9(b), in contrast, is a very dim image. Since the image contained a single hyper-reflective spot, the rest of the image remained dark after linear normalization. In future versions, the image can instead be non-linearly normalized (e.g. histogram equalization) prior to using the DME classifier. Finally, Fig. 9(c) shows the algorithm’s tendency to identify more regions of the ONL as fluid compared to the grader. This may be partially due to the small training data set size of 12 images.
Classification-based GTDP segmentation includes embedded denoising which is advantageous for reducing the noise present within the classification feature vectors (Fig. 5). While many different denoising methods may be also used to reduce the noise, the exciting properties of KR include its direct estimation of the denoised first and second order derivatives ( and ), as well as the generation of locally-weighted kernels which can be used directly as features or reused for non-linear feature denoising. A limitation of KR is the dependence of parameter values on the image noise level. Thus in order to optimally denoise images of varying scan types, future versions of the algorithm should adapt the KR parameter values based on the estimated noise level. Adapting alternative image representation models such as sparsity-based techniques, shown to be very efficient in representing retinal OCT images [76], may also improve the performance and is part of our ongoing work.
The need for GTDP segmentation is a limitation to this study, as the KR-based classification method was not sufficient to accurately identify retinal layers as a standalone method. While K-means clustering helped improve the classification of layers with varying thicknesses and positions (e.g. a thicker NFL on the nasal side versus a thinner NFL on the temporal side as shown in Fig. 4(b), the classified intraretinal layers did not merge at the fovea. Furthermore, three separate classifiers were required to achieve the final classification result. While provided a balance in accuracy between the layer and fluid classifiers, generated the most accurate layer classifications and achieved better fluid classifications but with more false positives. We therefore merged all three classifiers to achieve the best possible classification result. Finally, since there was no hard constraint on layer ordering, Fig. 4(b) shows the OS-RPE layer sometimes appearing above the ISE. Nevertheless, the classification example shown in Fig. 4(b) shows that our method provides a significant starting point that is helpful in simplifying the GTDP search region step. Other more advanced classification methods (e.g. support vector machines) may also be implemented in place of our described method to possibly improve the classifier performance.
The classifier features and weights shown in Fig. 6 and Table 5 provide insight into the characteristics that are important for fluid and layer identification. For example, the denoised intensity feature ( ) was the most important feature for identifying fluid, likely due to the dark intensity exhibited by fluid. In contrast, for both cases where the classifier was required to identify layers, the vertical position within the retina ( ) was most important. This is due to the layered nature of the retina. Adapting alternative image representation models such as sparsity-based techniques, shown to be very efficient in representing retinal OCT images [76], may improve the performance and is part of our ongoing work.
Finally, our randomly selected validation data set did not include many images with subretinal fluid, a relatively uncommon feature in DME. Thus, its performance for such images is unknown. Furthermore, our algorithm does not yet distinguish between cystic and non-cystic fluid. Incorporating the ability to identify and categorize different types of fluid (e.g. subretinal, cystic, non-cystic) is a necessary step in our future work, as this will enable rigorous DME subtyping. While the correlation between DME subtype and response to therapy is still unknown, the potential clinical implications of personalized therapy and improved treatment outcomes are vast.
10. Conclusion
We developed a fully automatic algorithm to identify fluid and eight retinal layer boundaries on SD-OCT images with DME pathology. Results showed that our automatic algorithm performed comparably to manual graders. This is the first validated, fully-automated, and multi-layer segmentation method which has been applied to real-world, clinical images containing severe DME. Accurate identification of DME imaging biomarkers is extremely important, as it will facilitate the quantification and understanding of DME.
Acknowledgments
This work was supported by the NIH R01 EY022691, K12 EY016333-08, 1R01EY023039, and the U.S. Army Medical Research Acquisition Activity Contract W81XWH-12-1-0397.
References and links
- 1.Davidson J. A., Ciulla T. A., McGill J. B., Kles K. A., Anderson P. W., “How the diabetic eye loses vision,” Endocrine 32(1), 107–116 (2007). 10.1007/s12020-007-0040-9 [DOI] [PubMed] [Google Scholar]
- 2.Yau J. W., Rogers S. L., Kawasaki R., Lamoureux E. L., Kowalski J. W., Bek T., Chen S. J., Dekker J. M., Fletcher A., Grauslund J., Haffner S., Hamman R. F., Ikram M. K., Kayama T., Klein B. E., Klein R., Krishnaiah S., Mayurasakorn K., O’Hare J. P., Orchard T. J., Porta M., Rema M., Roy M. S., Sharma T., Shaw J., Taylor H., Tielsch J. M., Varma R., Wang J. J., Wang N., West S., Xu L., Yasuda M., Zhang X., Mitchell P., Wong T. Y., Meta-Analysis for Eye Disease (META-EYE) Study Group , “Global prevalence and major risk factors of diabetic retinopathy,” Diabetes Care 35(3), 556–564 (2012). 10.2337/dc11-1909 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Kumar V., Abbas A. K., Fausto N., Robbins S. L., Cotran R. S., Robbins and Cotran pathologic basis of disease, 7th ed. (Elsevier Saunders, Philadelphia, 2005), pp. xv, 1525 p. [Google Scholar]
- 4.Pe’er J., Folberg R., Itin A., Gnessin H., Hemo I., Keshet E., “Upregulated expression of vascular endothelial growth factor in proliferative diabetic retinopathy,” Br. J. Ophthalmol. 80(3), 241–245 (1996). 10.1136/bjo.80.3.241 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Tolentino M. J., McLeod D. S., Taomoto M., Otsuji T., Adamis A. P., Lutty G. A., “Pathologic features of vascular endothelial growth factor-induced retinopathy in the nonhuman primate,” Am. J. Ophthalmol. 133(3), 373–385 (2002). 10.1016/S0002-9394(01)01381-2 [DOI] [PubMed] [Google Scholar]
- 6.Bringmann A., Pannicke T., Grosche J., Francke M., Wiedemann P., Skatchkov S. N., Osborne N. N., Reichenbach A., “Müller cells in the healthy and diseased retina,” Prog. Retin. Eye Res. 25(4), 397–424 (2006). 10.1016/j.preteyeres.2006.05.003 [DOI] [PubMed] [Google Scholar]
- 7.Bringmann A., Reichenbach A., Wiedemann P., “Pathomechanisms of cystoid macular edema,” Ophthalmic Res. 36(5), 241–249 (2004). 10.1159/000081203 [DOI] [PubMed] [Google Scholar]
- 8.Negi A., Marmor M. F., “The resorption of subretinal fluid after diffuse damage to the retinal pigment epithelium,” Invest. Ophthalmol. Vis. Sci. 24(11), 1475–1479 (1983). [PubMed] [Google Scholar]
- 9.Marmor M. F., “Mechanisms of fluid accumulation in retinal edema,” Doc. Ophthalmol. 97(3/4), 239–249 (1999). 10.1023/A:1002192829817 [DOI] [PubMed] [Google Scholar]
- 10.Early Treatment Diabetic Retinopathy Study Research Group , “Treatment techniques and clinical guidelines for photocoagulation of diabetic macular edema. Early Treatment Diabetic Retinopathy Study Report Number 2,” Ophthalmology 94(7), 761–774 (1987). 10.1016/S0161-6420(87)33527-4 [DOI] [PubMed] [Google Scholar]
- 11.Cunningham E. T., Jr, Adamis A. P., Altaweel M., Aiello L. P., Bressler N. M., D’Amico D. J., Goldbaum M., Guyer D. R., Katz B., Patel M., Schwartz S. D., Macugen Diabetic Retinopathy Study Group , “A phase II randomized double-masked trial of pegaptanib, an anti-vascular endothelial growth factor aptamer, for diabetic macular edema,” Ophthalmology 112(10), 1747–1757 (2005). 10.1016/j.ophtha.2005.06.007 [DOI] [PubMed] [Google Scholar]
- 12.Elman M. J., Aiello L. P., Beck R. W., Bressler N. M., Bressler S. B., Edwards A. R., Ferris F. L., 3rd, Friedman S. M., Glassman A. R., Miller K. M., Scott I. U., Stockdale C. R., Sun J. K., Diabetic Retinopathy Clinical Research Network , “Randomized trial evaluating ranibizumab plus prompt or deferred laser or triamcinolone plus prompt laser for diabetic macular edema,” Ophthalmology 117(6), 1064–1077 (2010). 10.1016/j.ophtha.2010.02.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Mitchell P., Bandello F., Schmidt-Erfurth U., Lang G. E., Massin P., Schlingemann R. O., Sutter F., Simader C., Burian G., Gerstner O., Weichselberger A., RESTORE study group , “The RESTORE study: ranibizumab monotherapy or combined with laser versus laser monotherapy for diabetic macular edema,” Ophthalmology 118(4), 615–625 (2011). 10.1016/j.ophtha.2011.01.031 [DOI] [PubMed] [Google Scholar]
- 14.Elman M. J., Bressler N. M., Qin H., Beck R. W., Ferris F. L., 3rd, Friedman S. M., Glassman A. R., Scott I. U., Stockdale C. R., Sun J. K., Diabetic Retinopathy Clinical Research Network , “Expanded 2-year follow-up of ranibizumab plus prompt or deferred laser or triamcinolone plus prompt laser for diabetic macular edema,” Ophthalmology 118(4), 609–614 (2011). 10.1016/j.ophtha.2010.12.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Do D. V., Schmidt-Erfurth U., Gonzalez V. H., Gordon C. M., Tolentino M., Berliner A. J., Vitti R., Rückert R., Sandbrink R., Stein D., Yang K., Beckmann K., Heier J. S., “The DA VINCI Study: phase 2 primary results of VEGF Trap-Eye in patients with diabetic macular edema,” Ophthalmology 118(9), 1819–1826 (2011). 10.1016/j.ophtha.2011.02.018 [DOI] [PubMed] [Google Scholar]
- 16.Gillies M. C., Simpson J. M., Gaston C., Hunt G., Ali H., Zhu M., Sutter F., “Five-year results of a randomized trial with open-label extension of triamcinolone acetonide for refractory diabetic macular edema,” Ophthalmology 116(11), 2182–2187 (2009). 10.1016/j.ophtha.2009.04.049 [DOI] [PubMed] [Google Scholar]
- 17.Kotsolis A. I., Tsianta E., Niskopoulou M., Masaoutis P., Baltatzis S., Ladas I. D., “Ranibizumab for diabetic macular edema difficult to treat with focal/grid laser,” Graefe's archive for clinical and experimental ophthalmology = Albrecht Von Graefes Arch. Klin. Exp. Ophthalmol. 248(11), 1553–1557 (2010). 10.1007/s00417-010-1428-4 [DOI] [PubMed] [Google Scholar]
- 18.Nguyen Q. D., Shah S. M., Khwaja A. A., Channa R., Hatef E., Do D. V., Boyer D., Heier J. S., Abraham P., Thach A. B., Lit E. S., Foster B. S., Kruger E., Dugel P., Chang T., Das A., Ciulla T. A., Pollack J. S., Lim J. I., Eliott D., Campochiaro P. A., READ-2 Study Group , “Two-year outcomes of the ranibizumab for edema of the mAcula in diabetes (READ-2) study,” Ophthalmology 117(11), 2146–2151 (2010). 10.1016/j.ophtha.2010.08.016 [DOI] [PubMed] [Google Scholar]
- 19.Massin P., Bandello F., Garweg J. G., Hansen L. L., Harding S. P., Larsen M., Mitchell P., Sharp D., Wolf-Schnurrbusch U. E., Gekkieva M., Weichselberger A., Wolf S., “Safety and efficacy of ranibizumab in diabetic macular edema (RESOLVE Study): a 12-month, randomized, controlled, double-masked, multicenter phase II study,” Diabetes Care 33(11), 2399–2405 (2010). 10.2337/dc10-0493 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Browning D. J., Altaweel M. M., Bressler N. M., Bressler S. B., Scott I. U., Diabetic Retinopathy Clinical Research Network , “Diabetic macular edema: what is focal and what is diffuse?” Am. J. Ophthalmol. 146(5), 649–655 (2008). 10.1016/j.ajo.2008.07.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Hunter A., Chin E. K., Telander D. G., “Macular edema in the era of spectral-domain optical coherence tomography,” Clin. Ophthalmol. 7, 2085–2089 (2013). 10.2147/OPTH.S49552 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Helmy Y. M., Atta Allah H. R., “Optical coherence tomography classification of diabetic cystoid macular edema,” Clin. Ophthalmol. 7, 1731–1737 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Sikorski B. L., Malukiewicz G., Stafiej J., Lesiewska-Junk H., Raczynska D., “The diagnostic function of OCT in diabetic maculopathy,” Mediators Inflamm. 2013, 434560 (2013). 10.1155/2013/434560 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Sonoda S., Sakamoto T., Shirasawa M., Yamashita T., Otsuka H., Terasaki H., “Correlation between reflectivity of subretinal fluid in OCT images and concentration of intravitreal VEGF in eyes with diabetic macular edema,” Invest. Ophthalmol. Vis. Sci. 54(8), 5367–5374 (2013). 10.1167/iovs.13-12382 [DOI] [PubMed] [Google Scholar]
- 25.Sim D. A., Keane P. A., Fung S., Karampelas M., Sadda S. R., Fruttiger M., Patel P. J., Tufail A., Egan C. A., “Quantitative analysis of diabetic macular ischemia using optical coherence tomography,” Invest. Ophthalmol. Vis. Sci. 55(1), 417–423 (2014). 10.1167/iovs.13-12677 [DOI] [PubMed] [Google Scholar]
- 26.Ghosh S., Bansal P., Shejao H., Hegde R., Roy D., Biswas S., “Correlation of morphological pattern of optical coherence tomography in diabetic macular edema with systemic risk factors in middle aged males,” Int. Ophthalmol. 35, 3–10 (2014). [DOI] [PubMed] [Google Scholar]
- 27.Antony B. J., Abràmoff M. D., Harper M. M., Jeong W., Sohn E. H., Kwon Y. H., Kardon R., Garvin M. K., “A combined machine-learning and graph-based framework for the segmentation of retinal surfaces in SD-OCT volumes,” Biomed. Opt. Express 4(12), 2712–2728 (2013). 10.1364/BOE.4.002712 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Chiu S. J., Li X. T., Nicholas P., Toth C. A., Izatt J. A., Farsiu S., “Automatic segmentation of seven retinal layers in SDOCT images congruent with expert manual segmentation,” Opt. Express 18(18), 19413–19428 (2010). 10.1364/OE.18.019413 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.DeBuc D. C., Somfai G. M., Ranganathan S., Tátrai E., Ferencz M., Puliafito C. A., “Reliability and reproducibility of macular segmentation using a custom-built optical coherence tomography retinal image analysis software,” J. Biomed. Opt. 14(6), 064023 (2009). 10.1117/1.3268773 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Ishikawa H., Stein D. M., Wollstein G., Beaton S., Fujimoto J. G., Schuman J. S., “Macular Segmentation with Optical Coherence Tomography,” Invest. Ophthalmol. Vis. Sci. 46(6), 2012–2017 (2005). 10.1167/iovs.04-0335 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Vermeer K. A., van der Schoot J., Lemij H. G., de Boer J. F., “Automated segmentation by pixel classification of retinal layers in ophthalmic OCT images,” Biomed. Opt. Express 2(6), 1743–1756 (2011). 10.1364/BOE.2.001743 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Shahidi M., Wang Z., Zelkha R., “Quantitative thickness measurement of retinal layers imaged by optical coherence tomography,” Am. J. Ophthalmol. 139(6), 1056–1061 (2005). 10.1016/j.ajo.2005.01.012 [DOI] [PubMed] [Google Scholar]
- 33.Yang Q., Reisman C. A., Chan K., Ramachandran R., Raza A., Hood D. C., “Automated segmentation of outer retinal layers in macular OCT images of patients with retinitis pigmentosa,” Biomed. Opt. Express 2(9), 2493–2503 (2011). 10.1364/BOE.2.002493 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Cabrera Fernández D., Salinas H. M., Puliafito C. A., “Automated detection of retinal layer structures on optical coherence tomography images,” Opt. Express 13(25), 10200–10216 (2005). 10.1364/OPEX.13.010200 [DOI] [PubMed] [Google Scholar]
- 35.Chiu S. J., Izatt J. A., O’Connell R. V., Winter K. P., Toth C. A., Farsiu S., “Validated Automatic Segmentation of AMD Pathology Including Drusen and Geographic Atrophy in SD-OCT Images,” Invest. Ophthalmol. Vis. Sci. 53(1), 53–61 (2012). 10.1167/iovs.11-7640 [DOI] [PubMed] [Google Scholar]
- 36.Lang A., Carass A., Hauser M., Sotirchos E. S., Calabresi P. A., Ying H. S., Prince J. L., “Retinal layer segmentation of macular OCT images using boundary classification,” Biomed. Opt. Express 4(7), 1133–1152 (2013). 10.1364/BOE.4.001133 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Mishra A., Wong A., Bizheva K., Clausi D. A., “Intra-retinal layer segmentation in optical coherence tomography images,” Opt. Express 17(26), 23719–23728 (2009). 10.1364/OE.17.023719 [DOI] [PubMed] [Google Scholar]
- 38.Paunescu L. A., Schuman J. S., Price L. L., Stark P. C., Beaton S., Ishikawa H., Wollstein G., Fujimoto J. G., “Reproducibility of Nerve Fiber Thickness, Macular Thickness, and Optic Nerve Head Measurements Using StratusOCT,” Invest. Ophthalmol. Vis. Sci. 45(6), 1716–1724 (2004). 10.1167/iovs.03-0514 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Mayer M. A., Hornegger J., Mardin C. Y., Tornow R. P., “Retinal Nerve Fiber Layer Segmentation on FD-OCT Scans of Normal Subjects and Glaucoma Patients,” Biomed. Opt. Express 1(5), 1358–1383 (2010). 10.1364/BOE.1.001358 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Mujat M., Chan R., Cense B., Park B., Joo C., Akkin T., Chen T., de Boer J., “Retinal nerve fiber layer thickness map determined from optical coherence tomography images,” Opt. Express 13(23), 9480–9491 (2005). 10.1364/OPEX.13.009480 [DOI] [PubMed] [Google Scholar]
- 41.Carass A., Lang A., Hauser M., Calabresi P. A., Ying H. S., Prince J. L., “Multiple-object geometric deformable model for segmentation of macular OCT,” Biomed. Opt. Express 5(4), 1062–1074 (2014). 10.1364/BOE.5.001062 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Somfai G. M., Tátrai E., Laurik L., Varga B. E., Ölvedy V., Smiddy W. E., Tchitnga R., Somogyi A., DeBuc D. C., “Fractal-based analysis of optical coherence tomography data to quantify retinal tissue damage,” BMC Bioinformatics 15(1), 295 (2014). 10.1186/1471-2105-15-295 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Sohn E. H., Chen J. J., Lee K., Niemeijer M., Sonka M., Abràmoff M. D., “Reproducibility of diabetic macular edema estimates from SD-OCT is affected by the choice of image analysis algorithm,” Invest. Ophthalmol. Vis. Sci. 54(6), 4184–4188 (2013). 10.1167/iovs.12-10420 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Huang Y., Danis R. P., Pak J. W., Luo S., White J., Zhang X., Narkar A., Domalpally A., “Development of a semi-automatic segmentation method for retinal OCT images tested in patients with diabetic macular edema,” PLoS ONE 8(12), e82922 (2013). 10.1371/journal.pone.0082922 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Lee J. Y., Chiu S. J., Srinivasan P. P., Izatt J. A., Toth C. A., Farsiu S., Jaffe G. J., “Fully automatic software for retinal thickness in eyes with diabetic macular edema from images acquired by cirrus and spectralis systems,” Invest. Ophthalmol. Vis. Sci. 54(12), 7595–7602 (2013). 10.1167/iovs.13-11762 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Chiu S. J., Toth C. A., Bowes Rickman C., Izatt J. A., Farsiu S., “Automatic segmentation of closed-contour features in ophthalmic images using graph theory and dynamic programming,” Biomed. Opt. Express 3(5), 1127–1140 (2012). 10.1364/BOE.3.001127 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Quellec G., Lee K., Dolejsi M., Garvin M. K., Abràmoff M. D., Sonka M., “Three-dimensional analysis of retinal layer texture: identification of fluid-filled regions in SD-OCT of the macula,” IEEE Trans. Med. Imaging 29(6), 1321–1330 (2010). 10.1109/TMI.2010.2047023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Chen X., Niemeijer M., Zhang L., Lee K., Abramoff M. D., Sonka M., “Three-dimensional segmentation of fluid-associated abnormalities in retinal OCT: probability constrained graph-search-graph-cut,” IEEE Trans. Med. Imaging 31(8), 1521–1531 (2012). 10.1109/TMI.2012.2191302 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Pilch M., Stieger K., Wenner Y., Preising M. N., Friedburg C., Meyer zu Bexten E., Lorenz B., “Automated segmentation of pathological cavities in optical coherence tomography scans,” Invest. Ophthalmol. Vis. Sci. 54(6), 4385–4393 (2013). 10.1167/iovs.12-11396 [DOI] [PubMed] [Google Scholar]
- 50.Roychowdhury S., Koozekanani D. D., Radwan S., Parhi K. K., “Automated localization of cysts in diabetic macular edema using optical coherence tomography images,” Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Conference 2013, 1426–1429 (2013). [DOI] [PubMed] [Google Scholar]
- 51.Lang A., Carass A., Swingle E. K., Al-Louzi O., Bhargava P., Saidha S., Ying H. S., Calabresi P. A., Prince J. L., “Automatic segmentation of microcystic macular edema in OCT,” Biomed. Opt. Express 6(1), 155–169 (2015). 10.1364/BOE.6.000155 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Tian J., Marziliano P., Baskaran M., Tun T. A., Aung T., “Automatic segmentation of the choroid in enhanced depth imaging optical coherence tomography images,” Biomed. Opt. Express 4(3), 397–411 (2013). 10.1364/BOE.4.000397 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Takeda H., Farsiu S., Milanfar P., “Kernel regression for image processing and reconstruction,” IEEE Trans. Image Process. 16, 349–366 (2007). [DOI] [PubMed] [Google Scholar]
- 54.Chiu S. J., Li X. T., Nicholas P., Toth C. A., Izatt J. A., Farsiu S., “Automatic segmentation of seven retinal layers in SDOCT images congruent with expert manual segmentation,” Opt. Express 18(18), 19413–19428 (2010). 10.1364/OE.18.019413 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Chiu S. J., Izatt J. A., O’Connell R. V., Winter K. P., Toth C. A., Farsiu S., “Validated automatic segmentation of AMD pathology including drusen and geographic atrophy in SD-OCT images,” Invest. Ophthalmol. Vis. Sci. 53(1), 53–61 (2012). 10.1167/iovs.11-7640 [DOI] [PubMed] [Google Scholar]
- 56.Takeda H., Farsiu S., Milanfar P., “Deblurring Using Regularized Locally Adaptive Kernel Regression,” IEEE Trans. Image Process. 17(4), 550–563 (2008). 10.1109/TIP.2007.918028 [DOI] [PubMed] [Google Scholar]
- 57.Seo H. J., Milanfar P., “Training-Free, Generic Object Detection Using Locally Adaptive Regression Kernels,” IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1688–1704 (2010). 10.1109/TPAMI.2009.153 [DOI] [PubMed] [Google Scholar]
- 58.Ruppert D., Wand M. P., “Multivariate locally weighted least squares regression,” Ann. Stat. 22(3), 1346–1370 (1994). 10.1214/aos/1176325632 [DOI] [Google Scholar]
- 59.Hansen P. C., “The truncated svd as a method for regularization,” BIT Numerical Mathematics 27(4), 534–553 (1987). 10.1007/BF01937276 [DOI] [Google Scholar]
- 60.X. G. Feng and P. Milanfar, “Multiscale principal components analysis for image local orientation estimation,” in Signals, Systems and Computers,2002. Conference Record of the Thirty-Sixth Asilomar Conference on, 2002), 478–482 vol.471. [Google Scholar]
- 61.Cormack R. M., “A review of classification,” J. R. Stat. Soc. [Ser A] 134(3), 321–367 (1971). 10.2307/2344237 [DOI] [Google Scholar]
- 62.S. B. Kotsiantis, “Supervised Machine Learning: A Review of Classification Techniques,” in Proceedings of the 2007 conference on Emerging Artificial Intelligence Applications in Computer Engineering: Real Word AI Systems with Applications in eHealth, HCI, Information Retrieval and Pervasive Technologies, (IOS Press, 2007), pp. 3–24. [Google Scholar]
- 63.Laws K. I., “Textured Image Segmentation,” (DTIC Document, 1980). [Google Scholar]
- 64.Shapiro L., Stockman G. C., “Computer Vision. 2001,” ed: Prentice Hall; (2001). [Google Scholar]
- 65.Jain A., Zongker D., “Feature selection: evaluation, application, and small sample performance,” IEEE Trans. Pattern Anal. Machine Intell. 19(2), 153–158 (1997). 10.1109/34.574797 [DOI] [Google Scholar]
- 66.Kohavi R., John G. H., “Wrappers for feature subset selection,” Artif. Intell. 97(1-2), 273–324 (1997). 10.1016/S0004-3702(97)00043-X [DOI] [Google Scholar]
- 67.Guyon I., Elisseeff A., “An introduction to variable and feature selection,” J. Mach. Learn. Res. 3, 1157–1182 (2003). [Google Scholar]
- 68.Mosteller F., Tukey J. W., Data analysis, including statistics (1968).
- 69.Lloyd S., “Least squares quantization in PCM,” IEEE Trans. Inf. Theory 28(2), 129–137 (1982). 10.1109/TIT.1982.1056489 [DOI] [Google Scholar]
- 70.Dijkstra E. W., “A note on two problems in connexion with graphs,” Numerische Mathematik. 1(1), 269–271 (1959). 10.1007/BF01386390 [DOI] [Google Scholar]
- 71.Chiu S. J., Lokhnygina Y., Dubis A. M., Dubra A., Carroll J., Izatt J. A., Farsiu S., “Automatic cone photoreceptor segmentation using graph theory and dynamic programming,” Biomed. Opt. Express 4(6), 924–937 (2013). 10.1364/BOE.4.000924 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Zuiderveld K., “Contrast limited adaptive histogram equalization,” in Graphics Gems IV, Paul S. H., ed. (Academic Press Professional, Inc., 1994), pp. 474–485. [Google Scholar]
- 73.Kohavi R., “A study of cross-validation and bootstrap for accuracy estimation and model selection,” in IJCAI, 1995), 1137–1145. [Google Scholar]
- 74.Sørensen T., “A method of establishing groups of equal amplitude in plant sociology based on similarity of species and its application to analyses of the vegetation on Danish commons,” Biol. Skr. 5, 1–34 (1948). [Google Scholar]
- 75.Dice L. R., “Measures of the amount of ecologic association between species,” Ecology 26(3), 297–302 (1945). 10.2307/1932409 [DOI] [Google Scholar]
- 76.Fang L., Li S., McNabb R. P., Nie Q., Kuo A. N., Toth C. A., Izatt J. A., Farsiu S., “Fast Acquisition and Reconstruction of Optical Coherence Tomography Images via Sparse Representation,” IEEE Trans. Med. Imaging 32(11), 2034–2049 (2013). 10.1109/TMI.2013.2271904 [DOI] [PMC free article] [PubMed] [Google Scholar]