Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2009 Apr 28.
Published in final edited form as: Exp Eye Res. 2007 Dec 7;86(4):562–575. doi: 10.1016/j.exer.2007.11.019

Automated, Computerized, Feature-Based Phenotype Analysis of Slit Lamp Images of the Mouse Lens

Jenny Yuen a, Yi Li a, Linda G Shapiro a, John I Clark b,c,d, Ernest Arnett b, E Helene Sage b,c, James F Brinkley a,b,e
PMCID: PMC2674242  NIHMSID: NIHMS46767  PMID: 18304532

Abstract

Longitudinal studies of a variety of transgenic mouse models for lens development can create substantial challenges in database management and analysis. We report a novel, automated, feature-based informatics approach to screening lens phenotypes in a large database of slit lamp images. Digital slit lamp images of normal and abnormal lenses in eyes of wild type (wt), SC1 null and SPARC null transgenic mice were recorded for quantitative evaluation of their structural phenotype. The images were processed to improve the contrast of structural features that corresponded to rings of opacity and fluctuations in scattering intensity in the lenses. Measurable attributes were assigned to the features in the lens images and given as an output vector of 46 dimensions. Characteristic patterns correlated with the structural phenotype of each mutant and wt lens and a statistical fit for each phenotype was defined. The genotype was identified correctly in nearly 85% of the slit lamp images on the basis of an automated computer analysis of the lens structural phenotype. The automated computer algorithm has the potential to evaluate a large database of slit lamp images and distinguish mouse genotypes on the basis of lens phenotypes objectively using a neural network analysis of the structural features observed in the slit lamp images. The neural network approach is a promising technology for objective evaluation of genotype/phenotype relationships based on structural features and light scattering in lenses. Further improvements in the automated method can be expected to simplify and increase the accuracy and efficiency of the feature based analysis of structural phenotypes linked to genetic variation.

Keywords: informatics, phenotype, genotype, slit lamp, imaging, lens

Introduction

Genetic variations in the eye are well known and often present structural phenotypes that can be mapped to specific, sometimes multiple, sites on chromosomes. In contrast, the ability to evaluate variations in structural phenotype observed in a large database of images can be complicated by difficult and subtle differences in the structural appearance of the eye that can depend on exposure conditions, lighting and the angle of imaging. This report describes the results of an effective, automated method for the quantitative comparison of a large database of slit lamp images using a neural network approach to the computerized analysis of variations in structure of three different lens phenotypes and their correlated genotype.

The lens is an excellent model for development of an objective analysis of structural phenotype because the lens structure is highly ordered and readily observed and recorded in living animals using a slit lamp. The lens is a sensitive indicator of hereditary and genetic disorders. A normal mammalian lens consists of symmetric layers of transparent cells containing a condensed protein solution that increases the refractive index to permit the lens to function as an optical element in the visual pathway. A complex network of special interactions between the concentrated cellular proteins favors short range, glass-like order and transparency across layers of lens cells (Benedek, 1983; Clark, 2004). Loss of transparency results when interactions between proteins are altered to produce inhomogeneities and large fluctuations in the index of refraction that scatter light. The opacity or cloudiness of the crystalline lens increases progressively with age (Harding, 1991; Kashima et al., 1993a; Thurston et al., 1997; Chylack, 2001; Thylefors et al., 2002). The spatial distribution and intensity of opacities within the lens varies with the molecular or genetic basis for the cataract (Phelps-Brown, 2000; Kojima et al., 2002; Fan, 2003; Seeberger et al., 2004).

Several methods for classification have been used effectively in clinical and epidemiological studies of cataract formation (Chylack et al., 1993a; Leske et al., 1996; Chylack, 2001; Klein et al., 2005) and it is widely recognized that specific patterns of opacification can be linked with hereditary and metabolic diseases including Alport's, Lowe's, Marfan's, and Down's, syndromes, as well as myotonic dystrophy, coraliform cataract, and chromosomal mutations (Phelps-Brown and Bron, 1996; Alizadeh et al., 2004; Graw, 2004). A distinctive form of opacity in Alzheimer's patients was identified using slit lamp imaging suggesting a relationships between structural phenotype and genotype (Goldstein et al., 2003). Mutant and transgenic mice have been linked to variations in the structural patterns of light scattering in lenses and in many cases, resemble human lens opacities (Clark and Carper, 1987; Rodriguez et al., 1992; Bettelheim et al., 1997; Graw, 1999; Norose et al., 2000; Tumminia et al., 2001; Yan et al., 2002). The success of the subjective classification of slit lamp images encourages the development of an automated computer based analyses of structural phenotypes that can be correlated with genetic mutations when the number of categories as well as the number of images increases in a very large database, as in a preclinical trial or genetic screen (Datiles et al., 1987; Chylack et al., 1995; Chylack, 2001; Seeberger et al., 2004; Wolf et al., 2005).

This report presents a novel method for computerized and quantitative evaluation of the progressive variations in lens structural features resulting from various mechanisms of opacification. No previous methods for the correlation of genotype with patterns of structural features in slit lamp images of mouse lenses have been reported. A Dynamic Light Scattering (DLS) fiber-optic probe that measured particle size distributions in animal lenses involves a different technology (Chenault et al., 2002; Datiles et al., 2002; Ansari, 2003; Simpanya et al., 2005) based on earlier quasielastic light scattering (QLS) instrumentation for lens opacification (Nishio et al., 1984; Benedek et al., 1987; Thurston et al., 1997). The DLS was used in the analysis of the distribution of particles of different sizes in the eyes of desert rodents, two groups of mice, guinea pigs and in a rat model for cataract formation (Ansari, 2003; Simpanya et al., 2005; Chenault et al., 2002). The DLS method is not automated and is limited in its ability to consider the spatial distribution of the scattering features in lens with respect to phenotype.

The density of opacities in different areas of the eye lens (posterior capsular, posterior cortical, nuclear, anterior cortical, and anterior capsular) and in different age ranges has been studied (Edwards et al., 1990; Chylack et al., 1993b; Kashima et al., 1993b). The minimum number of views required was studied in patients to avoid subjecting patient eyes to the repeated and extended exposure of intense light used for camera recordings and to reduce the number of images stored. Light exposure and photographic storage were problems that could limit the number of images per patient. Current imaging methods are significantly advanced with regard to light sensitivity for image acquisition using digital video imaging, and short exposure time for patients. The advancement in modern compression technologies allows storage of more images with higher resolution than was possible ten years ago.

A number of classification methods are in use to define clinically relevant features obtained from densitometric analyses that are correlated with opacity but are not for general phenotype analysis (Dimock et al., 2000; Seeberger et al., 2004). Some of these studies were automated for cataract classification of nuclear sclerosis from slit-lamp photographs (Fan, 2003) and gray scale processing with predefined landmarks was useful for classification of clinically important cataracts. The purpose of previous studies was different from the current approach which concentrates on patterns of subtle variations in structural features resulting from genetic mutation of mouse lenses at the earliest possible age.

Two mouse models that have abnormal lens phenotypes on the 129Sv background, the SC1/Hevin null and the SPARC (osteonectin, BM40) null mouse, were selected for the evaluation using the phenotype analysis (Brekken and Sage, 2001; Sullivan and Sage, 2004; Sullivan et al., 2006). The selection of the mouse models was based on the slow temporal progression of the opacification pattern in lens with age which involves zones of discontinuity characteristic of ageing in many mammalian lenses. The patterns are subtle without distinct or unique structural features. SC1/Hevin and SPARC are matricellular proteins that modulate interactions between the lens basement membrane or capsule and the epithelial cells. The absence of SC1/Hevin or SPARC alters mobility of differentiating fibers at the basal attachments to the lens capsule which results in a progressive loss of transparency in layers of lens fibers (Bassuk et al., 1999; Norose et al., 2000). The microscopic structure responsible for the slit lamp images remains to be determined using direct, systematic histological methods (Kuszak and Costello, 2004; Kuszak et al., 2006; Blankenship et al., 2007). In the current study, an innovative computerized approach was used to quantify collectively the patterns and features of the structural phenotype observed in images of transgenic mouse eyes. Structural features in the slit lamp image were assigned attributes that were concentrated into a vector of 46 dimensions for each image. The neural network assigned each input image to a set representing its genotype and labeled it accordingly. On the basis of the structural features alone the algorithm separated a testing set of slit lamp images into their three genetic categories, WT, SC1 null or SPARC null, with a high degree of accuracy. The results provide baseline data for improved informatics approaches to phenotype/genotype relationships in living animals. In a broader study the molecular basis for each phenotype will need to be investigated.

Materials and Methods

Slit-Lamp Eye Image Capture

We are given a set of k known cataract classes S = {C1,C2, …,Ck}, a training set of slit lamp mouse eye images for mice, TS, where each image t ∈ TS is labeled as corresponding to a category CI ∈ S, and a test image set TT such that TS ∩ TT = Ø. We want to CI ∈ S to which each image Q ∈ TT belongs. The wild type animals were the same strain, 129Sv, as the background for the mutants and the same animals were used as the wild type group for comparison with both mutant groups. The age range of the mice was 3 months to 25 months. The transgenic mouse models were examined and imaged using a Nikon FS-2 slit lamp ophthalmoscope and a Canon Optura 20 digital video camera. Frames from each examination were selected from video recordings, using Adobe Premiere, saved in TIF file format, and then cropped and oriented using Adobe Photoshop. The total image database is for over 100 different eyes in different mice. Duplicate images obtained under the same conditions were identical.

The set S of lenses used in this work is either: the WT (wild-type which has no laboratory-induced opacity), the Secreted Protein Acidic and Rich in Cysteine (SPARC) null (a matricellular protein), and the Synaptic Cleft (SC1) null.

S={WT,SC1,SPARC}

In both the SPARC-null and SC1-null animals, the gene encoding the specific protein was truncated or replaced so that the functional protein is no longer expressed. Figure 1 shows typical slit lamp images for each class. The original high-resolution images were resized to 300 by 300 pixel versions. The properties of the images in the datasets TS and TT are:

Figure 1.

Figure 1

Lens Phenotype for: Wild Type (WT), Synaptic Cleft 1 protein null (SC1) [16], and Secreted Protein Acidic and Rich in Cysteine (SPARC) null [1].

  • The eyes in the images are approximately the same size and are approximately centered at the same location in each image.

  • Each image is of a single eye in an individual mouse.

  • The illumination during the image capture process varies among the different experiments (figure 2).

  • There are artifacts caused by the illumination that are independent from the cataracts (figure 3).

  • The pattern in the center of the lens is directly related to the cataract.

  • A ring pattern can be observed in the WT class. Depending on the cataract, partial or total modification of this ring pattern can be observed.

  • Because of the manner in which the images are taken, the ring pattern is not circular but elliptical. This makes detection of the rings more difficult (figure 2).

  • The images contain structures such as eyelashes, which are not of interest.

Figure 2.

Figure 2

The shape of the eye in an image varies depending on both the mouse and the angle of incidence of the slit-lamp light source on the surface of the eye. Illumination was not controlled.

Figure 3.

Figure 3

Artifacts caused by the imaging process are marked by yellow ellipses.

The opacity or cloudiness of the lens changes the perception of the layers in the lens. The pattern of the rings or the relative colors between rings may vary, making this a useful feature in the characterization of the pattern of opacity in a lens. The angle of incidence of the slit-lamp light on the cornea accounts for the elliptical shape of the ring pattern shown in the image. Fitting circular arcs to the rings in the images was a good approximation to the layers of opacification. To include the largest amount of information, images with ring patterns as close to a circle as possible were preferred. Noise factors in the image were non-trivial because control of the imaging conditions was limited in the initial testing. Constant illumination and shape were not assumed (figure 3).

The analysis involves feature extraction and learning. Quantifiable features characteristic of particular patterns were defined. A defined set of features was extracted from each slit lamp image t ∈ TS, and these features were used to train a neural network (Russell, 1995). The training or learning process produced a model using the information from the training set. This model is used in the final analysis of each image.

Feature Extraction

A set of image features with measurable attributes was developed to discriminate between the different known genotypes. The features were ring, intensity profile, and histogram. For each slit-lamp image input into the feature extraction process, a numeric feature vector of 46 dimensions is output. The process is summarized in figure 4.

Figure 4.

Figure 4

Feature extraction process.

Ring Detection

Rings, or elliptical layers of varying transparency, are present in both the WT and mutant SC1 null and SPARC null lenses. Visualization of the layers of lens cells depends on the magnification and contrast. In the WT lens, the contrast between adjacent cells or layers of cells is small when observed in a slit lamp biomicroscope. The contrast increases when opacity occurs in some layers and not in adjacent layers. The increased contrast allows the rings to be observed in the slit-lamp. The pattern of rings can be correlated with the mutation. The proposed approach in identifying and quantifying these characteristics of the rings consists of a five-step process:

  1. Ring enhancement.

  2. Isolation of an elliptical sector of the lens.

  3. Transformation of the elliptical sector into a rectangular image containing only pixels corresponding to the lens.

  4. Compression of the rectangular image representation into a 1-dimensional array of mean intensities.

  5. Extraction of feature vector values from the 1-dimensional array produced in the previous step.

This process contributes 23 numerical values from the raw input image that will be used for the training phase.

Ring Enhancement

As shown in figure 5, the contrast between the rings is small. A standard solution for this problem is histogram equalization transform, which modifies the intensity values of image pixels to achieve nearly equal numbers of the distinct intensity values; this tends to stretch out the gray tones and increase the contrast. Figure 5b illustrates the histogram equalization transform applied to the image of figure 5a. Although there is some enhancement, the rings are not yet distinct. To solve this problem, a local histogram equalization transform that operates on local windows instead of the whole image is applied to the image. Each pixel value in the locally-equalized image is the result of applying histogram equalization to the sub-image formed by a window of fixed size centered at a pixel. The window size was 5 pixels square. Figure 5c shows the results of the local histogram equalization transform, which strongly enhanced the rings. This locally-equalized image will be the one used in the next steps for ring feature extraction.

Figure 5.

Figure 5

The original image (a), the original image after a histogram equalization (b), and the contrast enhancement (c).

Isolation of an elliptical sector and transformation to a rectangular image

Once the rings were enhanced, the parameters of the ring pattern were extracted. Let the coordinates of the center of the eye be (cx, cy). Consider the region that corresponds to the elliptical sector centered at (cx, cy) and between the angles α and , where 0 < -α < π/2. A ring was modeled as an elliptical arc with axes of length a and b parallel to the x and y axes respectively and centered at (cx, cy). Every point (x, y) in this elliptical arc was mapped to a point on a vertical line by the transformation Rect(x, y) = (cx + a, y) as shown in figure 6.

Figure 6.

Figure 6

The region inside an elliptical sector is mapped to a rectangular sector using the Rect(x, y) = (cx + a, y) transformation.

The Rect transformation was applied to every ellipse centered at (cx, cy) with a fixed c = a/b and 0 < a < w−cx, where w and h are the width and height of the image, respectively, to form a rectangular version of the elliptical eye. The resulting image of the Rect mapping was cropped to remove the cornea and the area outside the lens. The elliptical-to-rectangular transformation had three degrees of freedom: cx, cy, and c. The value of cy was restricted to h/2, working with two degrees of freedom for each image.

Compression of the Rectangular Eye Representation to a 1-dimensional Array of Data

The Rect transformation converts the elliptical rings into vertical lines for easier and more accurate analysis. It is important to choose the correct center (cx, cy) and c value for the lens. The previous step generates a large number of potential fits to the data. To evaluate the goodness of a chosen set of parameters, two vectors are constructed:

  • a vector M(cx, c) of values containing the mean intensity value for each column in the Rect(cx, c) image, and

  • a vector V(cx, c) containing the variance of each column in Rect(cx, c).

The score associated with each pair of values (cx, c) is defined as:

EllipseFitscore(cx,c)=vV(cx,c)v

Using this score, a pair of (cx, c) values that will be called (bestcx, bestc) is chosen. This pair has the property that ∀c'x, c'y such that:

|w2c'x|<ε

and

|h2c'y|<ε
EllipseFitscore(bestcx,bestcy)=EllipseFitscore(cx,cy)

for some ε > 0. Note that only the vector of variances V is used for parameter evaluation. The vector of means M was a simpler way of characterizing the ring information by exploiting property of homogeneous intensity of every ring. The result of the mapping by M will be used for the extraction of numerical features in the next step.

Ring Feature Extraction

The last transformation by M provided an easier method for identifying the boundaries of the rings, their average intensities, and other relationships that were useful features for the recognition of cataract patterns.

The mean vector M was analyzed to provide a feature vector of numeric attributes that was used for pattern recognition. The vector we used contained the following 23 ring attributes:

  • number, maximum, minimum, mean, and variance of maxima of the function;

  • number, maximum, minimum, mean, and variance of minima of the function;

  • maximum, minimum, mean, and variance of the distance between pairs of consecutive maxima;

  • maximum, minimum, mean, and variance of the distance between pairs of consecutive minima;

  • maximum, minimum, mean, and variance of the difference between each pair of consecutive maxima and a minima;

  • number of consecutive pairs of maxima and minima (regions formed by peaks and valleys).

These attributes captured several features of the rings. Globally, the attributes included the number of rings and the distribution of their properties such as mean intensity, width, and others. Locally, the attributes described the width and opacity of each ring and provided a comparison of these characteristics with the characteristics of other rings in the same image.

Intensity Profile

A fast Fourier Transform, FFT, was used to provide a parameter representing the spatial intensity fluctuations observed in a line scan across the diameter of slit lamp images of subtle opacities (Fig 9a). Intensity profiles were normalized to the corneal opacity for each mouse being examined, and duplicate images from the same eyes appear the same. For the function I(x), where the domain is each column in the image, and the range {0, 255} quantified the pixel values for intensity I(x), a polynomial was a poor fit to the spatial fluctuations in the noisy intensity profile across the center of the slit lamp image (Fig 9b). Instead, the entire intensity profile (Fig 9b) was the input for a Fast Fourier Transform, FourierI(x), which provided a power spectrum (fig 9c) containing the same information as the intensity profile in a broad, simplified reciprocal plot. The result is a much better polynomial fit. 12 coefficients of this polynomial (6 real and 6 imaginary) were concatenated to form a feature vector.

Figure 9.

Figure 9

The white line in the left image(a) indicates the row of pixels that was considered for the intensity profile feature. The image in the middle (b) shows a polynomial fit of degree 5 on the intensity profile. The image on the right (c) shows a polynomial fit on the output of the Fast Fourier Transform of the intensity profile.

Histogram of Western Quadrant of the Lens

There was a characteristic pattern of rings for each lens class. A histogram of the western quadrant of the lens was created and was fit to a 1-dimensional Gaussian using Maximum Likelihood. The variance of the Gaussian was included in the feature vector. Because of the variations in illumination caused by external sources in the image, only the variance, which was related to the distribution of intensities, and not the mean, which was directly associated with the change in external illuminations, was considered. Figure 11 shows the histogram of the images corresponding to each of the three of categories of images.

Figure 11.

Figure 11

The histogram of the western quadrant of the images of three different lens mutants. The variance of pixel intensities is smaller in SPARC null images and larger in SC1 null images, while the WT type images tend to have a variance value in between.

Segmented Least Squares Fitting on the Intensity Profile

While the intensity profile feature characterized some lenses, the function was assumed to be continuous and non-continuous fluctuations were approximated as smooth curves. Some classes such as SPARC null and WT lenses had similar polynomial fits but differed in the number of non-continuous changes. To encapsulate the “continuity,” a piecewise linear function was fit on I(x).

The standard least squares linear fitting algorithm fits a line to a set of n 2-dimensional points P = {(x1, y1), (x2, y2), …, (xn, yn)}. The line with the minimum error is y = ax + b, where

a=nixiyi(ixi)(iyi)nixi2(ixi)2
b=iyiaixin

The error of the fit is:

Error(L,P)=i=1n(yiaxib)2

The objective of the segmented least squares approximation is to find the partition of consecutive points that minimizes the error of the fit, which corresponds to the sum of the errors for each linear segment used in the fit with an additional cost C for each segment used.

Efit=piPError(li,pi)+C

where pi is a set of consecutive points in I(x), li is the least squares linear fit on this set, and C is the cost for each extra segment. We used values of C on the order of 500.

After fitting the piecewise linear function to I(x), only the left side of the lens was considered for this feature extraction. This was done by considering the segments up to the middle of the image and discarding the first segments that had a very small value for their slope and intersection at the origin. These segments corresponded to the dark section of the image, not a part of the mouse lens. The features extracted were:

  • number of segments;

  • means of the slope and the intersection at the origin (a,b) of each of the segments;

  • variances of the slope and the intersection at the origin (a,b) of each of the segments.

Learning

The analysis uses an inference engine mapping a feature space to a discrete set of labels for lens mutants. Each set is distinguished by the way it learns or obtains information to output decisions generalizing the training data presented to unseen situations. Supervised learning uses labeled training data as a knowledge base. Unsupervised learning determines patterns from unlabeled training input. Reinforcement learning involves a reward when the system output is correct and a penalty when the output is incorrect with the objective of identifying the output on the basis of the maximum reward. For the current problem supervised learning is expected to be most effective given the features of the images and their respective classes or labels.

Neural Networks

The information processing capacity of the human brain is thought to emerge from networks of neurons. This was the theoretical basis for the creation of artificial neural networks as inference engines. Neural networks are composed of nodes or units connected by directed links. A link from unit j to unit i serves to propagate the activation aj from j to i. Each link also has a numeric weight Wj,i associated with it, which determines the strength and sign of the connection. A network of units or a neural network is a structure with a set of units connected in a particular way and a set of defined weights for each unit. A network can be feed-forward, or recurrent. Feed-forward networks represent a function of their inputs, while recurrent networks feed their output back to their inputs. Feed-forward networks are arranged with layers of units such that each unit in a layer receives input from only the units in its preceding layer. A neural network represents the creation of a separator in an n-dimensional space (where n is the size of the feature vector or the number of inputs to the neural network). Neural networks can learn (adjust the weights of their units) using different techniques. Backpropagation is an iterative technique where an example is applied to the network (originally set with random weights) and the weights are readjusted based on the quality of the output. The value is then propagated backwards through the network and changes are made to the weights of the nodes in each layer. This process is repeated until the overall error is below some threshold.

Use of Neural Networks

The numeric attributes from the four types of features in a lens images can be concentrated into a 46-dimensional feature vector F = {f1, f2, …, f46}. For each training image IT ∈ TS, a feature vector F(IT) can be extracted and T = (F, c) can be constructed, where c indicates the known training vector F. The algorithm can be trained to discriminate between the different phenotypes and to become a decision function that, when given a vector U of attribute-values from a previously unseen lens image, will produce a value, c, that indicates the most likely genotype for the new image. The training process for obtaining the automated parameters is summarized in figure 13. The current analysis was implemented as a neural network with 15 hidden units in one hidden layer for training using the data mining software package Weka (Witten and Eibe, 2005). The model for the final system was trained with the whole dataset.

Figure 13.

Figure 13

Learning process.

Results

The testing phase (as observed in figure 14) used the parameters obtained from the training phase and accepted slit-lamp images outside the training dataset for the analysis. The resulting output was compared to the correct genotypes and an error was calculated. The analysis was tested using three image sets: SC1 null, SPARC null, and WT lenses. The testing was performed with a 10-fold cross validation. This method consisted of dividing each dataset into 10 subsets and using each of these subsets in one of the 10 iterations as the testing set while the rest were used as the training set. The dataset consisted of 107 images of SC1 null, 155 of SPARC null, and 121 of WT. The values listed in the following tables are the average for the ten individual experiments.

Figure 14.

Figure 14

Use of the final categorization.

Using only Ring features

The performance of the trained analysis on the ring feature alone was shown in table 1, expressed as a confusion matrix. The diagonal elements of the matrix (SC1,SC1), (SPARC, SPARC), and (WT, WT) indicated the number of instances in which an image was identified correctly. More than 69.19% of the images were identified correctly showing that the ring pattern feature correlated highly with the genotype. The off-diagonal elements show incorrect genotyping. Columns denote the underlying (real) categories and rows denote the inferred categories. For example, the entry in row SC1, column WT indicates that 17 SC1 images were wrongly categorized as WT.

Table 1. Confusion matrix for results using only ring features.

SC1 SPARC WT
SC1 81 9 17
SPARC 8 110 37
WT 12 35 74

Using only Intensity Profile features

The performance of a trained analysis on the intensity profile feature alone is shown in table 2. A significant improvement can be observed in the characterization of the SC1 null class in comparison to the ring feature. However, the analysis showed some confusion differentiating between SPARC null and WT. The percentage of correct identifications was 71.27%.

Table 2. Confusion matrix for the results trained using only intensity profile features.

SC1 SPARC WT
SC1 103 1 3
SPARC 1 109 45
WT 0 60 61

Using both Intensity Profile and Ring features

The performance for a combined ring-profile feature is shown in table 3. Using both features not only reduced the confusion between the WT and SPARC null mutants, but also increased the accuracy for the SC1 null mutant. This analysis achieved an overall accuracy of 77.2%.

Table 3. Confusion matrix for trained using ring and intensity profile features.

- SC1 SPARC WT
SC1 99 6 2
SPARC 3 111 41
WT 1 34 86

Using Ring, Intensity Profile, and Western Quadrant Histogram features

Integrating the histogram of one of the quadrants of the lens, a higher accuracy was achieved and more instances of the classes SPARC null and WT were identified correctly. Less accuracy was achieved with SC1 null in comparison to the previous setup. The overall accuracy is 78.79% and the confusion matrix was shown in table 4.

Table 4. Confusion matrix for the results using ring, intensity profile, and western quadrant histogram.

SC1 SPARC WT
SC1 97 6 4
SPARC 2 118 35
WT 1 33 87

Using Ring, Intensity Profile, Western Quadrant Histogram, and Segmented Least Squares Fitting features

The introduction of the segmented least squares fitting feature increased the number of correct identifications and reduced the confusion observed using only the ring and intensity curve features. The accuracy using this set of features was of 82.72% and the confusion matrix is shown in table 5.

Table 5. Confusion matrix for the results using ring, intensity profile, western quadrant.

SC1 SPARC WT
SC1 104 1 2
SPARC 1 122 32
WT 0 30 91

Discussion

A novel neural network approach for analysis of mouse lenses with different genotypes was developed using four types of features: the ring pattern, the intensity profile, the western quadrant histogram, and segmented least squares. The first two features were of major importance while the last two were additions that improved the accuracy of the analysis. The natural layering of transparent fibers in the lens varies with the genotype of the mice. The intensity profile is a feature that captured information from the rings, and the FFT analyzed the pattern formed by all the rings collectively not locally. This novel analysis permits categorization of large data bases of various types and stages of opacification on the basis of spatial and temporal differences in the pattern of opacities of the lens. Image analysis that accurately and consistently uses recognition of ring and intensity patterns in the identification of the genotype of mice can be expected to add a simple, new, automated and standardized method for computerized evaluation of patterns of early stage opacity and to establish phenotype – genotype relationships between factors contributing lens structure. The preliminary experimental data from tables 1-5 established the success of the new approach in distinguishing patterns of opacity in the SC1 null class from patterns of opacity in the SPARC null and WT classes with a 98.95% accuracy when four features were used. It was reasonably successful in distinguishing between the SPARC null and WT classes with a 77.45% accuracy. Overall, the analysis achieved a 82.76% accuracy. Repeated application of the algorithm to the same database produced the same results. The result was an impressive demonstration of a very promising method for an automated computer analysis of lens images that distinguishes selective categories of genetic mutation. The success of the algorithm is remarkable when considering the fact that key factors including age and cellular microstructure were not controlled, which suggests fundamental structural patterns remain at all ages. It needs to be appreciated that slit lamp images are low resolution. The underlying cellular basis for the slit lamp phenotype can only be determined by detailed and systematic microscopic analyses (Kuszak and Costello, 2004; Kuszak et al., 2006; Blankenship et al., 2007). The standard deviation in phenotype identification rates for the 10 fold-cross validation using all the features was 0.015 and shows a robust system with the current training samples. The results provide baseline data for improved informatics approaches for establishing phenotype/genotype relationships using slit lamp images of eyes in living animals and encourage continued research on automated and quantitative screening of phenotypes in large image databases.

Figure 7.

Figure 7

An iterative process selects a center point, isolates an elliptical sector of the lens which was mapped to a rectangle. The variance at each column of the rectangular map was computed, and the case where the sum of the column variances was the lowest was selected as the best ellipse sector isolation; its center corresponds to the best center of the lens.

Figure 8.

Figure 8

The ring extraction process allowed detection of the lens center and characterization of the lens rings as observed in the fitted elliptical arcs.

Figure 10.

Figure 10

Western quadrant of the lens.

Figure 12.

Figure 12

The non-continuous changes in the intensity profile were not well characterized when fit with the intensity profile to a continuous function. The upper graph was the intensity profile for the SPARC null lens on the top right and the bottom graph was the intensity profile for the WT, shown on the bottom right. The non-continuous fluctuations of intensities in the image were observed as sharp edges in the intensity profile that were characteristic in SPARC null, while their presence was not as prominent in the WT class.

Figure 15.

Figure 15

Percentage of correct identification using different combinations of features.

Figure 16.

Figure 16

A WT (left) and a SPARC null (right) image. Neither the expert nor the computer identified the mutants.

Acknowledgments

Support is gratefully acknowledged from LM000714 to JFB & LGS, EY13180 to JIC & EHS from the NIH and IIS0097329 to LGS, YL and JY from the NSF.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

  1. Alizadeh A, Clark J, Seeberger T, Hess J, Blankenship T, FitzGerald PG. Characterization of a mutation in the lens-specific CP49 in the 129 strain of mouse. Investigative ophthalmology & visual science. 2004;45:884–891. doi: 10.1167/iovs.03-0677. [DOI] [PubMed] [Google Scholar]
  2. Ansari RR, King JF, Seeberger T, Clark JI. Early detection of cataract and response to pantethine therapy with non-invasive static and dynamic light scattering. Proceedings of SPIE, Ophthalmic Technology XII. 2003;4951:168–178. [Google Scholar]
  3. Bassuk JA, Birkebak T, Rothmier JD, Clark JM, Bradshaw A, Muchowski PJ, Howe CC, Clark JI, Sage EH. Disruption of the Sparc locus in mice alters the differentiation of lenticular epithelial cells and leads to cataract formation. Experimental eye research. 1999;68:321–331. doi: 10.1006/exer.1998.0608. [DOI] [PubMed] [Google Scholar]
  4. Benedek G. Why the eye lens is transparent. Nature. 1983;302:383–384. doi: 10.1038/302383a0. [DOI] [PubMed] [Google Scholar]
  5. Benedek GB, Chylack LT, Jr, Libondi T, Magnante P, Pennett M. Quantitative detection of the molecular changes associated with early cataractogenesis in the living human lens using quasielastic light scattering. Current eye research. 1987;6:1421–1432. doi: 10.3109/02713688709044506. [DOI] [PubMed] [Google Scholar]
  6. Bettelheim FA, Churchill AC, Zigler JS., Jr On the nature of hereditary cataract in strain 13/N guinea pigs. Current eye research. 1997;16:917–924. doi: 10.1076/ceyr.16.9.917.5047. [DOI] [PubMed] [Google Scholar]
  7. Blankenship T, Bradshaw L, Shibata B, Fitzgerald P. Structural specializations emerging late in mouse lens fiber cell differentiation. Investigative ophthalmology & visual science. 2007;48:3269–3276. doi: 10.1167/iovs.07-0109. [DOI] [PubMed] [Google Scholar]
  8. Brekken RA, Sage EH. SPARC, a matricellular protein: at the crossroads of cell-matrix communication. Matrix Biol. 2001;19:816–827. doi: 10.1016/s0945-053x(00)00133-5. [DOI] [PubMed] [Google Scholar]
  9. Chenault VM, Ediger MN, Ansari RR. In vivo assessment of diabetic lenses using dynamic light scattering. Diabetes technology & therapeutics. 2002;4:651–659. doi: 10.1089/152091502320798295. [DOI] [PubMed] [Google Scholar]
  10. Chylack LT, Jr, Wolfe JK, Friend J, Khu PM, Singer DM, McCarthy D, del Carmen J, Rosner B. Quantitating cataract and nuclear brunescence, the Harvard and LOCS systems. Optom Vis Sci. 1993a;70:886–895. doi: 10.1097/00006324-199311000-00005. [DOI] [PubMed] [Google Scholar]
  11. Chylack LT, Jr, Wolfe JK, Singer DM, Leske MC, Bullimore MA, Bailey IL, Friend J, McCarthy D, Wu SY. The Lens Opacities Classification System III. The Longitudinal Study of Cataract Study Group. Archives of ophthalmology. 1993b;111:831–836. doi: 10.1001/archopht.1993.01090060119035. [DOI] [PubMed] [Google Scholar]
  12. Chylack LT, Jr, Wolfe JK, Friend J, Tung W, Singer DM, Brown NP, Hurst MA, Kopcke W, Schalch W. Validation of methods for the assessment of cataract progression in the Roche European-American Anticataract Trial (REACT) Ophthalmic epidemiology. 1995;2:59–75. doi: 10.3109/09286589509057085. [DOI] [PubMed] [Google Scholar]
  13. Chylack LT, Jr, Khu PM. Subjective classification & objective quantitation of human cataract. In: Albert DFJ, editor. Principles & Practice of Ophthalmology. W.B. Saunders Co; Philadelphia, PA: 2001. pp. 1449–1463. [Google Scholar]
  14. Clark JI, Carper D. Phase separation in lens cytoplasm is genetically linked to cataract formation in the Philly mouse. Proceedings of the National Academy of Sciences of the United States of America. 1987;84:122–125. doi: 10.1073/pnas.84.1.122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Clark JI. Order and disorder in the transparent media of the eye. Experimental eye research. 2004;78:427–432. doi: 10.1016/j.exer.2003.10.008. [DOI] [PubMed] [Google Scholar]
  16. Datiles MB, Edwards PA, Trus BL, Green SB. In vivo studies on cataracts using the Scheimpflug slit lamp camera. Investigative ophthalmology & visual science. 1987;28:1707–1710. [PubMed] [Google Scholar]
  17. Datiles MB, 3rd, Ansari RR, Reed GF. A clinical study of the human lens with a dynamic light scattering device. Experimental eye research. 2002;74:93–102. doi: 10.1006/exer.2001.1106. [DOI] [PubMed] [Google Scholar]
  18. Dimock J, Robman LD, McCarty CA, Taylor HR. Digital cataract photography. The Journal of audiovisual media in medicine. 2000;23:7–11. doi: 10.1080/014051100101790. [DOI] [PubMed] [Google Scholar]
  19. Edwards PA, Datiles MB, Unser M, Trus BL, Freidlin V, Kashima K. Computerized cataract detection and classification. Current eye research. 1990;9:517–524. doi: 10.3109/02713689008999591. [DOI] [PubMed] [Google Scholar]
  20. Fan S, Dyer CR, Hubbard L, Klein B. An automatic system for classification of nuclear sclerosis from slit-lamp photographs. In: Peters TM, E RE, editors. MICCAI 2003. LNCS 2878. Springer-Verlag; 2003. pp. 592–601. [Google Scholar]
  21. Goldstein LE, Muffat JA, Cherny RA, Moir RD, Ericsson MH, Huang X, Mavros C, Coccia JA, Faget KY, Fitch KA, Masters CL, Tanzi RE, Chylack LT, Jr, Bush AI. Cytosolic beta-amyloid deposition and supranuclear cataracts in lenses from people with Alzheimer's disease. Lancet. 2003;361:1258–1265. doi: 10.1016/S0140-6736(03)12981-9. [DOI] [PubMed] [Google Scholar]
  22. Graw J. Mouse models of congenital cataract. Eye (London, England) 1999;13(Pt 3b):438–444. doi: 10.1038/eye.1999.118. [DOI] [PubMed] [Google Scholar]
  23. Graw J. Congenital hereditary cataracts. The International journal of developmental biology. 2004;48:1031–1044. doi: 10.1387/ijdb.041854jg. [DOI] [PubMed] [Google Scholar]
  24. Harding J. Cataract: Biochemistry, epidemiology, and pharmacology. Chapman and Hall; London, UK: 1991. [Google Scholar]
  25. Kashima K, Trus BL, Unser M, Edwards PA, Datiles MB. Aging studies on normal lens using the Scheimpflug slit-lamp camera. Investigative ophthalmology & visual science. 1993a;34:263–269. [PubMed] [Google Scholar]
  26. Kashima K, Unser M, Datiles MB, Trus BL, Edwards PA. Minimum views required to characterize cataracts when using the Scheimpflug camera. Graefe's archive for clinical and experimental ophthalmology = Albrecht von Graefes Archiv fur klinische und experimentelle Ophthalmologie. 1993b;231:687–691. doi: 10.1007/BF00919282. [DOI] [PubMed] [Google Scholar]
  27. Klein BE, Hubbard L, Ferrier NJ, Klein R, Klein DJ, Lee KE, Ewen A, Jensen K, Evans MD. Detecting progression of nuclear sclerosis by using human grading versus semiautomated computer grading. Investigative ophthalmology & visual science. 2005;46:1155–1162. doi: 10.1167/iovs.04-0239. [DOI] [PubMed] [Google Scholar]
  28. Kojima M, Okuno T, Miyakoshi M, Sasaki K, Takahashi N. Environmental temperature and cataract progression in experimental rat cataract models. Developments in ophthalmology. 2002;35:125–134. doi: 10.1159/000060816. [DOI] [PubMed] [Google Scholar]
  29. Kuszak JR, Costello MJ. In: The Structure of the Vertebrate Lens in Development of the Ocular Lens. Lovicu FJ, Robinson JL, editors. Cambridge University Press; Cambridge, UK: 2004. pp. 71–118. [Google Scholar]
  30. Kuszak JR, Mazurkiewicz M, Zoltoski R. Computer modeling of secondary fiber development and growth: I. Nonprimate lenses. Molecular Vision. 2006:251–270. [PubMed] [Google Scholar]
  31. Leske MC, Chylack LT, Jr, Wu SY, Schoenfeld E, He Q, Friend J, Wolfe J. Incidence and progression of nuclear opacities in the Longitudinal Study of Cataract. Ophthalmology. 1996;103:705–712. doi: 10.1016/s0161-6420(96)30625-8. [DOI] [PubMed] [Google Scholar]
  32. Nishio I, Weiss JN, Tanaka T, Clark JI, Giblin FJ, Reddy VN, Benedek GB. In vivo observation of lens protein diffusivity in normal and X-irradiated rabbit lenses. Experimental eye research. 1984;39:61–68. doi: 10.1016/0014-4835(84)90115-5. [DOI] [PubMed] [Google Scholar]
  33. Norose K, Lo WK, Clark JI, Sage EH, Howe CC. Lenses of SPARC-null mice exhibit an abnormal cell surface-basement membrane interface. Experimental eye research. 2000;71:295–307. doi: 10.1006/exer.2000.0884. [DOI] [PubMed] [Google Scholar]
  34. Phelps-Brown NA, Bron AJ. Lens Disorders: A clinical manual of cataract diagnosis. Butterworth Heinemann; Oxford, UK: 1996. [Google Scholar]
  35. Rodriguez IR, Gonzalez P, Zigler JS, Jr, Borras T. A guinea-pig hereditary cataract contains a splice-site deletion in a crystallin gene. Biochimica et biophysica acta. 1992;1180:44–52. doi: 10.1016/0925-4439(92)90025-i. [DOI] [PubMed] [Google Scholar]
  36. Russell SPN. Prentice Hall Series on Artificial Intelligence. Prentice Hall; Englewood Cliffs, NJ: 1995. Artificial intelligence: a modern approach. [Google Scholar]
  37. Seeberger TM, Matsumoto Y, Alizadeh A, Fitzgerald PG, Clark JI. Digital image capture and quantification of subtle lens opacities in rodents. Journal of biomedical optics. 2004;9:116–120. doi: 10.1117/1.1630034. [DOI] [PubMed] [Google Scholar]
  38. Simpanya MF, Ansari RR, Suh KI, Leverenz VR, Giblin FJ. Aggregation of lens crystallins in an in vivo hyperbaric oxygen guinea pig model of nuclear cataract: dynamic light-scattering and HPLC analysis. Investigative Ophthalmology & visual science. 2005;46:4641–4651. doi: 10.1167/iovs.05-0843. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Sullivan MM, Sage EH. Hevin/SC1, a matricellular glycoprotein and potential tumor-suppressor of the SPARC/BM-40/Osteonectin family. The international journal of biochemistry & cell biology. 2004;36:991–996. doi: 10.1016/j.biocel.2004.01.017. [DOI] [PubMed] [Google Scholar]
  40. Sullivan MM, Barker TH, Funk SE, Karchin A, Seo NS, Hook M, Sanders J, Starcher B, Wight TN, Puolakkainen PA, Sage EH. Matricellular hevin regulates decorin production and collagen assembly. Journal of biological chemistry. 2006 doi: 10.1074/jbc.M510507200. [DOI] [PubMed] [Google Scholar]
  41. Thurston GM, Hayden DL, Burrows P, Clark JI, Taret VG, Kandel J, Courogen M, Peetermans JA, Bowen MS, Miller D, Sullivan KM, Storb R, Stern H, Benedek GB. Quasielastic light scattering study of the living human lens as a function of age. Current eye research. 1997;16:197–207. doi: 10.1076/ceyr.16.3.197.15410. [DOI] [PubMed] [Google Scholar]
  42. Thylefors B, Chylack LT, Jr, Konyama K, Sasaki K, Sperduto R, Taylor HR, West S. A simplified cataract grading system. Ophthalmic epidemiology. 2002;9:83–95. doi: 10.1076/opep.9.2.83.1523. [DOI] [PubMed] [Google Scholar]
  43. Tumminia SJ, Clark JI, Richiert DM, Mitton KP, Duglas-Tabor Y, Kowalak JA, Garland DL, Russell P. Three distinct stages of lens opacification in transgenic mice expressing the HIV-1 protease. Experimental eye research. 2001;72:115–121. doi: 10.1006/exer.2000.0934. [DOI] [PubMed] [Google Scholar]
  44. Wolf N, Penn P, Pendergrass W, Van Remmen H, Bartke A, Rabinovitch P, Martin GM. Age-related cataract progression in five mouse models for anti-oxidant protection or hormonal influence. Experimental eye research. 2005;81:276–285. doi: 10.1016/j.exer.2005.01.024. [DOI] [PubMed] [Google Scholar]
  45. Yan Q, Clark JI, Wight TN, Sage EH. Alterations in the lens capsule contribute to cataractogenesis in SPARC-null mice. Journal of cell science. 2002;115:2747–2756. doi: 10.1242/jcs.115.13.2747. [DOI] [PubMed] [Google Scholar]
  46. Witten IH, Eibe F. Data Mining: Practical machine learning tools and techniques. San Francisco, CA: Morgan Kaufmann; 2005. [Google Scholar]

RESOURCES