Abstract
Glaucoma is the leading cause of preventable blindness in the western world. Investigation of high-resolution retinal nerve fiber layer (RNFL) images in patients may lead to new indicators of its onset. Adaptive optics (AO) can provide diffraction-limited images of the retina, providing new opportunities for earlier detection of neuroretinal pathologies. However, precise processing is required to correct for three effects in sequences of AO-assisted, flood-illumination images: uneven illumination, residual image motion and image rotation. This processing can be challenging for images of the RNFL due to their low contrast and lack of clearly noticeable features. Here we develop specific processing techniques and show that their application leads to improved image quality on the nerve fiber bundles. This in turn improves the reliability of measures of fiber texture such as the correlation of Gray-Level Co-occurrence Matrix (GLCM).
OCIS codes: (170.4470) Ophthalmology, (010.1080) Active or adaptive optics, (100.2960) Image analysis
1. Introduction
Retinal imaging provides vital information for the study, diagnosis and treatment of retinal pathologies. This information can be obtained non-invasively using ophthalmoscopes. In recent years, ophthalmic devices have been integrated with adaptive optics (AO) technology to provide high resolution retinal images that reveal retinal microstructures. Highly accurate and reliable image processing techniques are necessary for the development of automated diagnostic tools that will facilitate the use of AO retinal imaging technology in regular clinical practice [1, 2].
The RNFL is formed by retinal ganglion cell axons and represents the innermost layer of the retina. The ganglion cell axons are spread out as a thin layer with striations (nerve fiber bundles). Retinal Nerve Fiber Layer Defects (RNFLD) can be used to diagnose glaucoma, which is the leading cause of preventable blindness across the western world [3–5]. A recent clinical review of glaucoma refers to it as a group of pathological eye conditions with several causes that, usually associated with ocular hypertension, result in damage to the optic nerve head (ONH) and loss of visual field (VF) [6]. The World Health Organization estimated that in 2010 glaucoma accounted for 2% of visual impairment and 8% of global blindness [7]. Loss of axonal tissue in the RNFL has been reported to be one of the earliest detectable glaucomatous changes, which can proceed to morphological changes in the ONH and VF loss [8, 9]. For this reason, many studies have focused on thinning of the RNFL and RNFLD using various imaging technologies [10, 11]. It is therefore essential to obtain high quality RNFL images for early and accurate diagnosis.
Adding AO to imaging systems such as flood-illuminated ophthalmoscopes, Scanning Laser Ophthalmoscopes (SLO) and OCT devices has recently allowed researchers to identify individual nerve fiber bundles, providing high-resolution images of both the RNFL and ONH [1, 2, 9, 12–15]. In this work, sequences of RNFL images are obtained using a commercial AO-assisted flood illumination device. The original RNFL images are often corrupted by intensity variations referred to as shading or intensity inhomogeneity. This is due to inherent imperfections of the image formation process such as non-uniform illumination, uneven spatial sensitivity of the sensor or camera imperfections [16]. It affects automatic image processing, such as segmentation, registration and characterization of retinal features [17]. We have previously described a novel wavelet based method of pre-processing for the correction of uneven illumination in AO flood retinal images [18]. Nevertheless, the final image quality is also limited by motions of the eye that occur on short time scales. These motions can lead to blurring and distortion of the retinal images. More complex warping is considered in the registration of SLO images due to the fact that the image is acquired point-by-point while the eye is moving [19, 20], and is not expected to apply to the snap-shot fundus imaging carried out here. In the case of flood illumination systems, rigid transformations to correct for global translation and rotation will be applicable. In medical image processing, cross-correlation techniques are usually used to measure displacements between images. These use image intensities for direct matching, without the need to detect image features. Cross-correlation has been used already to register images obtained using various AO assisted flood illumination retinal cameras [21–23]. Although the cross-correlation approach can work well, its performance can be severely compromised by factors such as changes in the image intensity, noise, and reduced overlap area when there is large image motion [24]. In this work, we show that the phase correlation technique is more robust than cross-correlation for the estimation of translational motion [25]. However, both correlation methods fail when there are changes in scale or rotation. Various methods have been described in the literature for the estimation of rotational motion. These include bilateral matching between local features of Harris-corner points detected in both input and reference images to measure rotation [26], and the Radon transform [27, 28]. The combination of Fourier based techniques and log-polar transformations has also been applied to measure rotation, translation and scaling [29–32]. Although many techniques claim to accurately measure rotational motion, we have encountered two main issues: (i) most commonly a reference image is matched against a rotated version of itself and (ii) they are intended to measure large angles of rotation (typically >1-2°) with relatively low accuracy. These techniques fail when estimating small rotation angles and/or the image quality or content is changing. In an earlier study, we found that the maximum value of rotation in the retinal images does not exceed 1°, which for an image size of 1k pixels, corresponds to 8-9 pixels at the edge of the image [33].
Our earlier work with AO retinal images focused on the cone photoreceptor layer using a peak tracking approach to correct for rotation [18], which demonstrated the improvement in image quality resulting from correcting rotation for angles less than 1°. On the other hand, the RNFL image sequence does not have clearly noticeable cone photoreceptors and the retinal vessels have natural movement or pulsing. Hence methods based on local features, control points or peak tracking are not optimal to estimate the rotation. In this work, we applied and compared approaches to register and de-rotate the RNFL image sequences, which included the Radon transform, log-polar transform and an exhaustive search. The effectiveness of the registration approaches was evaluated using the gray level co-occurrence matrix (GLCM). This technique provided a useful measure of the structural contrast of the nerve fiber layers and has the potential to be useful as an imaging bio-marker for glaucoma.
2. Methods
The sequences of images used here were obtained using a commercial AO-assisted fundus imager (rtx1, Imagine Eyes, Orsay, France). The deformable mirror of the rtx1 was used to select the appropriate plane of focus. The focal plane was adjusted to acquire images of the RFNL close to the ONH. Each sequence consisted of a series of 40 frames. Each frame had an exposure time of 9 ms and there was an interval of 105 ms between each frame in the series. This work was part of study protocol on the clinical application of AO retinal imaging approved by the local ethical committee (ASL RMA, Rome, Italy). All research procedures adhered to the tenets of the Declaration of Helsinki. Images of the RFNL were acquired from healthy volunteers older than 18 years old. For the purpose of this work, intended to disclose an effective method of registration of AO flood RFNL images, the results are shown only for one case.
Uneven illumination of the RFNL image was corrected before registration. A wavelet based approach was used to remove the variations in the uniformity of the illumination of the retina. The result of the wavelet based approach was compared with homomorphic filtering and spatial filtering methods, as described in previous work [18]. The quality of each image was analyzed in order to select the sharpest image in the sequence, which was used as a reference image during registration of the images. The sharpness metric, described by Fienup & Miller [34], was used as a quality metric to measure the image sharpness.
In this study, translational motion has been measured using cross-correlation (CC) and phase correlation (PC) techniques while rotational motion has been determined using Radon, log-polar and exhaustive search approaches. The cross-correlation approach has been described in many publications [35]; it is used to find the translation (or more complex transformation) which maximizes the cross-correlation between a template image and the images to be registered. In this work, we used the image with the highest sharpness value as the template. In phase correlation, the inverse Fourier transform of the cross-power spectrum of the template image and the image to be registered gives a delta function at a position corresponding to the required translation [31]. An advantage of this approach is that filtering in the Fourier domain can be effectively implemented to reduce the effect of noise.
2.1 Radon transform
For an arbitrary function f(x,y) the Radon transform is defined by [36],
(1) |
where Lr,θ is a line making an angle θ with the y axis at a distance r from the origin. If an image f(x,y) is rotated through an angle ψ then the Radon transform of the rotated image is given by:
(2) |
This means that rotation of an image corresponds to translation on the angular variable (θ) of the Radon transform. This translation is estimated using cross-correlation. The Radon transform of the input images is carried out using the Matlab Radon function.
2.2 Log-polar transform
The log-polar transformation is a nonlinear and non-uniform sampling of Cartesian co-ordinates to log-polar co-ordinates. In the log-polar transformation, radial lines in Cartesian space are mapped into horizontal lines, while arcs are mapped to vertical lines in the polar coordinate space. The pixels in a log-polar sampled image can be characterized by the ring number R(where) and the wedge or angular number W(where ) given by [37]:
(3) |
(4) |
where r is the radial distance from the centre (xc,yc) and θ is the angle, nr is the number of rings, nw is the number of wedges or sectors, rmin is the radius of the smallest ring in the log-polar samples and rmax is the radius of the largest ring in the log-polar samples. In general, the choice of values for the parameters of the log-polar transform is not obvious. In this work, Young’s method [37] was used to decide the parameters of the log-polar grid. In order to measure rotational motion between two log-polar converted images (i.e., reference and input), cross-correlation or phase correlation was applied. We refer to the first approach as rotation angle calculated using cross-correlation on the log-polar transformed image (LPNCC). The combination of log-polar representation with phase correlation for images which are translated and/or rotated is called the Fourier-Mellin (FM) transform [29].
2.3 Exhaustive search
We have implemented another approach to finding the rotation between the reference and the input images, which we refer to as exhaustive search (ES). In this work, we used ES only to measure rotation because translation has already been estimated and scale changes were found to be negligible. The input image, I2, which is to be aligned with reference image I1 is rotated over a range of angles, from −1° to + 1°. For each angle, the phase correlation between the input and reference images is obtained, and the angle θcoarse corresponding to the highest peak value is obtained. The angular step size is now reduced (typically by a factor of 10) and a new search carried out about θcoarse. The procedure is repeated until convergence or until a minimum step size (typically 0.01°) is reached. The accuracy of the method improves as the step size is reduced, but the array size and computation time increase. We refer to this method as exhaustive search phase correlation (ESPC).
2.4 Texture analysis
In order to investigate the effectiveness of the registration approaches, we performed texture analysis on the final RNFL images with translation correction alone and with translation correction and de-rotation using ESPC. We used the GLCM for this purpose [38]. The co-occurrence matrix of an image is defined as the distribution of co-occurring values at a given offset and angle. The co-occurrence matrix C for an n × m image I, parameterized by an offset (Δx, Δy), is defined as:
(5) |
where i and j are the image intensity values of the image, p and q are the spatial positions in the image I and the offset (Δx,Δy) depends on the direction θ used and the distance d at which the matrix is computed. The (Δx, Δy) parameterization makes the co-occurrence matrix sensitive to rotation. Choosing an offset vector, such that the rotation of the image is not equal to 180°, will result in a different co-occurrence matrix for the same (rotated) image. This can be avoided by forming the co-occurrence matrix using a set of offsets sweeping through 180° (i.e., 0, 45, 90, and 135 degrees) at the same distance to achieve a degree of rotational invariance. For images comprising repetitive texture patterns, the correlation of GLCM measured at four directions exhibits periodic behavior with a period equal to the spacing between adjacent texture primitives. When the texture is coarse, the correlation function drops off slowly, whereas for fine textures it drops off rapidly. The correlation of GLCM is used as a measure of periodicity of texture as well as a measure of the scale of the texture primitives [39]. The GLCM has been calculated using the Matlab graycomatrix function.
3. Results
In order to understand the effectiveness of different methods of uneven illumination correction, the average images obtained after correcting the translation (using PC) and rotational motion (using ESPC) and the corresponding correlation of GLCM have been calculated, as shown in Fig. 1 . The correlation curves of GLCM for the mean-subtraction and homomorphic methods of uneven illumination correction (Fig. 1(d), and 1(e)) show a slow fall-off and do not have periodic ripples corresponding to the nerve fiber bundles. It indicates that these two methods are not suitable for RNFL images when-analyzing the nerve fiber bundles. On the contrary, the Daubechies wavelets at decomposition level 3-4 provided the best illumination correction while maintaining the image information. The average registered image using wavelet based uneven illumination correction and its correlation of GLCM are shown in Fig. 1(c) and 1(f) respectively. The correlation curves clearly show the repeated pattern and contain deep fall-off at 135° orientation.
We have then carried out the image registration in two stages: (i) estimation of translation and (ii) estimation of rotation. In order to estimate the translation, (i) the sharpest image of the sequence has been used as the reference image and (ii) both cross-correlation and phase correlation with parabolic interpolation were used for sub-pixel measurement. The result of correlation techniques for measuring horizontal and vertical translational displacements is shown in Fig. 2 .
Visual examination of the registered sequence of images confirmed that the discrepancies between the two techniques occur when cross-correlation fails to correctly measure displacements (in this case for 11 of the 40 frames, 27.5%). We have also found that the translations measured using cross-correlation are sensitive to the size of the window used and the method of background correction. The best results were obtained when a window of side length approximately 75% of the total image side and wavelet background correction were used. We observed visually that the result of translation correction using phase correlation was better than cross-correlation in terms of clear visibility of RNFL structures, as shown in Fig. 3(a) and 3(b) respectively.
Although the integrated result of both methods looks visually similar (Fig. 3(a)-3(b)), the auto-correlation function (ACF) of each of the average images shown in Fig. 3(c)-3(d) indicates that the ACF from registration using phase correlation has very clear structure compared with using cross-correlation. An intensity line profile taken from the ACF of both images shows that the image striations, corresponding to the retinal nerve fiber bundles, are clearer when phase correlation is used for registration (Fig. 3(e)). In the case of cross-correlation, most of the images corresponding to incorrect residual motion have been identified as having poor quality, as shown in Fig. 4 . However, the phase correlation method measured the displacements for them too, which demonstrates its robustness.
In the case of rotation, we compared four approaches: (a) Radon transform (b) log-polar transform, (c) Fourier-Mellin and (d) exhaustive search. We firstly tested these methods for a single retinal frame rotated by a set of known values (test image) and then applied them to a full retinal image sequence with unknown rotations.
For the test image sequence, all of the techniques performed well down to an angular step size of 0.1°. In the case of a real retinal image sequence, the results of each approach are shown in Fig. 5 . The angle measured using the Radon transform is shown in Fig. 5(a). The rotation angles were applied to de-rotate the image sequence. However, residual rotation (even >1°) is visually evident, indicating that the measurement is not correct. We have also applied the measured angle of rotation from ESPC to de-rotate the image sequence and found that the residual rotation appears negligible. We therefore consider the ESPC approach as a standard while evaluating registration accuracy. The LPNCC and FM approaches have been compared with the ESPC method, as shown in Fig. 5(b). The rotation angle calculated using LPNCC and ESPC are very similar; while the FM approach seems to be significantly different (≥0.9°) for several frames. Typically, LPNCC only fails for a few poor quality frames. The RMS difference between the rotation angles measured using ESPC and LPNCC was 0.007° whereas it was 0.09° between ESPC and FM. The resultant average image after translation correction using PC plus rotation correction using ESPC is shown in Fig. 6(a) .
In order to assess the improvement in correcting rotation, a section of the translation corrected and de-rotated images of size 150 x 150 pixels was taken at the bottom left corner of the image, as shown in Fig. 6(b) and Fig. 6(c) respectively. These sampling windows have retinal nerve fiber bundles which visually appear oriented between 0° and 45°. In analyzing the fiber texture, we have used the GLCM as described in section 2.
Figure 7 shows the correlation of the GLCM at four directions (0°, 45°, 90° and 135°) corresponding to the final images obtained with translation correction alone and with translation correction and de-rotation using ESPC. Since the nerve fiber bundles are oriented along approximately 45°, we expect the correlation to fall off quickly at 135°. The comparison of results shows a deeper fall-off at 135° for the de-rotated sample area than for the translation corrected window. On the other hand, the GLCM of the translation-only corrected section (Fig. 7(a)) shows similar falloffs for orientations of 45° and 135°. The GLCM correlation was then obtained using different de-rotation methods: the Radon, log-polar and exhaustive search approaches have been compared and shown in Fig. 8 . The correlation of GLCM for the Radon transform approach (Fig. 7(a)) does not have the clear fall-off at 135° and does not exhibit the ripples which correspond to the retinal nerve fiber bundles. This is due to the incorrect measurement of the angle of rotation, as shown in Fig. 5(a). In contrast, the correlation curve of GLCM for an average image obtained using the LPNCC approach (Fig. 8(b)) shows a deep fall-off at 135° and ripples. This curve is very similar to the correlation curve for GLCM using the ESPC approach (Fig. 8(d)). Finally, the correlation of the GLCM for an average image obtained using FM is very similar to the ESPC result with a slightly less deep fall-off; the ripples, however, are not as clearly indicated as with the ESPC method.
4. Conclusion
The aim of this work was to identify the best approach for image registration of AO-flood illuminated RNFL images. We found that phase correlation is more robust than cross-correlation irrespective of the method of uneven illumination correction or size of the image while estimating the translational motion. The rotational motion has been measured using four different approaches: the Radon transform, LPNCC, FM and ESPC. We found that the result of measuring rotation using the ESPC approach is more accurate and robust irrespective of the method of uneven illumination correction or translation correction approach. The result of the ESPC approach has therefore been used as a standard to verify the registration accuracy and compared with the other methods of estimating rotation, as shown in Fig. 5. The result of LPNCC in measuring the angle of rotation is very close to ESPC.
In conclusion, the registration of RNFL images is best when the uneven illumination correction using the wavelet approach is combined with phase correlation for correcting translation and ESPC for correcting rotational motion. This is of particular interest, as the improvement obtained through rotational motion correction will be significant in the case of mosaicing images used to form a larger field of view, which aids improvement in diagnostic tools. Recently, a method using the Harris-Stephen interest point detector for AO flood-illuminated assisted images has been described [40]. The authors applied cross-correlation between small windows of detected feature-points followed by an affine model to minimize a least-square criterion between reference and input image in order to estimate translation, rotation and scale. In our study, the registration approach did not rely on retinal features as these may or may not be detectable in RNFL images. We found that LPNCC could be an effective approach for registering the RNFL images, which is computationally efficient. It fails only in the case of very poor quality frames. Currently, it takes approximately 3-4 minutes and 1-2 minutes on an 8-core PC for the best possible processing using the ESPC and LPNCC methods respectively, which could be considered a barrier for clinical implementation. However, the code has not been optimized for speed, and carrying this out, together with advances in computer hardware would certainly solve this problem.
We have proposed the correlation of GLCM to compare the effectiveness of the different processing techniques. The correlation of GLCM showed a sharp fall-off followed by periodic ripples that were clearest after the best processing (wavelet based uneven illumination correction + PC + ESPC) was carried out. Texture analysis based on correlation of GLCM was further shown to be potentially useful as a measure of the nerve fiber structure. The method retains information on spatial structures and for images comprising repetitive texture patterns, such as RNFL images, the correlation of GLCM measured at four directions show periodic behavior with the period equal to the spacing between adjacent texture primitives. The next stage of our work will involve applying this processing to RNFL images acquired from patients. AO imaging could potentially help to recognize early glaucomatous damage and to identify patients who could benefit from more intensive observation and management.
Acknowledgments
This research was supported by the National University of Ireland (Galway) funding (Devaney, Ramaswamy) and by the 5x1000 funding (Italy) for scientific research and universities (Lombardo, Ramaswamy).
References and links
- 1.Carroll J., Kay D. B., Scoles D., Dubra A., Lombardo M., “Adaptive optics retinal imaging--clinical opportunities and challenges,” Curr. Eye Res. 38(7), 709–721 (2013). 10.3109/02713683.2013.784792 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Lombardo M., Serrao S., Devaney N., Parravano M., Lombardo G., “Adaptive optics technology for high-resolution retinal imaging,” Sensors (Basel) 13(1), 334–366 (2013). 10.3390/s130100334 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Quigley H. A., “Glaucoma,” Lancet 377(9774), 1367–1377 (2011). 10.1016/S0140-6736(10)61423-7 [DOI] [PubMed] [Google Scholar]
- 4.Quigley H. A., Broman A. T., “The number of people with glaucoma worldwide in 2010 and 2020,” Br. J. Ophthalmol. 90(3), 262–267 (2006). 10.1136/bjo.2005.081224 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Quigley H. A., Sommer A., “How to use nerve fiber layer examination in the management of glaucoma,” Trans. Am. Ophthalmol. Soc. 85, 254–272 (1987). [PMC free article] [PubMed] [Google Scholar]
- 6.King A., Azuara-Blanco A., Tuulonen A., “Clinical review: Glaucoma,” BMJ 346, 29–33 (2013). 10.1136/bmj.f3518 [DOI] [PubMed] [Google Scholar]
- 7.WHO , “Global data on visual impairments 2010.” www.who.int/blindness/ GLOBALDATAFINALforweb.pdf
- 8.Alencar L. M., Zangwill L. M., Weinreb R. N., Bowd C., Sample P. A., Girkin C. A., Liebmann J. M., Medeiros F. A., “A comparison of rates of change in neuroretinal rim area and retinal nerve fiber layer thickness in progressive glaucoma,” Invest. Ophthalmol. Vis. Sci. 51(7), 3531–3539 (2010). 10.1167/iovs.09-4350 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Kocaoglu O. P., Cense B., Jonnal R. S., Wang Q., Lee S., Gao W., Miller D. T., “Imaging retinal nerve fiber bundles using optical coherence tomography with adaptive optics,” Vision Res. 51(16), 1835–1844 (2011). 10.1016/j.visres.2011.06.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Lim T. C., Chattopadhyay S., Acharya U. R., “A survey and comparative study on the instruments for glaucoma detection,” Med. Eng. Phys. 34(2), 129–139 (2012). 10.1016/j.medengphy.2011.07.030 [DOI] [PubMed] [Google Scholar]
- 11.Mansouri K., Leite M. T., Medeiros F. A., Leung C. K., Weinreb R. N., “Assessment of rates of structural change in glaucoma using imaging technologies,” Eye (Lond.) 25(3), 269–277 (2011). 10.1038/eye.2010.202 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Akagi T., Hangai M., Takayama K., Nonaka A., Ooto S., Yoshimura N., “In vivo imaging of lamina cribrosa pores by adaptive optics scanning laser ophthalmoscopy,” Invest. Ophthalmol. Vis. Sci. 53(7), 4111–4119 (2012). 10.1167/iovs.11-7536 [DOI] [PubMed] [Google Scholar]
- 13.Huang G., Qi X., Chui T. Y. P., Zhong Z., Burns S. A., “A clinical planning module for adaptive optics SLO imaging,” Optom. Vis. Sci. 89(5), 593–601 (2012). 10.1097/OPX.0b013e318253e081 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Ivers K. M., Li C., Patel N., Sredar N., Luo X., Queener H., Harwerth R. S., Porter J., “Reproducibility of measuring lamina cribrosa pore geometry in human and nonhuman primates with in vivo adaptive optics imaging,” Invest. Ophthalmol. Vis. Sci. 52(8), 5473–5480 (2011). 10.1167/iovs.11-7347 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Takayama K., Ooto S., Hangai M., Arakawa N., Oshima S., Shibata N., Hanebuchi M., Inoue T., Yoshimura N., “High-resolution imaging of the retinal nerve fiber layer in normal eyes using adaptive optics scanning laser ophthalmoscopy,” PLoS ONE 7(3), e33158 (2012). 10.1371/journal.pone.0033158 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.J. C. Russ, The image processing handbook (CRC/Taylor & Francis, 2007), Chap. 3. [Google Scholar]
- 17.Tomazevic D., Likar B., Pernus F., “Comparative evaluation of retrospective shading correction methods,” J. Microsc. 208(3), 212–223 (2002). 10.1046/j.1365-2818.2002.01079.x [DOI] [PubMed] [Google Scholar]
- 18.Ramaswamy G., Devaney N., “Pre-processing, registration and selection of adaptive optics corrected retinal images,” Ophthalmic Physiol. Opt. 33(4), 527–539 (2013). 10.1111/opo.12068 [DOI] [PubMed] [Google Scholar]
- 19.Li H., Lu J., Shi G., Zhang Y., “Tracking features in retinal images of adaptive optics confocal scanning laser ophthalmoscope using KLT-SIFT algorithm,” Biomed. Opt. Express 1(1), 31–40 (2010). 10.1364/BOE.1.000031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Vogel C. R., Arathorn D. W., Roorda A., Parker A., “Retinal motion estimation in adaptive optics scanning laser ophthalmoscopy,” Opt. Express 14(2), 487–497 (2006). 10.1364/OPEX.14.000487 [DOI] [PubMed] [Google Scholar]
- 21.Lombardo M., Serrao S., Ducoli P., Lombardo G., “Variations in image optical quality of the eye and the sampling limit of resolution of the cone mosaic with axial length in young adults,” J. Cataract Refract. Surg. 38(7), 1147–1155 (2012). 10.1016/j.jcrs.2012.02.033 [DOI] [PubMed] [Google Scholar]
- 22.Putnam N. M., Hofer H. J., Doble N., Chen L., Carroll J., Williams D. R., “The locus of fixation and the foveal cone mosaic,” J. Vis. 5(7), 632–639 (2005). 10.1167/5.7.3 [DOI] [PubMed] [Google Scholar]
- 23.Rha J., Jonnal R. S., Thorn K. E., Qu J., Zhang Y., Miller D. T., “Adaptive optics flood-illumination camera for high speed retinal imaging,” Opt. Express 14(10), 4552–4569 (2006). 10.1364/OE.14.004552 [DOI] [PubMed] [Google Scholar]
- 24.Gharabaghi S., Daneshvar S., Sedaaghi M. H., “Retinal image registration using geometrical features,” J. Digit. Imaging 26(2), 248–258 (2013). 10.1007/s10278-012-9501-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.C. D. Kuglin and D. C. Hines, “The Phase Correlation Image Alignment method,” in Proceedings of IEEE Cybernet Society (Institute of Electrical and Electronics Engineers, 1975), pp.163–165. [Google Scholar]
- 26.J. Chen, R. T. Smith, J. Tian, and A. F. Laine, “A Novel Registration method for Retinal Images based on Local Features,” in Proc. of IEEE Engineering in Medicine & Biology Society, (Institute of Electrical and Electronics Engineers, 2008), pp.2242–2245. 10.1109/IEMBS.2008.4649642 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.J. You, W. Lu, J. Li, G. Gindi, and Z. Liang, “Image matching for translation, rotation and uniform scaling by the radon transform,” in Proceedings of International Conference on Image Processing (Institute of Electrical and Electronics Engineers, 1998), pp.847–851. [Google Scholar]
- 28.Arodz T., “Invariant Object Recognition using Radon-based Transform,” Computing and Informatics 24, 183–199 (2005). [Google Scholar]
- 29.De Castro E., Morandi C., “Registration of Translated and Rotated Images using Finite Fourier Transforms,” IEEE Trans. Pattern Anal. Mach. Intell. 9(5), 700–703 (1987). 10.1109/TPAMI.1987.4767966 [DOI] [PubMed] [Google Scholar]
- 30.R. Matungka, Y. F. Zheng, and R. L. Ewing, “Object Recognition Using Log-Polar Wavelet Mapping,” in IEEE International Conference on Tools with Artificial Intelligence, (Institute of Electrical and Electronics Engineers, 2008), pp.559–563. [Google Scholar]
- 31.Reddy B. S., Chatterji B. N., “An FFT-based Technique for Translation, Rotation, and Scale-Invariant Image Registration,” IEEE Trans. Image Process. 5(8), 1266–1271 (1996). 10.1109/83.506761 [DOI] [PubMed] [Google Scholar]
- 32.G. Wolberg and S. Zokai, “Robust Image Registration using Log-Polar Transform,” in IEEE International Conference on Image Processing, (Institute of Electrical and Electronics Engineers, 2000), 493–496. [Google Scholar]
- 33.G. Ramaswamy, M. Lombardo, and N. Devaney, “Texture analysis of adaptive optics assisted retinal nerve fiber layer images,” in Photonics Ireland 2013, (Belfast, 2013). [Google Scholar]
- 34.Fienup J. R., Miller J. J., “Aberration correction by maximizing generalized sharpness metrics,” J. Opt. Soc. Am. A 20(4), 609–620 (2003). 10.1364/JOSAA.20.000609 [DOI] [PubMed] [Google Scholar]
- 35.Zitova B., Flusser J., “Image registration methods: a survey,” Image Vis. Comput. 21(11), 977–1000 (2003). 10.1016/S0262-8856(03)00137-9 [DOI] [Google Scholar]
- 36.S. Deans, The Radon Transform and some of its Applications (A Wiley-Interscience, 1983), Chap. 2. [Google Scholar]
- 37.Young D., “Straight lines and circles in the log-polar image,” in Proceedings of the 11th British Machine Vision Conference, 2000), pp.426–435 10.5244/C.14.43 [DOI] [Google Scholar]
- 38.Haralick R. M., “Statistical and Structural Approaches to Texture,” Proc. IEEE 67(5), 786–804 (1979). 10.1109/PROC.1979.11328 [DOI] [Google Scholar]
- 39.R. Jain, R. Kasturi, and B. G. Schunck, “Texture,” in Machine Vision (McGraw-Hill, Inc., 1995), pp. 234–248. [Google Scholar]
- 40.C. Kulcsar, G. L. Besnerais, E. Odlund, and X. Levecq, “Robust processing of images sequences produced by an adaptive optics retinal camera,” in Imaging and Applied Optics, OSA Technical Digest (online) (Optical Society of America, 2013), paper OW3A.3. http://www.opticsinfobase.org/abstract.cfm?URI=AOPT-2013-OW3A.3 10.1364/AOPT.2013.OW3A.3 [DOI]