Abstract
Adaptive optics(AO) has greatly improved retinal image resolution. However, even with AO, temporal and spatial variations in image quality still occur due to wavefront fluctuations, intra-frame focus shifts and other factors. As a result, aligning and averaging images can produce a mean image that has lower resolution or contrast than the best images within a sequence. To address this, we propose an image post-processing scheme called “lucky averaging”, analogous to lucky imaging (Fried, 1978) based on computing the best local contrast over time. Results from eye data demonstrate improvements in image quality.
Adaptive optics scanning laser ophthalmoscopy (AOSLO) provides high resolution in vivo retinal images and has become a reliable tool for studying the structure and function of the retina [1]. AOSLO systems typically provide real-time imaging, and thus record images over time. However, in almost all real time image sequences, image quality varies slightly over both space and time. That is, local regions have higher contrast in one frame than another. Based on the point-scanning mechanism of an AOSLO, the variance can be attributed to both time related and space related factors, most likely including: dynamic changes in human eye aberrations such as the tear film, local retinal shape changes due to ocular pulse, accommodative fluctuations, and variations in the retinal shape which can interact with the above factors. In addition, images from subjects who have retinal defects such as detachments, drusen and edema produce even more pronounced variations in the retinal shape, which interact to enhance these variations in image quality.
Conventionally, AOSLO raw images are post processed to decrease the impact of noise. Typically this is achieved by either manually or automatically choosing acceptable frames, omitting those with very large eye movements or blinks, then aligning to minimize the effect of both intra- and inter-frame movements [2], and finally averaging the aligned frames. The more frames used for the averaging, the greater the increase in signal to noise ratio. On the other hand, since the averaging is frame-based, the intra-frame contrast variance is not altered. Therefore the calculation of the mean can degrade the optimal image quality achieved from the image sequence. That is, the frame averaging is not taking the best advantage of the sequential nature of AOSLO imaging.
This problem is also true if recording views of the sky from a ground based telescope. By choosing only the best frames, images are selected when the atmospheric turbulence along the line of sight is minimal. That is, the process takes advantage of lucky moments when turbulence has a minimal effect and has therefore been called “lucky imaging” [3]. In analogy the AOSLO is measuring the retina point-by-point, and its high frequency pixel clock makes it well suited for this “lucky imaging” approach, but now within a frame as well as between frames.
Thus, we propose a region-based averaging scheme we call “lucky averaging”. Image acquisition and alignment are performed in the normal manner described above. However, before averaging, the image sequence is analyzed region by region. Small areas in the image represent short-exposure times and small spatial scales, greatly eliminating the intra-region variance in optical quality and enabling the process to capture those “lucky moments” when small region’s image quality is at its best. The region quality is evaluated using an appropriate objective image quality metric and assigned to the central pixel of the region. Each pixel becomes the center of a region, and consequently, pixel quality maps are built up in sequence. The final image is remapped by only including the times when each pixel has the highest image quality as evaluated using the metric. The scheme is shown in Fig. 1.
To test this approach, cone photoreceptor images were acquired from different subjects and processed. The AOSLO system has been described previously [4]. The pixel clock was 30MHz, and the frame rate was 30Hz. The AO control loop operated at 15Hz. Each image was 800 by 520 pixels, corresponding to 2.79°×1.78° imaging field in the retina. Each image series consisted of approximately 100 frames acquired sequentially in ~3 second. Once aligned, a window of 9 by 7 pixels was created and swept across the image, pixel-by-pixel to compute an image quality metric. A window of this size represented 450 μs in time and about 70 μm2 in space.
While any appropriate image quality metric can be used, for this paper we implemented two examples - a sharpness metric and a gray level co-occurrence matrix (GLCM). The sharpness metric has been used for incoherent imaging in astronomy and synthetic radar image reconstructions [5]. It represents a sum of a nonlinear point transformation of the image and is generally defined as
(1) |
where x,y are pixel coordinates, and β is the power of the pixel intensity, w(x,y) is the weight for each pixel, and I(x,y) is the pixel intensity.
In practice, we set β as 2, and w(x,y) as 1. Normalization is also considered, as [6] indicates. The metric is finally defined as,
(2) |
where x, y are pixel coordinates, x0 and y0 represent the coordinates of central pixels of the sampling window. Δx and Δy represent half the sampling window width and height respectively. The index represents the frame index number of the analyzed image within the sequence
The second image metric is GLCM, measuring how often different combinations of gray levels co-occur in an image region [7], computed as in equation (3). For a given Δd which represents pixel separation between a pair of pixels, we compute GLCMΔd by examining pixel pair along 8 different directions from 0 degree to 315 degree with 45 degree increment, where the element GLCMΔd (i,j) represents the number of the pixel pair with gray level pair(i,j). The parameter d represents the two dimensional position vector of the pixel in the image. We typically normalize the GLCM by dividing the whole matrix with the sum of all elements in the matrix. Once a GLCM has been constructed, several types of statistical analysis can be performed. Local contrast is one of these measures, taking GLCM as contrast weights, computed as in equation (4), where α is the power of the gray level difference between a pixel pair, and β is the power of GLCM. In practice, for cone photoreceptor images, we calculated GLCM3 or GLCM4. α and β are set as 2 and 1 respectively.
(3) |
(4) |
The calculation result S(x0, y0, frame index) from equation (2) or equation (4) is then assigned to the central pixel (x0,y0). A quality metric S(x, y, frame index) is then generated for every pixel in the image sequence.
For either metric the final step is to set a threshold for including a pixel. That is, for a given pixel position(x,y), we can include (for instance) the 30%of the instances within the sequence when those pixel instances have the highest value of the image quality metric. The final image is then the average of those pixels. This is computationally fairly simple, taking about 3 minutes for 100 frames, using a 2.8 GHz PC with 8GB of memory and implemented in MATLAB.
Figure 2 and 3 compare results from the different averaging methods. Similar results were obtained on two other subjects. In single frame images, we observe noise and contrast variance across the image (Fig. 2a). The average image is less noisy(Fig. 2b) and in an enlargement more details can be observed due to the decreased noise(Fig. 3b). However, some cones are dim and some regions are still blurry. For the lucky-averaging (Fig. 2c, 2d and Fig. 3c, 3d), the contrast reduction is decreased. Cones are somewhat brighter and more distinct, suggesting local contrast and resolution improvement. To quantify the difference between lucky averaging and normal averaging, a power spectral analysis was performed. We calculated the power ratio, as a function of frequency between lucky averaging and standard averaging (Fig. 4). Thus, a ratio greater than one indicates a higher spectral power for lucky averaging at a given frequency. The average increase in powers between either lucky averaging imaging metric and standard averaging is 20% and the improvement extends all the way to the optical cut-off frequency which is about 0.4 cycles/pixels based on the system resolution 2.5?m [8] for this subject’s pupil size of 6 mm. Beyond the cut off frequency, the ratio fluctuates around one, indicating that the noise level is close between the two schemes. The two metrics used for lucky averaging produced similar improvements for these images, but because the scheme can be used with most local quality evaluation metrics the approach is general and can be tuned to optimize extraction of the most appropriate information.
To investigate the role of the criterion for including pixels, we processed the same region in Fig. 3 with varied criteria to include the best 20%, 30% and 40% of pixels using GLCM method (Fig. 5). As expected, the improvement is largest with the lowest percentage criterion; however there is a limit, since as the criterion uses fewer samples, the impact of photon noise increases. Thus, the choice of criterion will be a balance between reducing random noise through averaging and improving contrast by picking the best “moments”. One simple solution to this tradeoff is to collect more data for the same region of interest. It is also possible to base the criterion on the information itself, optimizing S/N ratio.
In summary, “lucky averaging”, analogous to “lucky imaging”, has been applied to AOSLO image post processing. The image sequence is evaluated based on small regions instead of frames. The quality is enhanced by 20%, suggesting a better use of the data by this processing scheme. A larger dataset is necessary but collecting 3 or 4 seconds worth of data for each retinal region is realistic.
Acknowledgments
This work was supported by National Institutes of Health grants EY14375, EY04395 and P30EY019008
Footnotes
OCIS Codes: 100.0100, 100.2000, 100.2980
Reference
- 1.Roorda Austin. Applications of Adaptive Optics Scanning Laser Ophthalmoscopy. Optometry & Vision Science. 2010;87(4) doi: 10.1097/OPX.0b013e3181d39479. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Stevenson SB, Roorda A. Correcting for miniature eye movements in high resolution scanning laser ophthalmoscopy. In: Manns F, Söderberg P, Ho A, editors. Ophthalmic Technologies XV; Proceedings of SPIE; SPIE, Bellingham, WA. 2005.pp. 145–151. [Google Scholar]
- 3.Fried David L. Probability of getting a lucky shortexposure image through turbulence. Optical Society of America, Journal. 1978 Dec;68 [Google Scholar]
- 4.Ferguson R. Daniel, Zhong Zhangyi, Hammer Daniel X., Mujat Mircea, Patel Ankit H., Deng Cong, Zou Weiyao, Burns Stephen A. Adaptive optics scanning laser ophthalmoscope with integrated wide-field retinal imaging and tracking. JOSA A. 2010;27(11):A265–A277. doi: 10.1364/JOSAA.27.00A265. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Muller RA, Buffington A. Real-time correction of atmospherically degraded telescope images through image sharpening. J. Opt. Soc. Am. 1974;64(9):1200–1210. [Google Scholar]
- 6.Geng Ying, Anne Schery Lee, Sharma Robin, Dubra Alfredo, Ahmad Kamran, Libby Richard T., Williams David R. Optical properties of the mouse eye. biomedical optics express. 2011;2(4):717–738. doi: 10.1364/BOE.2.000717. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Haralick RM, Shanmugam K, Dinstein I. Textural Features for Image Classification. IEEE Trans. on Systems, Man, and Cybernetics. 1973;3(6):610–621. [Google Scholar]
- 8.Weiyao Zou, Xiaofeng Qi, Burns Stephen A. Woofertweeter adaptive optics scanning laser ophthalmoscopic imaging based on Lagrange-multiplier damped leastsquares algorithm. Biomedical Optics Express. 2011;2(7):1986–2004. doi: 10.1364/BOE.2.001986. [DOI] [PMC free article] [PubMed] [Google Scholar]