Abstract
Intact optical information of an object delivered through an imaging system is deteriorated by imperfect optical elements and unwanted defects. Image deconvolution has been widely exploited as a recovery technique due to its practical feasibility, and operates by assuming linear shift-invariant property of the imaging system. However, shift invariance does not rigorously hold in all imaging situations and is not a necessary condition for solving an inverse problem of light propagation. Several improved deconvolution techniques exploiting spatially variant point spread functions have been proposed in previous studies. However, the full characterization of an optical imaging system for compensating aberrations has not been considered. Here, we present a generalized method to solve the linear inverse problem of coherent light propagations without any regularization method or constraint on shift invariance by fully measuring the transmission matrix of the imaging system. Our results show that severe aberrations produced by a tilted lens or an inserted disordered layer can be corrected properly only by the proposed generalized image deconvolution. This work generalizes the theory of image deconvolution, and enables distortion-free imaging under general imaging condition.
Introduction
An optical imaging system delivers information about an object by collecting scattered light from the object and projecting it onto an imaging plane. In theory, the light scattered from each point is treated as a new point source, and forms a one-to-one correspondence with other points on the imaging plane. Ideally, the matrix of connected in-and-output spatial frequency components of an ideal imaging system can be represented as an identity matrix (Fig. 1a). In a practical imaging system, however, the image of a point object never forms a point image. This is because of the limited spatial bandwidth of the optical imaging system, which is determined by a finite numerical aperture (NA) and aberrations.
To compensate for aberrations, deconvolution (or inverse filtering), which has remarkable practical feasibility, has been widely used1–3. Deconvolution is based on two assumptions: the linearity and the shift invariant properties. Linearity in imaging implies that if the output responses to the input point sources are known, then the object image will be expressed as a linear combination of the output responses corresponding to the linear constituents of the input information. Shift invariance means that the translational shifts of a source in the object plane do not change the profile of the output response (Fig. 1b). For coherent and incoherent imaging systems, the linearity and shift invariant properties are applied to complex optical fields, and to light intensity profiles, respectively. When those assumptions hold, the output distribution of the complex amplitude (or intensity) at the imaging plane can be expressed as the input complex amplitude (or intensity) convolved with a unique point spread function (PSF) of the imaging system. In other words, it is assumed that one PSF works over the entire field of view. In the Fourier domain of a coherent imaging system, under this assumption, the optical spectrum of the complex amplitude distribution at the imaging plane is simply a multiplication of the object optical spectrum and the Fourier transform of the PSF, or a coherent transfer function (CTF).
Accordingly, the spatial frequency of an object is transferred to the imaging plane independently of other spatial frequencies, and the matrix in Fig. 1b connecting the input and output frequency components of the linear shift-invariant imaging system becomes a diagonal matrix. Deconvolution theory has been applied to a wide range of optical imaging systems, from telescope imaging4, positron-emission tomography5 to deconvolution microscopy6, confocal microscopy7, 8, and digital holography9, 10.
Although deconvolution has been used as a standard method for correcting aberrations, the shift invariance is approximately valid only when moderate aberrations exist. For instance, misalignment in an optical imaging system or contamination by dust can cause the PSF (or CTF) of an imaging system to be different at each position of the FOV, because the separated point sources in the object plane have spatially distinct optic paths. In adaptive optics11–14, one of the main issues for deep tissue imaging is to properly correct for spatially varying aberrations imposed by complex tissue structures. Recently, a wavefront shaping technique by utilizing multi-conjugate planes15, 16 or turbid layer conjugation17 has been proposed, but is practically limited to fully account the mixing of light spatial modes in an imaging plane due to a finite size of both a pixel and a total area of current spatial light modulators (SLMs). Previously, improved deconvolution techniques have been proposed in the fields of astronomy18–20 and tomography21–25, in order to correct spatially varying blurs. However, these techniques are only applicable to intensity imaging and focus primarily on finding an optimized solution of the spatially variant PSFs using the regularization method like Lucy-Richardson algorithm based on sparsely measured data, rather than restoring distortion-free images by solving the linear inverse problem solely based on fully measured spatially variant optical responses of the imaging system. The inverse problem for reversing optical aberrations has not yet been exactly solved, because this problem was found to be extremely sensitive to system noise levels26 and there have been technical limitations in accurately recording the optical responses of the system to a set of enormously many impulse signals capable of generating object space.
Even in an aligned imaging system, which maintains a high degree of shift invariance by keeping the FOV much smaller than the size of the optical elements, it is inevitable that the matrix in Fig. 1c characterizing the image delivery capability in the frequency domain has some non-diagonal components. These non-diagonal components can be an important issue, especially when a high level of aberration correction is required. Therefore, in order to completely compensate for aberrations imposed by imperfections in the imaging system, the PSF of a system which varies depending on the position of the object in the FOV, need to be systematically considered.
Here, we generalize the concept of image deconvolution by fully taking into account the varying nature of PSFs across the imaging FOV. By measuring all the responses of an imaging system to the translational shift of an input point source over the entire FOV with the help of wavefront shaping techniques27–32, the image forming capability of the imaging system is fully characterized. Inspired by recent advances in techniques33–40 for characterizing light propagations in scattering media, our approach measures the transmission matrix (TM) of a coherent imaging system in Fourier domain. In other words, we convert an ill-posed image deconvolution problem41 to a well-posed linear inverse problem of light propagations by obtaining sufficiently many independent linear equations from the measured TM of the imaging system and thus do not require any constraint on the sample and imaging system.
To investigate the formation of optical imaging – one of the fundamental principles in photonics - we revisit the inverse problem of light propagation through a coherent imaging system42. From the measured TM, we constitute a well-posed linear inverse problem and solve it without any regularization method or constraint on shift invariance. We employed an interferometric microscope equipped with a digital micro-mirror device (DMD), which enables the full-field measurements of an optical field, with the capability of controlling the wavefront of an incident beam onto a sample43 (see Methods). Using this system, the propagation of light or the TM information of the used imaging system was fully characterized by successively recording the transmitted light fields for various illumination angles of an incident beam that was controlled by a DMD. Then, to demonstrate the principle of the generalized image deconvolution, optical fields with both distorted amplitude and phase objects under various aberrations were corrected, using two approaches, the conventional and the generalized image deconvolutions, respectively. We demonstrate that the present approach effectively corrects system aberrations in nearly all imaging situations, whereas the conventional deconvolution only works when the linear shift invariance of an imaging system holds.
Principles
The full characterization of an imaging system was achieved by measuring the TM of the system. The optical responses of the imaging system were successively recorded to a set of impulses spanning the frequency space. Here, we displayed a set of input binary patterns generated by the Lee hologram method43, 44 as impulses on a DMD, and recorded the holograms of the transmitted beam with a Mach-Zehnder interferometric microscope (see Methods and Materials). However, it should be noted that the formulas that follow are applicable to any type of SLMs, including a liquid crystal on silicon and a galvanometer mirror, and any input basis that is capable of generating object space.
Suppose a total number of N input fields, spanning the spatial frequency space, are generated within a chosen NA. Let it be called a scan NA. Then, an input frequency spectrum of the i th binary image y i can be expressed as , where forms a basis of the DMD frequency space up to the NA of the condenser lens, and the coefficient d ji is directly determined by taking the Fourier transform of the i th binary image. Likewise, the output frequency spectrum of an image captured by the camera which corresponds to the i th binary input becomes . In this case, expands the spatial frequency space of the camera plane up to the NA of the imaging objective lens. Let D and C be matrices of the input and output optical spectra whose elements are and , respectively. The TM of an imaging system T connects the spatial frequency components of the input optical spectrum in the basis to ones of the output spectrum in the basis as follows: z i = Ty i. Inserting the expansions of the y i and z i in and bases into the last equation, we obtain , where t γj is the (γ, j)-entry of T for . Multiplying , the (i, δ)-entry of D −1, to both sides and summing over the index i finally gives
1 |
In Fig. 2b, the square matrix parts of the C and D matrices are represented. The calculated TM is physically the same as having the information about all the CTFs (or PSFs) at each point of the FOV. Thus, the CTF at each point in the FOV of an imaging system can be retrieved from the measured TM. We retrieved the representative CTF of the imaging system by propagating a point image source at the center of the FOV, which corresponds to the scan NA, through the TM (Fig. 2c). If we expand the CTF, O, using the basis as , the coefficient o j can be mathematically written as,
2 |
Unless otherwise stated in this paper, the representative CTF refers to the CTF that the imaging system has at the center of the FOV.
By fully characterizing an imaging system by measuring the TM, any distortion in the imaging can be corrected by using a back propagation of the complex optical fields. From the measured hologram of a sample, the complex optical fields can be retrieved using the proper field retrieval algorithms45 (Fig. 3a). Let the optical spectrum of the sample image S be expressed as . Then, in the conventional deconvolution approach, the distortion-free optical spectrum of the sample S′ can be obtained by dividing S by O from Eq. (2) as,
3 |
In contrast, in the proposed approach, the generally deconvolved optical spectrum of the sample S″ can be retrieved from S by applying the inverse of TM, T −1, from Eq. (1) as,
4 |
The remaining practical task for retrieving S″ is to obtain an inverse of C. This can be numerically achieved when C decomposes into UΣV † via the single value decomposition, where U and V are complex unitary matrices and Σ is a diagonal matrix. Accordingly, C −1 can be directly obtained by calculating V(1/Σ) U †. Taking the inverse Fourier transform of S′ and S″ then finally gives the conventionally and generally deconvolved sample images, respectively (Fig. 3b,c).
Results and Discussion
Violations in linear shift invariance of an imaging system
In order to demonstrate how imposed aberrations on the imaging system deteriorate the condition of shift invariance, two simulated point sources, whose optical spectrums are uniform up to NA of 0.6 (Fig. 4a), were propagated through the measured TMs under various optically misaligned conditions. If the imaging system has the linear shift invariant property, then the PSF and the CTF will not be changed by laterally translating an input point source. Therefore, the degree of shift invariance of the imaging system can be mapped over the entire FOV by calculating the correlations with the CTF or the profile of the PSF corresponding to the center of the FOV.
As an example, the CTFs of the lens-tilted imaging system at three different positions (α, β, and O in Fig. 4a) are shown in Fig. 4b. The two-dimensional (2-D) correlation map of the PSFs over the entire FOV of the lens-tilted system, representing the degree of shift invariance of the system, is presented in Fig. 4c. Following the same procedure, the 2-D PSF correlation maps of the aligned, optically defocused, lens tilted, and scattered systems are given in Fig. 4d–g, respectively. The output complex optical fields of the two point sources (α and β in Fig. 4a) through the measured TMs, were also presented right to the correlation map of the corresponding imaging system.
The correlation map of the aligned optics setup (Fig. 4d) shows a high degree of the shift invariant property. The output amplitude and phase profiles of the propagated two point sources are also similar to those of the input light field (Fig. 4a), except for minor distortions originating from unintentional misalignments and defects in the optical elements.
In the case of the optically defocused system (Fig. 4e), the imaging system still exhibits a high degree of shift invariance. This can be explained by the nature of light propagation. Optical defocus, or light propagation in general, only creates phase shifts on the Fourier optical spectra, so the generation of new spatial wave vectors from other spatial frequencies does not occur. Accordingly, in principle, short translations of a lens along the optic axis do not impose violations of shift invariance of the imaging system. In the propagated output field, two point foci no longer exist due to the dissatisfaction of the imaging condition; only weak interference patterns can be observed (Fig. 4e). In the phase map, however, two developed concentric phase rings are observed and this is consistent with the theory of light propagations.
In the case of tilting the tube lens (Fig. 4f), however, it severely violates the shift invariance of the imaging system. This is because optical responses upon translational shifts of an input source in the object plane are not isotropic in the imaging plane anymore. Instead, the shift invariant property of the system is only preserved along the particular direction which corresponds to the axis of the lens rotation. It is also observed that the output complex optical fields of the two point sources are squeezed along the corresponding direction.
When a disordered layer is inserted in the optic axis, which diffuses an incident plane wave into multispectral random waves, it effectively abolishes the shift invariant property of the aligned imaging system (Fig. 4g). As a result, the 2-D PSF correlation map of the scattered imaging system has a remarkably narrow range in which the shift invariance is preserved, and this narrow region at the FOV center is where the widely-known optical memory effect46–49 comes in. Also, when the point light sources are propagated through the measured TM, the output complex field is diffused out around the original point locations.
Restoration of aberration-free complex optical field images
In order to demonstrate the capability of distortion-free imaging using the generalized image deconvolution techniques, we corrected the optical field images of objects measured under optically aligned and misaligned conditions. A 1951 USAF target and a polystyrene bead with a diameter of 10 μm were used as an amplitude object and a phase object, respectively. A set of binary patterns was generated using the DMD to achieve both the maximum spatial frequency and the FOV up to 0.92 μm−1 and 20 × 20 μm2, respectively (see Methods and Materials).
The retrieved complex optical fields of the samples under various imaging conditions are presented in Fig. 5. The measured representative CTFs of the aligned (Fig. 5a), optically defocused (Fig. 5b), lens tilted (Fig. 5c), and scattered (Fig. 5d) imaging systems are given. Restored images obtained with both the conventional deconvolution and the proposed generalized deconvolution are presented. To quantitatively analyze the quality of the image reconstruction, the correlation value of each sample field with the generally deconvolved field in the aligned imaging system is indicated.
The quality of image reconstruction using the present generalized deconvolution method is significantly enhanced in comparison to the conventional devolution method. The conventionally deconvolved optical images exhibit the correct profile of the object only when the imaging system is aligned (Fig. 5a) or optically defocused (Fig. 5b). In contrast, the restored optical fields obtained with the generalized deconvolution give the correct shapes of the object even in the lens-tilted (Fig. 5c) and scattered (Fig. 5d) systems, notwithstanding some speckle noises. In the optically aligned imaging system (Fig. 5a), the profiles of both conventionally and generally deconvolved complex optical fields are similar to those of the original holograms, as expected. While conventional deconvolution preserves the non-uniform background phases of the original hologram, the generalized deconvolution further corrects the background phase profile so that it is uniform. This aspect can also be found in the optically defocused and lens-tilted imaging systems. Consequently, the restored complex optical fields based on the generalized image deconvolution exhibit a flat background phase, even in the scattered imaging system.
Next, in the optically defocused system (Fig. 5b), the corrected profiles of the samples again validate both the conventional and generalized image deconvolution approaches for correcting aberrations. The concentric phase rings of the CTFs clearly show the well-known phase profile of a spherical wavefront kernel.
In the tube-lens-tilted imaging system (Fig. 5c), however, we can see that only the generalized deconvolution properly restored the fine shapes of the number six and the spherical bead from the distorted optical fields. Despite some de-blurring effects, conventional deconvolution could not reshape the distorted images of either the USAF target or the polystyrene bead, because a tilted lens in the imaging system violates the shift invariance of the imaging system.
Imperfections in the distortion correcting capability of the conventional deconvolution seem more apparent when it comes to the scattered imaging system, which mimics an extreme situation that occurs with many microscopic scatters in the optic beam path (Fig. 5d). In this case, the application of conventional deconvolution to the raw holograms even strengthens the image distortions. The corresponding CTF of the scattered system also exhibits speckle-like inhomogeneous amplitude and phase profiles. The generalized image deconvolution, however, still restores the shape of both the amplitude and phase objects with moderate spatial noises.
Conclusions
In this paper, we introduced a generalized image deconvolution concept which does not require any assumption related to a spatially invariant system, and experimentally demonstrated its image reconstruction performance. To achieve this goal, the light propagating behaviors of the imaging system were fully characterized by successively measuring the optical responses of the system to a set of illumination beams. Our results show that the assumption of spatial invariance is satisfied only for some ideal cases. The present method correctly restored the complex amplitude images of both amplitude and the phase objects despite various types of aberrations, including a tilted lens and the insertion of a scattering layer. In contrast, the conventional deconvolution method could only correct image distortions when the distortions did not violate the shift invariance, as in the case of optical defocus.
The present work demonstrates the proof-of-principle of the suggested idea for a coherent imaging system. However, the present approach is general and also readily applicable to any imaging configuration including incoherent, x-ray, terahertz, and infrared imaging systems. For an incoherent system such as fluorescence microscopy, the intensity measurements of the points sources at various locations within the FOV can be used.
Although the present work is focused on the generalized image deconvolution, the approach presented in this work is sufficiently broad and general, and it will directly offers novel approaches for various applications where a TM of an optical system is measured. For example, the present method can be combined with various quantitative phase imaging techniques50, 51 including ptychography52, digital holographic microscopy53, 2-D quantitative phase imaging54, 55, and 3-D holotomography approaches56–58, and enhance the imaging quality significantly.
The 2-D PSF correlation maps in Fig. 4 representing the degree of shift invariance of the imaging system provide a neat compromise between the conventional and generalized image deconvolutions. The main advantage of conventional image deconvolution is that image restoration is simple and fast based on the invariant CTF (or OTF) of the imaging system. Meanwhile, generalized image deconvolution can compensate any complex aberrations by eliminating the assumption of shift invariance in the imaging system, but complicated calculations are inevitable since all changes in the CTF based on the positions within the FOV must be considered. From a practical point of view, if the FOV of an imaging system can be divided into small sections where the shift invariance holds locally, even complicated aberrations which violate the global shift invariance of the system can also be effectively corrected, by assigning the proper CTFs to the corresponding subsections. The most efficient way to compensate for aberrations would be to minimize the number of small sections that maintain the shift invariance locally and constitute the entire FOV of the imaging system.
In sum, considering that the shift invariance is always deteriorated to some extent, even in optically aligned systems, we expect that this generalized imaging deconvolution method can be effectively exploited, particularly in industrial fields and biomedical applications which require a high level of aberration corrections for the purpose of inspecting optical elements and high-performance telescopes, and calibrating microscopes.
Methods and Materials
Experimental optics setup
To demonstrate the aberration correcting capability of the generalized image deconvolution, we employed Mach-Zehnder interferometry (Fig. 6a). A diode-pumped solid-state laser with a single longitudinal mode (λ = 532 nm, 50 mW, Cobolt Co., Solna, Sweden) served as the light source. A 2 × 2 fiber optic coupler (λ = 532 ± 15 nm, 90:10 split, Thorlabs, USA) connected to a fiber collimator (f = 4.34 mm, NA = 0.57, Thorlabs, USA) divides the incident laser beam into a sample and a reference arm. The illumination angles of the laser beams impinging onto the sample are systematically controlled by using a DMD (V-9501, Texas Instruments Inc., USA), which is located in the conjugate plane of the sample. When the Lee hologram, corresponding to the normal illumination, is loaded onto the DMD, the 1st order diffraction beam was chosen to serve as an optic beam path in the imaging system. Two objective lenses (UPlanSApo, 60×, water immersion, NA = 1.2, infinity corrected, Olympus Inc., USA) were used as a condenser and an objective lens, respectively. However, when imaging a thick object, an objective lens with a long working distance (LUCPlanFL N, 60×, WD = 1.7 mm, NA = 0.7, Olympus Inc., USA) was used as a condenser lens. Total magnification of the optical system is 100×, contributed by a telescopic 4-f system which consists of a 60× objective lens and an aspheric lens with a focal length of 300 mm. The sample and the reference arm were then recombined by a 50:50 beam splitter in front of the high-speed sCMOS camera (528 × 512 pixels with a pixel size of 6.5 μm, Neo sCMOS, ANDOR Inc., Northern Ireland, U.K.), and two polarizers were used to adjust the intensities of the sample and the reference beam intensities to maximize the fringe visibility.
Various aberrations imposed on the optical imaging system
To impose various aberrations on the Mach-Zehnder setup, the optical components were intentionally misaligned via image defocus, lens tilt, and insertion of weakly scattering media in the optic axis, as shown in Fig. 6b. Defocus was achieved by slightly moving the objective lens from a focused position. To apply more severe aberrations, the aspheric tube lens of the 4-f imaging system in front of the beam splitter was tilted by 10 degrees. To generate an extreme situation, we inserted two pieces of lens papers between the imaging and pupil planes.
Transmission matrix recording
To record responses of the optical imaging system for various light illumination angles, i.e., measuring the TM, we constructed a signal triggering circuit which connects both the DMD and the sCMOS camera. The DMD controls the angle of light illumination and then a trigger signal is sent to the camera, so that the corresponding interferogram is captured before the next DMD pattern is projected.
The control of illumination angles was achieved by projecting Lee holograms onto the DMD44. A ideal DMD pattern generating a plane wave with spatial frequencies of k x and k y, f (x, y), can be written as,
5 |
where x and y are radial coordinates of the DMD, p is the DMD pixel size, and M denotes a magnification of the imaging system. The binary hologram pattern to be uploaded on the DMD, h (x, y), is then determined as h (x, y) = 1, where f (x, y) > 0.5 and h (x, y) = 0, otherwise. Detailed information can be found elsewhere43. To measure the TM, the angular frequency space of the imaging system was sequentially scanned. The total number of DMD patterns is determined by both the scan NA and the step size in angular frequency space. Because inverse filtering is a band-limited technique, the scan NA is fundamentally limited by the value of min (NAc, NAi), where NAc and NAi denote the NAs of the condenser and imaging objective lenses, respectively. Meanwhile, the step size determines the maximum FOV of objective images to be corrected without aliasing artifact. The speed of the TM recording process is determined by the frame rate of the camera.
In the experiments with a 10-μm-diameter polystyrene bead (72986, Sigma-Aldrich Co., USA) and a 2″ × 2″ positive 1951 USAF target (#57-896, Edmund Optics Inc., USA), we chose the scan NA = 0.6 and FOV = 20 × 20 μm2. Accordingly, the scanning step size was set to be 1/20 μm−1 and the total step numbers were then 1,517. The frame rate of the camera was 67 Hz. The entire TM measurement takes less than 30 sec.
During the TM measurement, there could exist globally varying phase noises in the measured optical fields, which mainly attribute to air fluctuation and mechanical vibration affecting interferometry. To correct the globally varing phase noises, an additional invariant reference pattern with the area of 0.54 × 0.54 mm2 was simultaneously displayed on the corner of the DMD. Since the global phase varitations of this reference pattern were experimentally measured and these varitations are independent to Lee hologram patterns, the effects of globally varing phase noises can be cancled out without loss of generality.
Acknowledgements
This work was supported by KAIST, BK21+ program, Tomocube, and the National Research Foundation of Korea (2015R1A3A2066550, 2014M3C1A3052567, 2014K1A3A1A09063027).
Author Contributions
Y.P. conceived and supervised the project. S.L. performed the experiments and analyzed the data. K.L. and S.S. provided analysis tools. All authors wrote the manuscript.
Competing Interests
Prof. Park, Mr. KyeoReh Lee and Mr. Seungwoo Shin have financial interests in Tomocube Inc., a company that commercializes optical diffraction tomography and quantitative phase imaging instruments and is one of the sponsors of the work.
Footnotes
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Frieden, B. R. In Picture Processing and Digital Filtering 177–248 (Springer, 1975).
- 2.Stroke G, Zech R. A posteriori image-correcting “deconvolution” by holographic fourier-transform division. Physics Letters A. 1967;25:89–90. doi: 10.1016/0375-9601(67)90358-1. [DOI] [Google Scholar]
- 3.Castleman, K. Digital image processing. (1993).
- 4.Starck J, Pantin E, Murtagh F. Deconvolution in astronomy: A review. Publications of the Astronomical Society of the Pacific. 2002;114:1051. doi: 10.1086/342606. [DOI] [Google Scholar]
- 5.Colsher JG. Fully-three-dimensional positron emission tomography. Physics in medicine and biology. 1980;25:103. doi: 10.1088/0031-9155/25/1/010. [DOI] [PubMed] [Google Scholar]
- 6.McNally JG, Karpova T, Cooper J, Conchello JA. Three-dimensional imaging by deconvolution microscopy. Methods. 1999;19:373–385. doi: 10.1006/meth.1999.0873. [DOI] [PubMed] [Google Scholar]
- 7.Shaw PJ, Rawlins DJ. The point‐spread function of a confocal microscope: its measurement and use in deconvolution of 3‐D data. Journal of Microscopy. 1991;163:151–165. doi: 10.1111/j.1365-2818.1991.tb03168.x. [DOI] [Google Scholar]
- 8.Sarder P, Nehorai A. Deconvolution methods for 3-D fluorescence microscopy images. IEEE Signal Processing Magazine. 2006;23:32–45. doi: 10.1109/MSP.2006.1628876. [DOI] [Google Scholar]
- 9.Cotte Y, Toy MF, Pavillon N, Depeursinge C. Microscopy image resolution improvement by deconvolution of complex fields. Optics Express. 2010;18:19462–19478. doi: 10.1364/OE.18.019462. [DOI] [PubMed] [Google Scholar]
- 10.Cotte Y, et al. Marker-free phase nanoscopy. Nature Photonics. 2013;7:113–117. doi: 10.1038/nphoton.2012.329. [DOI] [Google Scholar]
- 11.Booth MJ. Adaptive optics in microscopy. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences. 2007;365:2829–2843. doi: 10.1098/rsta.2007.0013. [DOI] [PubMed] [Google Scholar]
- 12.Kubby, J. A. Adaptive Optics for Biological Imaging (CRC press, 2013).
- 13.Girkin JM, Poland S, Wright AJ. Adaptive optics for deeper imaging of biological samples. Current opinion in biotechnology. 2009;20:106–110. doi: 10.1016/j.copbio.2009.02.009. [DOI] [PubMed] [Google Scholar]
- 14.Booth MJ. Adaptive optical microscopy: the ongoing quest for a perfect image. Light: Science & Applications. 2014;3:e165. doi: 10.1038/lsa.2014.46. [DOI] [Google Scholar]
- 15.Simmonds RD, Booth MJ. Modelling of multi-conjugate adaptive optics for spatially variant aberrations in microscopy. Journal of Optics. 2013;15:094010. doi: 10.1088/2040-8978/15/9/094010. [DOI] [Google Scholar]
- 16.Wu T-W, Cui M. Numerical study of multi-conjugate large area wavefront correction for deep tissue microscopy. Optics express. 2015;23:7463–7470. doi: 10.1364/OE.23.007463. [DOI] [PubMed] [Google Scholar]
- 17.Park J-H, Sun W, Cui M. High-resolution in vivo imaging of mouse brain through the intact skull. Proceedings of the National Academy of Sciences. 2015;112:9236–9241. doi: 10.1073/pnas.1505939112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Lauer, T. R. Deconvolution with a spatially-variant PSF. arXiv preprint astro-ph/0208247 (2002).
- 19.Nagy JG, O’Leary DP. Restoring images degraded by spatially variant blur. SIAM Journal on Scientific Computing. 1998;19:1063–1082. doi: 10.1137/S106482759528507X. [DOI] [Google Scholar]
- 20.Bardsley J, Jefferies S, Nagy J, Plemmons R. A computational method for the restoration of images with an unknown, spatially-varying blur. Optics express. 2006;14:1767–1782. doi: 10.1364/OE.14.001767. [DOI] [PubMed] [Google Scholar]
- 21.Formiconi A, Pupi A, Passeri A. Compensation of spatial system response in SPECT with conjugate gradient reconstruction technique. Physics in medicine and biology. 1989;34:69. doi: 10.1088/0031-9155/34/1/007. [DOI] [PubMed] [Google Scholar]
- 22.Rathee S, Koles ZJ, Overton TR. Image restoration in computed tomography: Estimation of the spatially variant point spread function. IEEE transactions on medical imaging. 1992;11:539–545. doi: 10.1109/42.192689. [DOI] [PubMed] [Google Scholar]
- 23.Temerinac-Ott, M., Ronneberger, O., Nitschke, R., Driever, W. & Burkhardt, H. In Biomedical Imaging: From Nano to Macro, 2011 IEEE International Symposium on. 899–904 (2011).
- 24.Munger P, Crelier GR, Peters TM, Pike GB. An inverse problem approach to the correction of distortion in EPI images. IEEE Transactions on Medical Imaging. 2000;19:681–689. doi: 10.1109/42.875186. [DOI] [PubMed] [Google Scholar]
- 25.Barbee DL, Flynn RT, Holden JE, Nickles RJ, Jeraj R. A method for partial volume correction of PET-imaged tumor heterogeneity using expectation maximization with a spatially varying point spread function. Physics in medicine and biology. 2009;55:221. doi: 10.1088/0031-9155/55/1/013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Den Dekker A, V den Bos A. Resolution: a survey. JOSA A. 1997;14:547–557. doi: 10.1364/JOSAA.14.000547. [DOI] [Google Scholar]
- 27.Vellekoop IM, Mosk A. Focusing coherent light through opaque strongly scattering media. Optics letters. 2007;32:2309–2311. doi: 10.1364/OL.32.002309. [DOI] [PubMed] [Google Scholar]
- 28.Park J-H, et al. Subwavelength light focusing using random nanoparticles. Nature photonics. 2013;7:454–458. doi: 10.1038/nphoton.2013.95. [DOI] [Google Scholar]
- 29.Yu H, et al. Recent advances in wavefront shaping techniques for biomedical applications. Current Applied Physics. 2015;15:632–641. doi: 10.1016/j.cap.2015.02.015. [DOI] [Google Scholar]
- 30.Yoon J, Lee K, Park J, Park Y. Measuring optical transmission matrices by wavefront shaping. Optics Express. 2015;23:10158–10167. doi: 10.1364/OE.23.010158. [DOI] [PubMed] [Google Scholar]
- 31.Rotter, S. & Gigan, S. Light fields in complex media: mesoscopic scattering meets wave control. arXiv preprint arXiv:1702.05395 (2017).
- 32.Yoon, J. et al. Optogenetic control of cell signaling pathway through scattering skull using wavefront shaping. Scientific reports5 (2015). [DOI] [PMC free article] [PubMed]
- 33.Popoff S, et al. Measuring the transmission matrix in optics: an approach to the study and control of light propagation in disordered media. Physical review letters. 2010;104:100601. doi: 10.1103/PhysRevLett.104.100601. [DOI] [PubMed] [Google Scholar]
- 34.Yu H, et al. Measuring Large Optical Transmission Matrices of Disordered Media. Physical review letters. 2013;111:153902. doi: 10.1103/PhysRevLett.111.153902. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Popoff S, Lerosey G, Fink M, Boccara AC, Gigan S. Controlling light through optical disordered media: transmission matrix approach. New Journal of Physics. 2011;13:123021. doi: 10.1088/1367-2630/13/12/123021. [DOI] [Google Scholar]
- 36.Choi Y, et al. Overcoming the diffraction limit using multiple light scattering in a highly disordered medium. Physical review letters. 2011;107:023902. doi: 10.1103/PhysRevLett.107.023902. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Mosk AP, Lagendijk A, Lerosey G, Fink M. Controlling waves in space and time for imaging and focusing in complex media. Nature photonics. 2012;6:283–292. doi: 10.1038/nphoton.2012.88. [DOI] [Google Scholar]
- 38.Tao X, Bodington D, Reinig M, Kubby J. High-speed scanning interferometric focusing by fast measurement of binary transmission matrix for channel demixing. Optics express. 2015;23:14168–14187. doi: 10.1364/OE.23.014168. [DOI] [PubMed] [Google Scholar]
- 39.Boniface A, Mounaix M, Blochet B, Piestun R, Gigan S. Transmission-matrix-based point-spread-function engineering through a complex medium. Optica. 2017;4:54–59. doi: 10.1364/OPTICA.4.000054. [DOI] [Google Scholar]
- 40.Sarma R, Yamilov A, Cao H. Enhancing light transmission through a disordered waveguide with inhomogeneous scattering and loss. Applied Physics Letters. 2017;110:021103. doi: 10.1063/1.4973459. [DOI] [Google Scholar]
- 41.Bertero, M. & Boccacci, P. In From Cells to Proteins: Imaging Nature across Dimensions 349–370 (Springer, 2005).
- 42.Kim T, Zhou R, Goddard LL, Popescu G. Solving inverse scattering problems in biological samples by quantitative phase imaging. Laser & Photonics Reviews. 2016;10:13–39. doi: 10.1002/lpor.201400467. [DOI] [Google Scholar]
- 43.Shin S, Kim K, Yoon J, Park Y. Active illumination using a digital micromirror device for quantitative phase imaging. Optics Letters. 2015;40:5407–5410. doi: 10.1364/OL.40.005407. [DOI] [PubMed] [Google Scholar]
- 44.Lee W-H. Binary computer-generated holograms. Applied Optics. 1979;18:3661–3669. doi: 10.1364/AO.18.003661. [DOI] [PubMed] [Google Scholar]
- 45.Debnath SK, Park Y. Real-time quantitative phase imaging with a spatial phase-shifting algorithm. Optics Letters. 2011;36:4677–4679. doi: 10.1364/OL.36.004677. [DOI] [PubMed] [Google Scholar]
- 46.Feng S, Kane C, Lee PA, Stone AD. Correlations and fluctuations of coherent wave transmission through disordered media. Physical review letters. 1988;61:834. doi: 10.1103/PhysRevLett.61.834. [DOI] [PubMed] [Google Scholar]
- 47.Bertolotti J, et al. Non-invasive imaging through opaque scattering layers. Nature. 2012;491:232–234. doi: 10.1038/nature11578. [DOI] [PubMed] [Google Scholar]
- 48.Katz O, Small E, Silberberg Y. Looking around corners and through thin turbid layers in real time with scattered incoherent light. Nature photonics. 2012;6:549–553. doi: 10.1038/nphoton.2012.150. [DOI] [Google Scholar]
- 49.Katz O, Heidmann P, Fink M, Gigan S. Non-invasive single-shot imaging through scattering layers and around corners via speckle correlations. Nature photonics. 2014;8:784–790. doi: 10.1038/nphoton.2014.189. [DOI] [Google Scholar]
- 50.Popescu, G. Quantitative phase imaging of cells and tissues (McGraw-Hill, 2011).
- 51.Lee K, et al. Quantitative phase imaging techniques for the study of cell pathophysiology: from principles to applications. Sensors. 2013;13:4170–4191. doi: 10.3390/s130404170. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Zheng G, Horstmeyer R, Yang C. Wide-field, high-resolution Fourier ptychographic microscopy. Nature photonics. 2013;7:739–745. doi: 10.1038/nphoton.2013.187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Kemper B, Langehanenberg P, Von Bally G. Digital holographic microscopy. Optik & Photonik. 2007;2:41–44. doi: 10.1002/opph.201190249. [DOI] [Google Scholar]
- 54.Wang Z, et al. Spatial light interference microscopy (SLIM) Optics express. 2011;19:1016–1026. doi: 10.1364/OE.19.001016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Park Y, Popescu G, Badizadegan K, Dasari RR, Feld MS. Diffraction phase and fluorescence microscopy. Optics express. 2006;14:8263–8268. doi: 10.1364/OE.14.008263. [DOI] [PubMed] [Google Scholar]
- 56.Kim K, Kim KS, Park H, Ye JC, Park Y. Real-time visualization of 3-D dynamic microscopic objects using optical diffraction tomography. Opt Express. 2013;21:32269–32278. doi: 10.1364/OE.21.032269. [DOI] [PubMed] [Google Scholar]
- 57.Kim K, et al. High-resolution three-dimensional imaging of red blood cells parasitized by Plasmodium falciparum and in situ hemozoin crystals using optical diffraction tomography. Journal of biomedical optics. 2014;19:011005–011005. doi: 10.1117/1.JBO.19.1.011005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Kim K, et al. Optical diffraction tomography techniques for the study of cell pathophysiology. Journal of Biomedical Photonics & Engineering. 2016;2:020201. [Google Scholar]