Abstract
In recent years, there has been a significant transformation in the field of incoherent imaging with new possibilities of compressing three-dimensional (3D) information into a two-dimensional intensity distribution without two-beam interference (TBI). Most incoherent 3D imagers without TBI are based on scattering by a random phase mask exhibiting sharp autocorrelation and low cross-correlation along the depth axis. Consequently, during reconstruction, high lateral and axial resolutions are obtained. Scattering based-Imaging requires a wasteful photon budget and is therefore precluded in many power-sensitive applications. This study develops a proof-of-concept 3D incoherent imaging method using a rotating point spread function termed 3D Incoherent Imaging with Spiral Beams (3DI2SB). The rotation speed of the point spread function (PSF) with displacement and the orbital angular momentum has been theoretically analyzed. The imaging characteristics of 3DI2SB were compared with a direct imaging system using a diffractive lens, and the proposed system exhibited a higher focal depth than the direct imaging system. Different computational reconstruction methods such as the Lucy–Richardson algorithm (LRA), non-linear reconstruction (NLR), and the Lucy–Richardson–Rosen algorithm (LRRA) were compared. While LRRA performed better than both LRA and NLR for an ideal case, NLR performed better than both under real experimental conditions. Both single plane imaging, as well as synthetic 3D imaging, were demonstrated. We believe that the proposed approach might cause a paradigm shift in the current state-of-the-art incoherent imaging, fluorescence microscopy, and astronomical imaging.
Supplementary Information
The online version contains supplementary material available at 10.1186/s11671-022-03676-6.
Keywords: Orbital angular momentum, Incoherent holography, Diffractive optics, Imaging, Microscopy
Introduction
Incoherent holography technologies have an interesting history [1]. In the beginning, incoherent holography systems had quite complicated architectures due to spatial and temporal incoherence. Common path interferometers [2], rotational shear interferometers [3], and conoscopic holography [4] were some of the previously widely used architectures. Even though incoherent holography is capable of broad applicability, the above bulky, complicated configurations hindered realizing its full potential. With the development of active devices, the implementation of incoherent holography became relatively simpler. One well-known compact incoherent holography method was the Fresnel incoherent correlation holography (FINCH) which supported two optical channels in a single physical space and independently modulated one of the channels allowing easy application to fluorescence microscopy [5]. Later, FINCH was found to have a capability to achieve super-resolution by breaking the Lagrange invariant condition, which motivated further research and development in the area of incoherent holography [6]. Today FINCH has evolved into an advanced technique investigated by many prominent researchers across the globe. [7]
While FINCH followed a natural path of evolution, retaining its roots from classical holography, another branch of today’s incoherent holography called coded aperture imaging (CAI) has been parallelly developed based on computational optical methods. The CAI methods have been indirect imaging approaches developed to perform imaging with X-rays and Gamma rays, where manufacturing lenses for these wavelengths is challenging [8, 9]. In this direction, uniformly redundant array (URA) [10], modified URA [11], spectral imaging methods [12] were developed consisting of special patterns for the aperture masks. While the above are useful techniques, a pure phase aperture has many interesting features, such as the average speckle size equal to the diffraction-limited spot size [13, 14]. By recording the speckle intensity patterns for an object and a point object followed by a cross-correlation between the two intensity distributions, the image of the object can be reconstructed closer to the resolution limit of conventional imaging.
But unlike FINCH, CAI methods were initially developed to perform 2D imaging. Another main difference between FINCH and CAI was that FINCH is intense on optical experiments and less intense in computational, but CAI was the opposite. In 2016, a holography method called coded aperture correlation holography (COACH) was developed, which blurred the above distinct differences between FINCH and CAI [15]. COACH was intense in both optics as well as computational parts which connected FINCH and CAI. COACH was able to encode not only 3D spatial but also spectral information effectively [16]. As the 3D imaging capabilities were found to be unaffected even without two-beam interference conditions, COACH dropped the two-beam interference requirement and evolved into Interferenceless COACH (I-COACH) [17]. Later, many such techniques based on I-COACH were reported, such as diffuserCam [18] and scatter plate microscope [19] with different computational reconstruction mechanisms.
Even though the scattering mask-based CAI methods can encode spatial and spectral information in a single shot, the photon budget requirement is enormous compared to conventional lens-based imagers [20]. This problem prevents the implementation of such CAI methods to power-sensitive applications such as astronomical imaging and fluorescence microscopy. A modified approach was developed to overcome this problem; Instead of scattering light uniformly, the light was collected and focused into a random array of points in the sensor plane [21]. In this way, the efficiency of imaging was improved. However, the method could not be applied directly for multiplane imaging without multiplexing masks for the required planes.
In one of the recent attempts of implementing CAI methods in the infrared beamline of the Australian Synchrotron, a better understanding of imaging concepts was obtained [22]. The requirement of a high photon budget and the beam characteristics of Cassegrain objective lenses (COL) resulted in a new imaging path. The light diffracted from COL had four sharp intensity peaks over a large depth, and so the autocorrelation function was sharp along the depth axis while the cross-correlation was lower as the spacing between the spots increased with distance. This is exactly the condition needed for performing indirect 3D imaging. This observation resulted in coining a new optical field condition for 3D imaging with CAI—sharp autocorrelation and low cross-correlation along depth (SALCAD) field. Therefore, the scattered light is random SALCAD field, while the case with COL is a deterministic SALCAD field. For imaging 3D information using the SALCAD fields generated by COL, a special reconstruction algorithm called Lucy–Richardson–Rosen algorithm (LRRA) [23] was developed by combining the non-linear reconstruction method (NLR) developed recently [24] and the well-known Lucy–Richardson algorithm (LRA) [25, 26].
The above study opened the possibility of implementing deterministic SALCAD fields with energy concentrated in a small area similar to conventional imaging but suitable for 3D imaging. In this direction, beams carrying orbital angular momentum (OAM) exhibit interesting characteristics, including SALCAD property. An interesting family of beams is the OAM beams with rotating intensity distributions along the depth axis [27–29]. There have been some reports on implementing rotating point spread functions to improve the localization and sensing along the depth axis [30–34]. Most of the studies [30–32], employed a double helix pattern to improve 3D localization. In conventional fluorescence imaging, when the fluorescent particles lie beyond the depth of field of the imaging system, they appear blurred with a larger size resulting in loss of resolution. By converting every point image into a double helix pattern, the depth of field is increased, and the location of the particle can be identified by the rotation angle of the two spots. This approach improved the 3D localization of fluorescent particles. In 2017, 3D imaging was demonstrated with a double helix pattern using the widely used deconvolution algorithms, namely Weiner filter and the Lucy–Richardson algorithm [33]. But the method was demonstrated only for a short axial distance. In this study, we have investigated the characteristics of 3D Incoherent Imaging using Spiral Beam (3DI2SB) and have demonstrated 3D imaging. Two recently developed reconstruction algorithms, namely NLR [24] and LRRA [23] have been applied to expand the limits of 3D imaging [25, 26]. The manuscript consists of five sections. In the second section, the design of the phase mask, theoretical analysis of beam propagation to understand the rate of rotation, and OAM estimation were described. In the third section, simulation results of imaging are presented. The experimental results are presented in the fourth section. The conclusion and future perspectives of this research work are discussed in the final section.
Methods
The optical configuration of 3DI2SB is shown in Fig. 1. Basically, this configuration includes the observed object, a multifunctional diffractive optical element (MDOE) and the image sensor. This section consists of three sections: design and analysis of MDOE, estimation of OAM, and imaging methodology.
Design and Analysis of MDOE
The MDOE that can generate a rotating intensity distribution has been adapted from Ref. [31]. This study uses spatially incoherent and temporally coherent illumination, so the design and calculation have been carried out for a single wavelength. The MDOE is formed by the modulo-2π phase addition of the phase distributions of a diffractive lens (DL) and a spiral element. The topological charge distribution of the spiral element follows an arithmetic progression dependent upon the zone number of annular zones with equal areas starting from the center. Let us consider the MDOE in the following form:
1 |
where is the DL in the MDOE and is the spiral element in the following form:
2 |
3 |
where r is the radial coordinate and is the azimuthal coordinate, f is the focal length of the DL given as assuming paraxial approximation [35], λ is the wavelength and n is the order of circular zones, zs and zh are the object distance and image distances, respectively. For a point object with unit amplitude located along the optical axis at z = zs + Δ from the MDOE, where Δ is the defocus distance in the object plane, the complex amplitude after the MDOE using Kirchhoff-Fresnel integral is given as
4 |
where R is the radius of the beam at the MDOE plane.
Considering the focusing part of the MDOE (1), we can rewrite expression (4) as follows:
5 |
Obviously, for z = zs we get a focused field, which is the Fourier transform of the spiral element , and at z ≠ zs we get a defocused field.
When we substitute (2)–(3) into (5) we obtain:
6 |
For each ring of the spiral element , the integration by d can be performed analytically:
7 |
If the rings are sufficiently narrow, then instead of (7) we can approximately write:
8 |
Let us rewrite expression (8) in a form convenient for analysis:
9 |
where , , . Assuming a special 2f condition, zs = zh, . Expression (9) is a superposition of Bessel beams. The behavior of such beams during propagation was studied in various works [27, 29, 32, 36], where it was shown that the rotation of the intensity of the beam (9) is determined by the interference terms and the phase difference:
10 |
Substituting the expressions for the radii of the rings , instead of (10) we obtain:
11 |
Then the rotation speed of the beam intensity depending on the distance shift from object Δ will be (omitting the smaller term):
12 |
Note that displacement Δ can be positive or negative, which means rotation in different directions. Moreover, for large values Δ, the rotation speed will differ depending on the sign due to the nonlinearity of the expression (12). At small displacements, the rotation speed will be linear:
13 |
Calculation of the Orbital Angular Momentum (OAM) of the Field Generated by the Spiral Element
The normalized OAM of an arbitrary scalar field is defined as follows [36]:
14 |
Since the OAM is an invariant characteristic of the field and is conserved during propagation and focusing, the quantity (15) can be calculated in any plane, including the initial plane (at z = 0):
15 |
Since MDOE (1) is a purely phase element, the integral in the denominator of expression (15) corresponds to the energy of a bounded plane illuminating beam . Considering that the focusing part of MDOE (1) does not change the value of OAM, the integral in the numerator of expression (15) can be calculated as follows:
16 |
Considering equality [27], we obtain the following expression:
17 |
Thus, the normalized OAM of the beams generated by the N-zone spiral element (2)–(3) is equal to the number of zones N.
Incoherent Imaging
In the previous sections, we described the design of the DOE, theoretical analysis of the special beam, and estimation of OAM. The above studies were carried out for a single point. This analysis for a single point is true for both coherent as well as incoherent imaging systems. However, when moving on from a single point to a complicated 2D object consisting of multiple points, the description of imaging systems with coherent and incoherent sources are completely different except for some specially imposed conditions of limited field of view and sparse distribution of points [37]. The current imaging configuration converts an object point into a rotating intensity distribution with two main intensity peaks that rotate around the optical axis as a double helix. When the object point’s location changes laterally, the intensity distribution shifts, and when the location changes axially, then the intensity distribution rotates either clockwise or anticlockwise depending upon the direction of the axial shift. The point spread function of the incoherent imaging system is given as . In this study, we are interested in creating an incoherent imaging system that is linear in intensity, and so for a 2D object O, the intensity distribution is given as , where ‘⊗’ is the 2D convolutional operator. The image reconstruction is , where ‘*’ is the 2D correlation operator. Rewriting the above equation of IR by substituting the composition of IO, gives; , where is the autocorrelation function that reconstructs the image function. The width of the autocorrelation function is twice that of the diffraction-limited spot size, which is 1.22λzs/D in the object plane, assuming a regular correlation, also called matched filter, is selected [38]. Alternative a correlation method with phase-only filter [38] has a relatively sharper autocorrelation function. Recently, the NLR method sharpened the autocorrelation function with a width equal to the diffraction-limited spot size [24]. A spherical lens can create a sharp PSF at the sensor if the imaging condition is satisfied, where f, z1, and z2 are the focal length, point-lens distance, and lens-sensor distance, respectively. However, when the imaging condition is violated, the PSF sharpness decreases. In the proposed 3DI2SB, the rotating PSF has sharp intensity peaks over a relatively long distance. While the method does not directly create a well-defined image on the sensor, the autocorrelation function is expected to be sharper, facilitating indirect imaging over a long distance without needing a scattering mask.
Simulation Results
The simulation studies were carried out for the following specifications of the imaging system. Number of pixels = 500 × 500, λ = 600 nm, mesh grid pixel size = 10 μm, zs = 30 cm, f = 15 cm, zh = 30 cm and D = 5 mm. The intensity distribution for z = 15 cm to 45 cm in steps of 5 cm for the current system and a reference system consisting of only a DL without a spiral element are shown in Fig. 2a–n, respectively. As seen from Fig. 2a–g, the rotation appears to be faster when the defocus distance in the object plane Δ is smaller, and with a larger increase, the rotation slows down. Secondly, the rotation is faster when the point object is closer to the MDOE. The autocorrelation function is calculated for the case with MDOE and DL, and plotted for the different object locations for MDOE (red) and diffractive lens (blue) in Fig. 2o–u, respectively. It is seen that except for the imaging condition i.e., z = 30 cm, for all other cases, the autocorrelation function obtained for the MDOE is sharper with lesser background noise than the case with a DL.
The axial behavior is studied next. In this study, the simulated intensity distribution at z = 30 cm is cross-correlated with matched filter (along the transverse coordinates) with other values from z = 15 cm to 45 cm, i.e., . The plot of for the MDOE and the intensity variation for the DL are shown in Fig. 3. The FWHM of the above profile is approximately the axial resolution of the system ~ λ/NA2, where NA is the numerical aperture. It is seen that the focal depth of the MDOE is longer than that of the DL. In other words, the axial resolution of imaging using a DL is expected to be better than the 3DI2SB method. The simulation study reveals two main characteristics of imaging: lateral resolution and axial resolution. The lateral resolution given by the width of the autocorrelation function is better for 3DI2SB method than imaging using a DL except for the case when the imaging condition is satisfied. The axial resolution of the 3DI2SB method is lower than that of imaging using a DL. The videos of the variation in the intensity distribution for the DL and MDOE are given in Additional file 1: Video S1 and Additional file 2: Videos S2, respectively. The videos of the variation of the intensity distribution of the cross-correlation matrix Ca(x,y,z) for the MDOE when the distance was varied from z = 15 cm to 45 cm are given in Additional file 3: Videos S3.
A test object with the words “Structured light” was selected for the simulation studies. In imaging with a single DL, it is expected that the image resembles the object at the image plane and is blurred anywhere else. In the case of MDOE, the image of the object will probably appear distorted for all distances. It is necessary to reconstruct the image by processing it with the PSF as described in “Incoherent Imaging” section. The processing can be achieved either using cross-correlation between two intensity matrices or by iteratively estimating the maximum likelihood solution as in LRA [25, 26]. In some of the studies involving scattering and with Cassegrain objective lenses, the NLR was found to exhibit a better quality of image reconstruction [22]. In a recent study [23], the LRRA was applied to distorted images recorded with a Cassegrain objective lens and found to perform better than both LRA and NLR. All three reconstruction methods mentioned above have not yet been applied to an exotic intensity distribution such as the rotating PSF.
The NLR is given as , where α and β are tuned between − 1 and 1, to obtain the minimum entropy given as , where (m,n) are the indexes of the correlation matrix, and C(m,n) is the cross-correlation distribution. The LRRA is shown in Fig. 4. The LRRA has three parameters namely α, β and number of iterations. When α = 1, β = 1, the algorithm is LRA.
The intensity distributions for the test object were simulated for both DL and MDOE for distances from z = 15 cm to 45 cm, as shown in Fig. 5. The reconstruction results using NLR reconstruction (α = − 0.3, β = 0.6), LRA (iterations = 50) and LRRA (α = 0, β = 0.5 to 0.7, iterations = 6) are shown in Fig. 5. In all these reconstruction methods, each transverse plane is reconstructed with its own unique PSF. The optimal reconstruction (α and β) for NLR was determined when the entropy of the reconstructed image was minimal. The optimal iteration number for LRA and iteration number, α and β values of LRRA were obtained when the difference between the reconstructed image and test object was minimal. The number of parameters to be optimized is one (number of iterations), two (α and β), and three (all the above) for LRA, NLR, and LRRA, respectively. However, once the values are optimized, for one case, the values remained almost the same when the experimental conditions were not changed. It must be noted that LRRA is obtained by applying NLR to LRA, so LRRA does not require any special workflow but just an additional parameter to optimize. The library of PSFs is computed by displaying a point object in the system input. Knowing its number in the library list for the PSF yielding the best in-focus image enables one to locate the observed object along the z-axis.
It is seen that NLR always performs better than LRA while LRRA performs better than both. One disadvantage of LRRA is that it involves a certain number of iterations even though significantly lower than LRA, which can prevent the application of LRRA to real-time imaging as NLR. Another observation is that the performance of LRA for MDOE is better than DL. The mean squared error (MSE) was calculated using the direct imaging of the test object using a DL as a reference. The MSE distribution for the different cases is shown in Fig. 6. An important observation of LRRA was that it is highly sensitive to the relative locations of the IPSF and IO. The reconstruction methods LRRA and NLR are compared next when there is a relative horizontal shift error of 5, 10, 20, and 50 pixels between IPSF and IO. The reconstruction results of NLR and LRRA are shown in Fig. 7. It is seen that when there is a shift error, the performance of LRRA is significantly affected, while in the case of NLR, there is only a shift in the location of the reconstructed information. The previous success of LRRA [23] on the data from the Australian synchrotron can be attributed to the lower number of pixels in the image sensor (64 × 64 pixels), in which the chances of error are lower. However, in most imaging and holography experiments, the scientific camera consists of at least 1 Megapixels. Consequently, LRRA may not be suitable for such cases without additional stabilization approaches to improve the resilience of LRRA in the presence of errors.
Experiments
The schematic of the experimental setup is shown in Fig. 8. The same optical setup was used for both 2D and 3D experiments. An incoherent source (Thorlabs LED625L, 12 mW, λ = 625 nm, ∆λ = 15 nm) was used for illuminating the object, elements 4, 5, and 6 (both digits and gratings) of group 5 of negative USAF target were chosen as test object for the experiments. The SLM (Holoeye PLUTO, 1920 × 1080 pixels, 8 µm pixel pitch, phase-only modulation) was used to modulate the light beam by displaying the vortex phase mask shown in Fig. 1, along with the lens function having a focal length of 14 cm. The distance between the SLM and the digital camera (Retiga R6-DCC3260M, pixel size 4.54 μm × 4.54 μm) was 14 cm. A polarizer was used to allow light only along the active axis of the SLM. A pinhole with a size of 15 µm was used to record the IPSFs. The location of the pinhole was shifted by Δ = 0, 1.4, 2.8, and 4.4 cm and the PSF library was recorded. The USAF object was then mounted at exactly the same locations as the pinhole and the object intensity distributions were recorded. The images of the recorded IPSF and IO (the system response to an object in the input) for Δ = 0, 1.4, 2.8, and 4.4 cm and the reconstruction results using NLR (α = 0, β = 0.4) followed by the application of a median filter for MDOE are shown in Fig. 9. The reference images recorded for DL Δ = 0, 1.4, 2.8, and 4.4 cm are also shown in Fig. 9. As seen, the images obtained using MDOE are sharper in comparison to the ones obtained using DL.
For 3D experiments, the fundamental properties of incoherent imaging—the linearity and space invariance in intensity were exploited. It is necessary to understand what a 3D object is. When the object points exist along the z direction, it is a 3D object. Some 3D objects have points continuously along z while others have points distributed in only a few planes. In our study, we used the 2D data recorded at different axial locations to test the 3D imaging capabilities of the method. The images of the intensity distributions from four separated planes at Δ = 0, 1.4, 2.8, and 4.4 cm are shown in the upper line of Fig. 10. The reconstructed images using the IPSF recorded at all four planes are shown in lines 2–5 of Fig. 10. It is seen that only the object information at the planes of the IPSF was in focus, while the information from other planes was blurred. The experiment described in Fig. 10 is identical to recording the total intensity distribution simultaneously from multiple planes of an object. In the current experiment, the out-of-focus and the in-focus intensities are computed in the computer instead of appearing on the sensor.
Discussion and Conclusion
In this study, a new incoherent imaging system, 3DI2SB, is investigated. The concept has been proposed as a possible replacement for the existing scattering-based 3D imaging systems [17]. The 3DI2SB approach has been studied only for a beam consisting of two intensity peaks that rotate about the optical axis. Apparently, the two peaks exist over a long focal depth enabling a higher signal-to-noise ratio than scattering-based imagers and also a DL. The lateral resolution given by the autocorrelation function was found to be sharper for 3DI2SB than the case with a DL. The focal depth of the 3DI2SB was found to be slightly higher than that of the imager using a diffractive lens. Three different reconstruction methods, namely LRA, NLR, and LRRA, were compared. LRRA and NLR were found to perform better than LRA. For an ideal case, LRRA performed better than NLR, while for a practical case, NLR performed better than LRRA. The results shown in Fig. 5 show an interesting result. When the autocorrelation function of 3DI2SB and regular imaging system with DL were compared, the former was found to be sharper than that of the case with DL. But upon applying NLR and LRRA, the correlation function became shaper for both cases. However, we expect that in practice, with various kinds of noise, the performance of 3DI2SB with any reconstruction method will be better because the intensity distribution is more concentrated on a small area along with its propagation than the case of the DL. Further studies are needed in both NLR and LRRA to improve the reconstruction results. We believe that our study will open a pathway for implementing different incoherent exotic beams and structured light for multidimensional and multispectral imaging technologies [20].
Supplementary Information
Acknowledgements
S.N. Khonina thanks the financial support from the Ministry of Science and Higher Education of the Russian Federation (state contract with the “Crystallography and Photonics” Research Center of the RAS).
Abbreviations
- 3DI2SB
Three-dimensional incoherent imaging using spiral beams
- CAI
Coded aperture imaging
- COACH
Coded aperture correlation holography
- DL
Diffractive lens
- FINCH
Fresnel incoherent correlation holography
- I-COACH
Interferenceless coded aperture correlation holography
- LED
Light emitting diode
- LRA
Lucy–Richardson algorithm
- LRRA
Lucy–Richardson–Rosen algorithm
- MDOE
Multifunctional diffractive optical element
- NLR
Non-linear reconstruction
- OAM
Orbital angular momentum
- PSF
Point spread function
- SALCAD
Sharp autocorrelation and low cross-correlation along depth
- SLM
Spatial light modulator
- TBI
Two beam interference
- URA
Uniformly redundant array
- MURA
Modified uniformly redundant array
- USAF
United States Air Force
Authors' Contributions
Imaging concepts—VA and JR, Rotating PSF theoretical calculations and OAM theory—SK; Simulation—VA and ANKR; Algorithm NLR—JR; Algorithm LRRA—VA; Validation—SK, SJ and JR; Experiments—ND and RK; Supervision—SK, JR and SJ. All authors contributed to writing the manuscript. All authors read and approved the final manuscript.
Funding
We are grateful for the financial support via ARC Linkage LP190100505 project. V. Anand acknowledges the European Union’s Horizon 2020 research and innovation programme grant agreement No. 857627 (CIPHR). A. Naresh Kumar Reddy acknowledges the support from the State Education Development Agency (SEDA), Republic of Latvia (Project Number: 1.1.1.2/VIAA/3/19/436).
Data Availability
All data generated or analysed during this study are included in this published article.
Declarations
Competing Interests
The authors declare no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Vijayakumar Anand and Svetlana Khonina have contributed equally to this work.
References
- 1.Rosen J, Vijayakumar A, Kumar M, Rai MR, Kelner R, Kashter Y, Bulbul A, Mukherjee S. Recent advances in self-interference incoherent digital holography. Adv Opt Photonics. 2019;11(1):1–66. doi: 10.1364/AOP.11.000001. [DOI] [Google Scholar]
- 2.Peters PJ. Incoherent holograms with a mercury light source. Appl Phys Lett. 1966;8(8):209–210. doi: 10.1063/1.1754558. [DOI] [Google Scholar]
- 3.Bryngdahl O, Lohmann A. Variable magnification in incoherent holography. Appl Opt. 1970;9(1):231–232. doi: 10.1364/AO.9.000231. [DOI] [PubMed] [Google Scholar]
- 4.Sirat G, Psaltis D. Conoscopic holography. Opt Lett. 1985;10(1):4–6. doi: 10.1364/OL.10.000004. [DOI] [PubMed] [Google Scholar]
- 5.Rosen J, Brooker G. Digital spatially incoherent Fresnel holography. Opt Lett. 2007;32(8):912–914. doi: 10.1364/OL.32.000912. [DOI] [PubMed] [Google Scholar]
- 6.Rosen J, Siegel N, Brooker G. Theoretical and experimental demonstration of resolution beyond the Rayleigh limit by FINCH fluorescence microscopic imaging. Opt Express. 2011;19(27):26249–26268. doi: 10.1364/OE.19.026249. [DOI] [PubMed] [Google Scholar]
- 7.Rosen J, et al. Roadmap on recent progress in finch technology. J Imaging. 2021;7(10):197. doi: 10.3390/jimaging7100197. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Ables JG. Fourier transform photography: a new method for X-ray astronomy. Publ Astron Soc Austral. 1968;1(4):172–173. doi: 10.1017/S1323358000011292. [DOI] [Google Scholar]
- 9.Dicke RH. Scatter-hole cameras for X-rays and gamma rays. Astrophys J. 1968;3:L101. doi: 10.1086/180230. [DOI] [Google Scholar]
- 10.Fenimore EE, Cannon TM. Uniformly redundant arrays: digital reconstruction methods. Appl Opt. 1981;20(10):1858–1864. doi: 10.1364/AO.20.001858. [DOI] [PubMed] [Google Scholar]
- 11.Gottesman SR, Fenimore EE. New family of binary arrays for coded aperture imaging. Appl Opt. 1989;28(20):4344–4352. doi: 10.1364/AO.28.004344. [DOI] [PubMed] [Google Scholar]
- 12.Wagadarikar A, John R, Willett R, Brady D. Single disperser design for coded aperture snapshot spectral imaging. Appl Opt. 2008;47(10):B44–51. doi: 10.1364/AO.47.000B44. [DOI] [PubMed] [Google Scholar]
- 13.Lee K, Park Y. Exploiting the speckle-correlation scattering matrix for a compact reference-free holographic image sensor. Nat Commun. 2016;7(1):1–7. doi: 10.1038/ncomms13359. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Dainty JC, editor. Laser speckle and related phenomena. Berlin: Springer; 2013. [Google Scholar]
- 15.Vijayakumar A, Kashter Y, Kelner R, Rosen J. Coded aperture correlation holography—a new type of incoherent digital holograms. Opt Express. 2016;24(11):12430–12441. doi: 10.1364/OE.24.012430. [DOI] [PubMed] [Google Scholar]
- 16.Vijayakumar A, Rosen J. Spectrum and space resolved 4D imaging by coded aperture correlation holography (COACH) with diffractive objective lens. Opt Lett. 2017;42(5):947–950. doi: 10.1364/OL.42.000947. [DOI] [PubMed] [Google Scholar]
- 17.Vijayakumar A, Rosen J. Interferenceless coded aperture correlation holography—a new technique for recording incoherent digital holograms without two-wave interference. Opt Express. 2017;25(12):13883–13896. doi: 10.1364/OE.25.013883. [DOI] [PubMed] [Google Scholar]
- 18.Antipa N, Kuo G, Heckel R, Mildenhall B, Bostan E, Ng R, Waller L. DiffuserCam: lensless single-exposure 3D imaging. Optica. 2018;5(1):1–9. doi: 10.1364/OPTICA.5.000001. [DOI] [Google Scholar]
- 19.Singh AK, Pedrini G, Takeda M, Osten W. Scatter-plate microscope for lensless microscopy with diffraction limited resolution. Sci Rep. 2017;7(1):1–8. doi: 10.1038/s41598-016-0028-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Anand V, Ng SH, Maksimovic J, Linklater D, Katkus T, Ivanova EP, Juodkazis S. Single shot multispectral multidimensional imaging using chaotic waves. Sci Rep. 2020;10(1):1–3. doi: 10.1038/s41598-019-56847-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Rai MR, Rosen J. Noise suppression by controlling the sparsity of the point spread function in interferenceless coded aperture correlation holography (I-COACH) Opt Express. 2019;27(17):24311–24323. doi: 10.1364/OE.27.024311. [DOI] [PubMed] [Google Scholar]
- 22.Anand V, Ng SH, Katkus T, Maksimovic J, Klein AR, Vongsvivut J, Bambery KR, Tobin MJ, Juodkazis S. Exploiting spatio-spectral aberrations for rapid synchrotron infrared imaging. J Synchrotron Radiat. 2021;28(5):1616–1619. doi: 10.1107/S1600577521007104. [DOI] [PubMed] [Google Scholar]
- 23.Anand V, Han M, Maksimovic J, Ng SH, Katkus T, Klein AR, Bambery KR, Tobin MJ, Vongsvivut J, Juodkazis S. Single-shot mid-infrared incoherent holography using Lucy Richardson Rosen algorithm. Opto-Electron Sci. 2022;1(3):210006. doi: 10.29026/oes.2022.210006. [DOI] [Google Scholar]
- 24.Rai MR, Vijayakumar A, Rosen J. Non-linear adaptive three-dimensional imaging with interferenceless coded aperture correlation holography (I-COACH) Opt Express. 2018;26(14):18143–18154. doi: 10.1364/OE.26.018143. [DOI] [PubMed] [Google Scholar]
- 25.Richardson WH. Bayesian-based iterative method of image restoration. JoSA. 1972;62(1):55–59. doi: 10.1364/JOSA.62.000055. [DOI] [Google Scholar]
- 26.Lucy LB. An iterative technique for the rectification of observed distributions. Astron J. 1974;79:745. doi: 10.1086/111605. [DOI] [Google Scholar]
- 27.Kotlyar VV, Khonina SN, Skidanov RV, Soifer VA. Rotation of laser beams with zero of the orbital angular momentum. Opt Commun. 2007;274(1):8–14. doi: 10.1016/j.optcom.2007.01.059. [DOI] [Google Scholar]
- 28.Schulze C, Roux FS, Dudley A, Rop R, Duparré M, Forbes A. Accelerated rotation with orbital angular momentum modes. Phys Rev A. 2015;91(4):043821. doi: 10.1103/PhysRevA.91.043821. [DOI] [Google Scholar]
- 29.Kotlyar VV, Soifer VA, Khonina SN. An algorithm for the generation of laser beams with longitudinal periodicity: rotating images. J Mod Opt. 1997;44(7):1409–1416. doi: 10.1080/09500349708230745. [DOI] [Google Scholar]
- 30.Pavani SRP, Piestun R. Three dimensional tracking of fluorescent microparticles using a photon-limited double-helix response system. Opt Express. 2008;16(26):22048–22057. doi: 10.1364/OE.16.022048. [DOI] [PubMed] [Google Scholar]
- 31.Prasad S. Rotating point spread function via pupil-phase engineering. Opt Lett. 2013;38(4):585–587. doi: 10.1364/OL.38.000585. [DOI] [PubMed] [Google Scholar]
- 32.Baránek M, Bouchal Z (2014) Optimizing the rotating point spread function by SLM aided spiral phase modulation. In: 19th Polish–Slovak–Czech optical conference on wave and quantum aspects of contemporary optics, vol 9441. International Society for Optics and Photonics, p 94410N
- 33.Wang Z, Cai Y, Liang Y, Zhou X, Yan S, Dan D, Bianco PR, Lei M, Yao B. Single shot, three-dimensional fluorescence microscopy with a spatially rotating point spread function. Biomed Opt Express. 2017;8(12):5493–5506. doi: 10.1364/BOE.8.005493. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Wang W, Situ G. Interferometric rotating point spread function. Sci Rep. 2017;7(1):1–7. doi: 10.1038/s41598-016-0028-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Vijayakumar A, Bhattacharya S. Characterization and correction of spherical aberration due to glass substrate in the design and fabrication of Fresnel zone lenses. Appl Opt. 2013;52(24):5932–5940. doi: 10.1364/AO.52.005932. [DOI] [PubMed] [Google Scholar]
- 36.Khonina SN, Kotlyar VV, Soifer VA, Pääkkönen P, Simonen J, Turunen J. An analysis of the angular momentum of a light field in terms of angular harmonics. J Mod Opt. 2001;48(10):1543–1557. doi: 10.1080/09500340108231783. [DOI] [Google Scholar]
- 37.Hai N, Rosen J. Interferenceless and motionless method for recording digital holograms of coherently illuminated 3D objects by coded aperture correlation holography system. Opt Express. 2019;27(17):24324–24339. doi: 10.1364/OE.27.024324. [DOI] [PubMed] [Google Scholar]
- 38.Horner JL, Gianino PD. Phase-only matched filtering. Appl Opt. 1984;23(6):812–816. doi: 10.1364/AO.23.000812. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All data generated or analysed during this study are included in this published article.