Abstract
To address the major challenges to obtain high spatial resolution in snapshot hyperspectral imaging, 3D printed glass lightguide array has been developed to sample the intermediate image in high spatial resolution and redistribute the pixels in the output end to achieve high spectral resolution. Curved 3D printed lightguide array can significantly simplify the snapshot hyperspectral imaging system, achieve better imaging performance, and reduce the system complexity and cost. We have developed two-photon polymerization process to print glass lightguide array, and demonstrated the system performance with biological samples. This new snapshot technology will catalyze new hyperspectral imaging system development and open doors for new applications from UV to IR.
Graphical Abstract
3D printed glass lightguide array has been developed to sample the intermediate image in high spatial resolution and redistribute the pixels in the output end to achieve high spectral resolution. Curved 3D printed lightguide array can significantly simplify the snapshot hyperspectral imaging system, achieve better imaging performance, and reduce the system complexity and cost.
Introduction
Snapshot hyperspectral imaging is an emerging imaging modality to obtain spectral information of the dynamic target in real time. Several snapshot hyperspectral imaging technologies have been developed for various applications ranging from astronomy to agriculture and from surveillance to medicine.1–4 These technologies include integral field imaging, image mapping, multi-aperture filtering, spectrally resolving detector array, coded aperture, computed tomography imaging, tunable echelle imager, and Fourier transform spectral imaging.1 Compared to other methods, integral field imaging method has a number of key advantages, such as easy to implement with regular camera, less computation, and low cost.
The key in integral field imaging technology is how to sample the intermediate field. Figure 1(a) is the schematic configuration of a typical snapshot hyperspectral imaging system. The imaging optics images the object to the input end of the field sampler which redistributes the image points to different locations in the output end. The collimation optics then collimates the light from the re-sampled image points. After passing through the dispersion element, the dispersed collimated light is focused onto the digital sensor. The hyperspectral images can then be reconstructed from the dispersed image. Pinhole array, faceted mirror, fiber bundle, and lenslet array have been used to sample the field, each of them has its unique properties and limitations.5–10 Pinhole array method is simple and low cost to implement, but light efficiency and spatial resolution are two major limitations, preventing its adoption in practical applications. Lenslet array approach partially addresses the light efficiency issue, but the spatial resolution is still the bottleneck. Faceted mirror method is relatively difficult to implement as it is difficult to fabricate, and the edges of the faceted mirrors can cause some artifacts. Fiber bundle approach is very straightforward, with one end of the fiber array arranged as compact as possible to sample the intermediate image and the other end of the fiber array specially arranged so that the spectral information can be distinguished in the detection sensor with a dispersion element, such as prism or grating. The resolution is determined by the pitch of the fiber bundle in the input end. Due to the cladding layer, the spatial resolution is still sacrificed, and the light efficiency is not high either.
Three-dimensional (3D) printing is an emerging fabrication method for precision optics, it is attractive due to its flexibility in building complex shapes through an additive process.11–22 Most of the research in printing optics to date has focused on organic polymer or resin-based systems using stereolithography (STL),13 direct ink writing (DIW),14 projection microstereolithography (PμSL),15,16 and two-photon polymerization (TPP) processes.19–22 Optical element printed with organic polymer or resin through UV curing process has a number of limitations in hardness, transparency in UV and NIR, thermal resistance, chemical resistance and tenability. Another issue is that the printed part becomes yellowish gradually, the transmission in short wavelength is further degraded. Compared to optics made from optical polymers and optical silicones, inorganic glass optics is preferred for many applications because of its excellent optical, chemical, and thermal properties. For the applications in UV and IR, glass optics is the only solution.
We have developed a compact, high-resolution snapshot hyperspectral imaging (HSHI) system with 3D printed glass lightguide array as shown in Figure 1(a). The imaging optics images the object onto the intermediate image plane, and the input end of the 3D printed lightguide array samples the intermediate image in high resolution since each lightguide can be printed as small as the pixel size of a typical digital sensor. The output end of the lightguide array is specially arranged so that the light spectrum from each lightguide, after passing through the collimation optics, dispersion element, and focusing optics, can be disentangled in the sensor with minimum crosstalk. In addition to the high spatial resolution, the proposed HSHI system can be very compact since the input and output surfaces of the lightguide array can be printed in freeform shape, including concave surfaces. With concave surfaces, the optical systems can be simplified dramatically as there are no needs to correct field curvature, which is one of the driving factors for complex optical system.23–26
In this paper, we will first introduce the concept of compact, high-resolution snapshot hyperspectral imaging with 3D printed glass lightguide array, and then demonstrate the system performance with mask pattern and biological slides. We will also discuss the future research and potential applications.
Results
Concept of lightguide array for HSHI system
The goal of snapshot hyperspectral imaging is to obtain spectral information of an object with a single capture. To obtain high spatial and spectral resolution, due to the two-dimensional (2D) structure of the digital sensor it is necessary to re-arrange the images of object points in the image plane so that their positions are sparsely located in the digital sensor. The pixels between adjacent lightguides are reserved for capturing spectral information.
Inspired by the unique features of vertebrate and arthropod eyes, we have developed lightguide array to address the major hurdle in snapshot hyperspectral imaging. Figure 1(b) is the schematic longitudinal section of an anterior median eyes of the funnel-web spider Agelena labyrinthica.27 Below the cuticular lens lie the vitreous cells; they are enclosed by a cup of pigment cells. Similar to the human eye, the retina is curved to cover a larger field of view (FOV) and obtain better image quality. The retina consists of slim photoreceptor cells under a thin layer of pigment-free pigment cell processors. The light absorbing rhabdomeres extend on the distal sides of each visual cell. The axons of adjacent sensory cells meet near the middle of the eye cup. Figure 1(c) is the schematic structure of an apposition eye.28 In most apposition eyes the eight receptor cells in each ommatidium contribute to a single radial structure, known as a rhabdom. Having a slightly higher refractive index than its surroundings, the rhabdom behaves as a lightguide, so that the light entering its distal tip travels down the structure, trapped by total internal reflection (TIR). Inspired by the unique features of the eyes in Figures 1(b) and 1(c), we have developed various lightguide arrays for hyperspectral imaging. Figure 1(d) is the schematic configuration of a lightguide array with lenses and dispersion prism in the output end. An imaging optical system is needed to image the object to the flat input surface of this lightguide array. The dispersed light from the output end will be detected by the sensor. Figure 1(e) is the schematic configuration of another lightguide array with imaging lens in each lightguide in the input end, lenses and dispersion prism in the output end to obtain the spectral information. This apposition eye configuration will enable the lightguide array to capture information from large FOV and obtain hyperspectral information from the flat sensor in the output end.
Figures 1(f)–1(h) are the schematic structures of the lightguide arrays with a small number of lightguides to transform the spatial distribution of the intermediate image formed by the optical imaging system to the optimal distribution for obtaining spectral information with minimum crosstalk. The lightguide array in Figure 1(f) has a concave input surface and convex output surface, mimicking the eye structure in Figure 1(b). The concave input surface can simplify the imaging optics as there is no need to correct field curvature. The convex output surface can display the information at large angles. The lightguide array in Figure 1(g) has concave surfaces on both ends, significantly simplifying both the imaging optics in front of the lightguide array and collimation optics after the lightguide array. The lightguide array in Figure 1(h) has flat surfaces on both ends, it is much simpler, suitable for the imaging optics and collimation optics designed for the flat field.
The function of the lightguide array is to transform the spatial distribution of pixels from continuous mode in the intermediate image plane to sparse mode in the output end as shown in Figure 1(a). The lightguides in the input end are arranged in grid format, like the pixel layout in the monochromatic digital sensor. This allows the intermediate image to be sampled with high spatial resolution, and the hyperspectral image can be simply reconstructed in grid format from the dispersed image. Figures 2(a) and 2(b) are two layouts of the input ends of the lightguide arrays with square and circular lightguides. Square-shape lightguide has a better light efficiency and less light leakage. In contrast, the gaps between circular lightguides will leak light, reducing image contrast and introducing errors in reconstructing hyperspectral image. The lightguide shape can be maintained from the input end to the output end, the lightguide arrays in Figures 1(g)–1(h) and 2(c) are some examples. The shape can also change gradually from one shape to another, for example the lightguide in Figure 2(d) transforms the shape from square in the input end to circular in the output end. The arrangement of the lightguide array in the output surface is optimized to efficiently utilize the pixels in the detector based on the required spectral resolution.
Since the output surface of the lightguide is typically much larger than the input surface, all but the central lightguide are curved to some degrees, depending on the locations in the input and output ends as shown in Figure 2(e). One design guideline is to maintain the TIR condition along the entire lightguide. For the lightguide arrays with a large number of lightguides, the height H in Figure 2(e) should be large enough to ensure that the local radius R of the curve is large enough to avoid light leakage for the lightguide with large W.29 Another design consideration is that the lightguide in both input and output ends should be normal to the input and output surfaces, regardless of being flat or curve, as shown in Figure 2(e). This ensures high light efficiency in coupling light into the lightguide array from the imaging system and coupling light out of the lightguide array for the collimation optics. The diameter of the individual lightguide should be small enough, and the fill-factor in the input end of the lightguide array should be high enough to achieve high spatial resolution imaging. Printed lightguide will be an ideal solution as there is no need to have the cladding layer. More discussions on the design of lightguides are provided in the Supplementary Materials.
Printed lightguide array
3D printing is attractive due to its flexibility in building complex shapes through an additive process.30 Most reported studies use polymeric materials to print organic optics with decent performance. However, organic optics printed by polymer-based components are limited in practical applications due to their poor thermal stability, low transmission in short wavelengths, and low refractive indices. Glass optics is preferred as it has better optical, chemical, thermal, and mechanical properties. In the past few years, we have developed an TPP printing process with a solvent-free, pre-condensed liquid silica resin (LSR) for fabricating micro glass optics with relatively complex structures.19,20 Transparent glass optics can be obtained after thermal treatment at 600 °C in the air with linear shrinkage as low as 17%. The peak-to-valley deviation of the printed surface is better than 100 nm and the surface roughness is better than 6 nm.19 The solvent-free LSR is synthesized based on acid-catalyzed polymerization of tetramethoxysilane (TMOS) together with a sub-stoichiometric amount of water (water solution) and 6.5 mol% of methacryloxymethyltrimethoxysilane (MMTS) as photocurable moiety. More discussions on material preparation are provided in the Supplementary Materials. Figure 3 shows the TPP printing process with LSR, Figure 3(a) is the schematic diagram of focusing the laser light to a small spot to cure LRS locally in the focal point and Figure 3(b) illustrates the pyrolysis process of the cured structure. Figure 3(c) is the SEM image of the printed micro-spectrometer consisting of a prism and a collimation lens with grating on the flat surface, and Figure 3(d) is the SEM image of a printed micro-objective.20
To demonstrate the capability of the printing system and materials, we have printed lightguide arrays with circular and square lightguides, different input and output surface shapes. Figure 4(a) is the SEM image of the printed lightguide array with convex input and output surfaces, Figure 4(b) is the SEM image of the 1/4 cross-section of the printed lightguide array with concave input and output surfaces, and Figure 4(c) is the SEM image of the lightguide array with flat input and output surfaces. Figures 4(d)–4(f) show the details of lightguide arrays around the input end, output end, and middle region, demonstrating smoothness of the lightguide TIR surface. All lightguide arrays in Figure 4 have the supporting structures to support the input and output surfaces. The surface quality of the printed lightguide is discussed in the Supplementary Materials
Due to the limitations of the home-built 3D TPP printer without high precision motion stages, it is challenging to print the lightguide array with large height H, limiting its capability in printing square shape lightguides without breaking TIR condition because it is necessary to bend in two-dimension for the lightguides not in x-z and y-z planes. In this paper, the lightguide arrays with circular lightguides were printed to demonstrate the HSHI concept and evaluate system performance. With long-range 3-axis high precision motion stages, it is feasible to print the proposed lightguide array having more than 640×480 lightguides in either square or circular shape.
Characterization of the lightguide array
To demonstrate the concept with commercially available microscope objective and other imaging optics designed for flat object and image planes, the lightguide array with circular lightguides in Figure 1(h) was chosen for evaluation. Due to the limited printing volume of the home-built 3D TPP printer, a lightguide array with 15×9 lightguides was printed. The diameter of each lightguide was 5 μm, and the distance between the input and output surfaces was 100 μm. To measure the transmission of the lightguide, the light was first focused to a small spot and then coupled into the lightguide as shown in Figure 5(a). The ratio of the light received by the sensor with and without the lightguide is the transmission of the light guide. The light loss includes the light loss from Fresnel reflection on both surfaces, light leakage due to printing defects which break TIR condition, and material absorption. By changing the light spectrum, we measured the transmission of the lightguide at different wavelengths. Figure 5(b) plots the averaged transmission of the 9 lightguides.
To evaluate the performance of the 3D printed glass lightguide array in snapshot hyperspectral imaging, we built the system in Figure 6. A broadband light source (NKT Photonics FIU-15) provided illumination on the object with the help of the illumination optics (L1 and L2). The object, a Bayer filter segment, was relayed onto the input surface of the lightguide array using two microscope objectives (OBJ1: OLYMPUS UPlanFLN NA0.13 and OBJ2: NeoDPlan NA0.4). A third microscope objective (OBJ3: OLYMPUS LUCPlanFLN NA0.60) imaged the intermediate image from the output surface of the lightguide array to infinity. The collimated light was dispersed by the custom dispersion prism and then focused onto the digital sensor (IDSUI-3080SE-C) with the f=200 mm imaging lens (L3). With the above components used in the system, we can achieve a spatial resolution of 1.25μm. The spatial resolution can be improved if different objectives are used. The hyperspectral images were decoded from the dispersed image directly, and color image was reconstructed as superposition of the hyperspectral images according to the quantum efficiency of the color CMOS sensor. With the current custom dispersion prism and lens parameters, the spectral resolution is 7nm. The spectral resolution can also be improved with a prism with stronger dispersion.
The HSHI system integrates the spectral dimension with the spatial coordinates according to the lightguide structure. It encodes a 3D hypercube into a 2D dispersed image as
(1) |
where f denotes the encoding mapping from the spatial-spectral coordinates (x, y, λ) to the pixel coordinates (x′, y′) on the sensor. To retrieve the hypercube, a decoding operation is necessary. Typically, the decoding operation consists of spatial and spectral calibrations. Spatial calibration decides the mapping between lightguide index (x, y) and pixel coordinates on the dispersed image (x′, y′).
Spatial calibration
Spatial calibration intends to determine the pixel coordinates on the dispersed image for every lightguide in the input end. Since the relative position of the lightguide array in the input and output ends was known as shown in Figures 7(a) and 7(b), the even columns in the output end were shifted half period to keep interleaved with the odd columns. This interleaved pattern allowed sufficient pixels to cover the dispersed light from each lightguide and made full use of the pixels on the sensor without spectral overlap as shown in Figure 7(c) captured under visible illumination without the sample in the object plane (this image was captured with color CMOS sensor for illustration only). The circles in Figures 7(a) and 7(b) indicated the input end and output end of the same lightguide, the rectangular box in Figure 7(c) was the dispersed spectrum of the light from the corresponding lightguide in Figure 7(b). The pixel coordinate for each lightguide was recorded for spatial calibration. More discussions on spatial calibration are preparation are provided in the Supplementary Materials
Spectral calibration
The purpose of spectral calibration is to locate the spectral position on the dispersed image for each lightguide. To obtain accurate spectral location, an NKT SuperK EXTREME supercontinuum white light laser was used to provide illumination step-by-step from 450nm to 650nm with a 10nm increment. The corresponding spectral image was captured to pin the central location for each wavelength band. The pixel distances from the central pixel of 640–650nm waveband to the rest wavebands were calculated and plotted in Figure 7(d). The total pixel distance of the whole wavelength range was approximately 496 pixels, leaving a gap of around 42 pixels between the head and tail of the dispersed spectra adjacent lightguides. It made efficient use of pixels between neighboring light guides and effectively prevented crosstalk. After spatial and spectral calibration, the central pixel coordinates for 15 × 9 light guides at 21 wavebands were obtained and plotted in Figure 7(e) with the marker’s color matching the wavelength.
Hyperspectral reconstruction
To retrieve the transmittance/reflectance signature of the object, flat-field correction is required to exclude the effect of illumination non-uniformity and the defects in the printed lightguides. In practice, an unload image IX under the same illumination without the sample was captured pair-wisely with the dispersed image IO with the sample. Two corresponding hypercubes HSX and HSO were then obtained after applying the spatial and spectral calibration on the two dispersed images. Their relationship can be described as Eqs. (2) and (3).
(2) |
(3) |
TO and TLG represent the transmittance of the sample and light guide separately, and iWL denotes the white light illumination distribution. In experiments, both illumination and lightguide experienced certain inhomogeneities and it was difficult to isolate them from each other. Flat-field correction compensated for such illumination and system non-uniformity by dividing HSO by HSX as described in Eq. (4).
(4) |
Besides non-uniformity, another factor which degraded the reconstruction quality was noise. The noise mainly came from the light leakage from the gaps between the circular lightguide in the input end and the leakage during propagating through the lightguide due to the local surface defects. In addition, flat-field correction might boost the noise especially in areas with low intensity. A background noise term was removed for each lightguide spectral profile to reduce the impact of light leakage. The background noise was determined experimentally based on the noise in the surrounding black area. A thresholding operation was then applied to reduce the noise caused by flat-field correction, which appeared as large singular values. This strategy thresholded these singular values according to the spectral information of the neighboring lightguides.
Figure 8(a) illustrates the image reconstruction pipeline with a Bayer filter sample in Figure 8(b). After spatial and spectral calibration, two 2D dispersive images IX and IO were decoded into two 3D hyperspectral cubes HSX and HSO. With flat-field correction and denoising, TO exhibited a clear partitioning of the three filter bars, as demonstrated by the hyperspectral images with 10nm bandwidth in Figure 8(c). From the hyperspectral images, an RGB color image was then reconstructed as shown in Figure 8(d) with the quantum efficiency of the CMOS camera used to capture Figure 8(b). Compared to the color image captured without the lightguide array and the dispersion prism, the HSHI outcome exhibited a general resemblance with few differences due to the imperfections of the 3D printed lightguide array. The first issue was that the number of lightguides was insufficient, leading to the pixelated structure. The second issue was the light leakage in the central portion of the lightguide array, which could be clearly observed from IX where the central region was brighter, and the contrast was lower. The light leakage was due to the fact that the input end of the lightguide was circular, not square, leaving gaps between lightguides. This issue could be solved by printing the lightguide with square input end so that the fill factor is close to 100%, regardless of square or round output end. The length of the lightguide needs to be long enough so that the radius of curved region of the lightguide is large enough to avoid the light leakage during light propagating inside the lightguide.
Experimental results
To demonstrate the concept and validate imaging performance, a manmade flower with a red flower petal and a green stem (Figure 9(a)) was imaged first. The red petal and green stem were made from the transparent color film lighting gel filters. Figure 9(b) is the dispersed image captured by the camera, the extracted hyperspectral images as well as the reconstructed color image are illustrated in Figure 9(c) and 9(d) respectively. The hyperspectral images at each wavelength band demonstrated the variation of the spectral response at different locations. The reconstructed color image displayed a good consistency compared with the ground truth image captured by a color CMOS camera, considering under-sampling issue due to the limited number of lightguides. To validate the imaging performance, the measured spectral profiles of the light from two lightguides, which are two pixels in Figure 9(d), are plotted and compared with the transmitted spectral profiles of the color film filters measured by Ocean Optics spectrometer in Figures 9(e) and 9(f), demonstrating that the HSHI system had the capability of capturing hyperspectral images with accurate spectral information. The small discrepancy between the two spectral profiles was mainly due to the light leakages between the circular lightguides in the input end, as discussed in the above section.
To demonstrate the potential application of the HSHI system in biology, an experiment was implemented with polytrichum leaf WM prepared microscope slide. Figure 10(a) illustrates the color image of the polytrichum sample, and its down-sampled image is plotted in Figure 10(b). The polytrichum sample showed distinct spatial inhomogeneity. Specifically, the left side of the sample exhibited a more greenish color, while the right side was more oriented to the yellow spectrum. Such inhomogeneity was also captured by the HSHI system and reflected upon the extracted hyperspectral images (Figure 10(d)) and the reconstructed color image (Figure 10(e)). In 460nm and 480nm, the left lightguides gave relatively strong response while in longer wavelength the right lightguides possessed higher intensity, which agreed with the spatial variation of the sample color. Because of the limited number of lightguides, to better evaluate the recovery performance, the color image in Figure 10(a) was down-sampled until it shared the same pixel number as the HSHI system, as shown in Figure 10(b). Under the same pixel scale, the HSHI output possessed a better similarity with the down-sampled image, demonstrating the potential application of the HSHI system in biology. The spectral profiles from the two selected lightguides are plotted in Figure 10(f) and 10(g). The spectral features of point A tended to be inclined to wavelength longer than 570 nm, while the spectral features of point B had a clear blue shift compared with A. It matched the RGB colors of the two points as A is more yellowish and B looks greener.
To demonstrate the snapshot imaging capability, a dandelion fuzz sample in Figure 11(a) was imaged by a dynamically varying illumination. In details, the illumination spectrum switched from 450nm to 650nm sequentially with a bandwidth of 10nm, along with one with 100nm bandwidth centered at 580nm. For each illumination spectrum, a dispersed image was captured from HSHI system, and the same hyperspectral reconstruction operations were applied to extract hypercube and reconstruct color image. To evaluate these outcomes, the ground-truth color images of the dandelion fuzz sample under the same illumination conditions were captured for comparison. The results of total 22 illumination scenarios were recorded to show the dynamic spectral recovery of the proposed HSHI system. Among 22 frames, the results of two snapshots with illumination centered at 510nm and 640nm are displayed in Figures 11(b)–11(e) and Figures 11(f)–11(i), respectively. Figures 11(c) and 11(g) are the unprocessed dispersed image captured from HSHI system. From 510nm to 640nm, the bright spots underwent a leftward shift which agreed with the spectral calibration result. Compared with the ground-truth color images in Figures 11(b) and 11(f), the recovered color image shown in Figures 11(d) and (h) demonstrated accurate spatial distributions and RGB colors, considering the pixelated effect due to the limited number of lightguides. The light spectral profiles for one selected lightguide (the yellow point in Figures 11(d) and 11(h)) under two illumination spectra are plotted in Figures 11(e) and 11(i), showing accurate center wavelength and bandwidth. All the spectral profiles were normalized based on the maximum value within the 22 frames.
Conclusion and Discussion
We have developed a compact, high-resolution snapshot hyperspectral imaging technology with 3D printed glass lightguide array. Compared to other methods, the 3D printed lightguide array has several unique advantages as an image sampler: it enables high spatial resolution imaging as the lightguide can be printed as small as the pixel size of the digital sensor; it can dramatically simplify the optical systems before and after the lightguide array as the input and output surfaces of the lightguide array can be curved; and potentially it is low cost as both the new 3D printing technologies and the simplified optical systems can reduce the system cost significantly. In addition, the glass lightguide array will enable the potential applications in UV and NIR, which are not possible with polymer lightguide array.
While we have demonstrated that the 3D TPP printing process is an excellent method for fabricating the lightguide array, some further developments are still needed. The first is to print large format lightguide arrays with much more lightguides, for example more than 640×480 lightguides. Increasing the number of the lightguides will increase the aspect ratio of the lightguide dramatically. While TPP 3D printers with long-range high precision motion stages, such as Nanoscribe’s Quantum X Shape and Upnano’s NanoOne, can print large structures with high resolution, significant efforts are still needed to optimize the printing process to print the lightguide array as each lightguide has a small dimension to achieve high spatial resolution and it is necessary to maintain TIR condition. Another challenge is the printing material, the cured structure should be strong enough without deformation during the printing process as it floats within the printing material. Therefore, the printing material should be optimized to ensure that the cured structure is strong enough. Since the light propagates through TIR inside the lightguide and there is no cladding layer like optical fiber, any dust on the lightguide surface will break the TIR condition, reducing light efficiency and introducing crosstalk. Therefore, the printed lightguide array should be encapsulated so that no dust will accumulate on the TIR surface of the lightguide.
Our near-term plans are to update the home built TPP printer with a high power 780 nm femtosecond fiber laser and 3-axis high precision stage with 26 mm travel range and 10 nm resolution, optimize pre-condensed liquid silica resin to improve curing speed and improve the strength of the cured structure, and optimize the printing process to print large format lightguide. We will also design custom optical systems to demonstrate the HSHI system with lightguide array which has concave input and output surfaces.
The proposed concept of lightguide array and the TPP process for printing precision optical elements will address some major limitations in current hyperspectral imaging systems. 3D printing process will not only fabricate individual components, such as lightguide array and lens, it can also fabricate the integrated system without the need of fixtures and active alignment, significantly reducing the cost and system dimension. Figure 12 demonstrates one lightguide with collimation lens, dispersion prism, and focusing lens. It opens new doors for compact HSHI systems, such as the systems in Figures 1(d) and 1(e).
With unique properties in resolution, compactness, and cost, we expect that the HSHI system will have a variety of applications in astronomy, remote sensing, surveillance, biotechnology, medical diagnosis, agriculture, pharmaceuticals, food analysis, oil and gas detection, machine vision, and etc., from UV to IR.
Supplementary Material
Funding
National Cancer Institute R21CA268190.
Footnotes
Disclosures
R.L. is the founder of Light Research Inc.
Z. H, Y. S, and P. Y contributed equally to this work. Z. H, P. Y, Y. S, D. L, and R. L conceived the idea and designed the study. P. Y prepared materials and measured the printed samples, Z. H performed printing experiments, and Y. S processed image. Z. H, Y. S, P. Y, D. L, and R. L analyzed and interpreted the results and wrote the manuscript.
References
- 1.Hagen Nathan A., and Kudenov Michael W.. “Review of snapshot spectral imaging technologies.” Optical Engineering 52, no. 9 (2013): 090901. [Google Scholar]
- 2.Lu Guolan, and Fei Baowei. “Medical hyperspectral imaging: a review.” Journal of biomedical optics 19, no. 1 (2014): 010901. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Boldrini Barbara, Kessler Waltraud, Rebner Karsten, and Kessler Rudolf W.. “Hyperspectral imaging: a review of best practice, performance and pitfalls for in-line and on-line applications.” Journal of near infrared spectroscopy 20, no. 5 (2012): 483–508. [Google Scholar]
- 4.Dong Xingchen, Jakobi Martin, Wang Shengjia, Köhler Michael H., Zhang Xiaoxing, and Koch Alexander W.. “A review of hyperspectral imaging for nanoscale materials research.” Applied Spectroscopy Reviews 54, no. 4 (2019): 285–305. [Google Scholar]
- 5.Bowen IS “The Image-Slicer a Device for Reducing Loss of Light at Slit of Stellar Spectrograph.” The Astrophysical Journal 88 (1938): 113. [Google Scholar]
- 6.Weitzel L, Krabbe A, Kroker H, Thatte N, Tacconi-Garman LE, Cameron M, and Genzel R. “3D: The next generation near-infrared imaging spectrometer.” Astronomy and Astrophysics Supplement Series 119, no. 3 (1996): 531–546. [Google Scholar]
- 7.Wang Ye, Pawlowski Michal E., Cheng Shuna, Dwight Jason G., Stoian Razvan I., Lu Jiawei, Alexander David, and Tkaczyk Tomasz S.. “Light-guide snapshot imaging spectrometer for remote sensing applications.” Optics Express 27, no. 11 (2019): 15701–15725. [DOI] [PubMed] [Google Scholar]
- 8.Gao Liang, Kester Robert T., Hagen Nathan, and Tkaczyk Tomasz S.. “Snapshot image mapping spectrometer (IMS) with high sampling density for hyperspectral microscopy.” Optics express 18, no. 14 (2010): 14330–14344. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Fletcher-Holmes David William, and Harvey Andrew Robert. “Real-time imaging with a hyperspectral fovea.” Journal of Optics A: Pure and Applied Optics 7, no. 6 (2005): S298. [Google Scholar]
- 10.Courtes G “An integral field spectrograph (IFS) for large telescopes.” In International Astronomical Union Colloquium, vol. 67, pp. 123–128. Cambridge University Press, 1982. [Google Scholar]
- 11.Ngo Tuan D., Kashani Alireza, Imbalzano Gabriele, Nguyen Kate TQ, and Hui David. “Additive manufacturing (3D printing): A review of materials, methods, applications and challenges.” Composites Part B: Engineering 143 (2018): 172–196. [Google Scholar]
- 12.Berglund Gregory, Wisniowiecki Anna, Gawedzinski John, Applegate Brian, and Tkaczyk Tomasz S.. “Additive manufacturing for the development of optical/photonic systems and components.” Optica 9, no. 6 (2022): 623–638. [Google Scholar]
- 13.Shrotri Abhijeet, Beyer Micha, and Stübbe Oliver. “Manufacturing and analyzing of cost-efficient fresnel lenses using stereolithography.” In 3D Printed Optics and Additive Photonic Manufacturing II, vol. 11349, pp. 33–42. SPIE, 2020. [Google Scholar]
- 14.Sung Yu-Lung, Jeang Jenn, Lee Chia-Hsiung, and Shih Wei-Chuan. “Fabricating optical lenses by inkjet printing and heat-assisted in situ curing of polydimethylsiloxane for smartphone microscopy.” Journal of biomedical optics 20, no. 4 (2015): 047005. [DOI] [PubMed] [Google Scholar]
- 15.Chen Xiangfan, Liu Wenzhong, Dong Biqin, Lee Jongwoo, Ware Henry Oliver T., Zhang Hao F., and Sun Cheng. “High‐speed 3D printing of millimeter‐size customized aspheric imaging lenses with sub 7 nm surface roughness.” Advanced Materials 30, no. 18 (2018): 1705683. [DOI] [PubMed] [Google Scholar]
- 16.Sun C, Fang N, Wu DM, and Zhang X. “Projection micro-stereolithography using digital micro-mirror dynamic mask.” Sensors and Actuators A: Physical 121, no. 1 (2005): 113–120. [Google Scholar]
- 17.Blachowicz Tomasz, Ehrmann Guido, and Ehrmann Andrea. “Optical elements from 3D printed polymers.” e-Polymers 21, no. 1 (2021): 549–565. [Google Scholar]
- 18.Hong Zhihan, and Liang Rongguang. “IR-laser assisted additive freeform optics manufacturing.” Scientific reports 7, no. 1 (2017): 1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Hong Zhihan, Ye Piaoran, Loy Douglas A., and Liang Rongguang. “Three-dimensional printing of glass micro-optics.” Optica 8, no. 6 (2021): 904–910. [Google Scholar]
- 20.Hong Zhihan, Ye Piaoran, Loy Douglas A., and Liang Rongguang. “High-Precision Printing of Complex Glass Imaging Optics with Precondensed Liquid Silica Resin.” Advanced Science (2022): 2105595. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Gissibl Timo, Thiele Simon, Herkommer Alois, and Giessen Harald. “Two-photon direct laser writing of ultracompact multi-lens objectives.” Nature photonics 10, no. 8 (2016): 554–560. [Google Scholar]
- 22.Juodkazis Saulius. “3D printed micro-optics.” Nature Photonics 10, no. 8 (2016): 499–501. [Google Scholar]
- 23.Rim Seung-Bum, Catrysse Peter B., Dinyari Rostam, Huang Kevin, and Peumans Peter. “The optical advantages of curved focal plane arrays.” Optics Express 16, no. 7 (2008): 4965–4971. [DOI] [PubMed] [Google Scholar]
- 24.Guenter Brian, Joshi Neel, Stoakley Richard, Keefe Andrew, Geary Kevin, Freeman Ryan, Hundley Jake et al. “Highly curved image sensors: a practical approach for improved optical performance.” Optics express 25, no. 12 (2017): 13010–13023. [DOI] [PubMed] [Google Scholar]
- 25.Lee Gil Ju, and Song Young Min. “Compact zooming optical systems for panoramic and telescopic applications based on curved image sensor.” Journal of Optical Microsystems 2, no. 3 (2022): 031204. [Google Scholar]
- 26.Olivas Stephen J., Arianpour Ashkan, Stamenov Igor, Morrison Rick, Stack Ron A., Johnson Adam R., Agurok Ilya P., and Ford Joseph E.. “Image processing for cameras with fiber bundle image relay.” Applied Optics 54, no. 5 (2015): 1124–1137. [DOI] [PubMed] [Google Scholar]
- 27.Schröer Wolfgang-D. “Fine structure of the anterior median eyes of the funnel-web spider Agelena labyrinthica (Araneae: Agelenidae).” Arthropod structure & development 46, no. 2 (2017): 196–214. [DOI] [PubMed] [Google Scholar]
- 28.Duke-Elder Stewart. “The eye in evolution.” System of Ophthal mology 1 (1958). [Google Scholar]
- 29.Gupta Anurag, Lee Junwon, and John Koshel R. “Design of efficient lightpipes for illumination by an analytical approach.” Applied optics 40, no. 22 (2001): 3640–3648. [DOI] [PubMed] [Google Scholar]
- 30.Vaezi Mohammad, Seitz Hermann, and Yang Shoufeng. “A review on 3D micro-additive manufacturing technologies.” The International Journal of Advanced Manufacturing Technology 67, no. 5 (2013): 1721–1754. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.