Abstract
In this paper, we present a prototype pseudo-direct time-of-flight (ToF) CMOS image sensor, achieving high distance accuracy, precision, and robustness to multipath interference. An indirect ToF (iToF)-based image sensor, which enables high spatial resolution, is used to acquire temporal compressed signals in the charge domain. Whole received light waveforms, like those acquired with conventional direct ToF (dToF) image sensors, can be obtained after image reconstruction based on compressive sensing. Therefore, this method has the advantages of both dToF and iToF depth image sensors, such as high resolution, high accuracy, immunity to multipath interference, and the absence of motion artifacts. Additionally, two approaches to refine the depth resolution are explained: (1) the introduction of a sub-time window; and (2) oversampling in image reconstruction and quadratic fitting in the depth calculation. Experimental results show the separation of two reflections 40 cm apart under multipath interference conditions and a significant improvement in distance precision down to around 1 cm. Point cloud map videos demonstrate the improvements in depth resolution and accuracy. These results suggest that the proposed method could be a promising approach for virtually implementing dToF imaging suitable for challenging environments with multipath interference.
Keywords: pseudo-dToF imaging, charge-domain time-compressive sensing, multipath interference
1. Introduction
Time-of-flight (ToF) depth imaging calculates the distance of an object through illuminating a scene with a modulated light source and measuring the travel time of light emitted from the camera system to the object and reflected back to the camera. This technology enables high-accuracy and high-frame-rate 3D imaging using a compact system consisting of a CMOS image sensor together with an active modulated light source. Because of these advantages, ToF measurement has been realized in a wide range of machine vision applications, such as gesture capture and behavior recognition, and visual positioning and guidance [1,2,3]. However, the current mainstream ToF sensors can suffer from a relatively low spatial resolution [4,5] due to their large digital circuit areas, or scene-dependent errors due to multipath interference (MPI) problems [6]. Light from multiple paths, caused by interreflections and scattering, enters the same pixel, resulting in depth errors. Therefore, ToF sensors with both high spatial resolution and robustness to MPI are desired for realizing simpler and more accurate depth camera systems.
There are two major methodologies for traditional ToF depth imaging: direct ToF (dToF) and indirect ToF (iToF). The dToF sensors directly measure the ToF of incident light and express it as a waveform of reflected light intensity. This approach necessitates fast-response photodetectors with single-photon sensitivity, and in most cases, single-photon avalanche diodes (SPADs) [7,8] are utilized. The dToF sensors are relatively robust against MPI because it is easier and simpler to compensate for MPI since the entire reflected light waveform is observed. However, they yield lower spatial resolution because of the large area required for complex circuitry such as time-to-digital converters (TDCs) and histogram builders. In contrast, iToF sensors [9] use the correlation between the incident light and the time windows to estimate the ToF or the temporal phase difference. The correlation is given through the number of modulated photoelectrons corresponding to each time window. The iToF sensors are suitable for large pixel counts thanks to their simple circuits. However, they are susceptible to MPI due to the loss of reflected light waveform information after correlation. In addition, the multiple measurements for different time window timings often used in the amplitude modulation continuous wave (AMCW) method lead to motion artifacts in dynamic scenes [10]. Therefore, with the conventional approaches, it is difficult to achieve an appropriate balance among high spatial resolution, high accuracy, and tolerance to MPI.
In a paper at IISW 2023 [11], we presented a novel method called pseudo-direct time-of-flight (pseudo-dToF) depth imaging, which provides the advantages of both dToF and iToF image sensors. This method utilizes high-speed charge modulators for iToF, allowing for high-spatial-resolution imaging like conventional iToF image sensors [12,13]. Random exposure codes are applied to time windows in demodulation to perform time-compressive sensing in the charge domain [14,15,16]. This key technique allows us to acquire the entire light waveform in a single shot like those obtained using dToF sensors. Therefore, pseudo-dToF imaging provides high accuracy, robustness to MPI, and motion-artifact-free depth measurement. However, the depth precision is limited by the maximum modulation frequency of the time window, which corresponds to the bin size in the dToF. To overcome this drawback without any hardware modification, we presented two approaches. First, a sub-time window is introduced to halve the time resolution. Second, the sensor responses are oversampled to increase the number of reconstructed data points, and quadratic fitting is performed in the depth calculation to refine the distance results. The fitting is similar to the technique used in dToF to realize sub-bin depth resolution [17,18].
In this paper, we additionally describe the detailed sensor configuration and the reconstruction signal processing. ToF measurement videos are also presented, visually confirming the benefits of the proposed method in enhancing the time resolution without changing the hardware. Videos S1–S4 are available as Supplementary Materials.
2. Pseudo-dToF Depth Imaging
In our scheme, ToF imaging can be divided into three phases: pre-measurement, sensing, and signal reconstruction.
In the pre-measurement phase, all data used for the subsequent signal reconstruction are prepared. Exposure codes are generated, and the instrument response functions (IRFs) of the imaging optics, the light source, and the image sensor are measured in advance. Note that this preparation step needs to be performed only once for all measurements. During the sensing phase, the camera captures temporal signals of the incident light emitted from a synchronized laser and reflected to the camera through a single-aperture lens. The image sensor consists of macro-pixels based on charge modulators, as shown in Figure 1. During the image acquisition, exposure codes, i.e., temporal electronic shutters, are provided for each tap of the charge modulator in each pixel. Thus, incident light signals are temporally compressed in the charge domain, yielding four temporally compressed images in one shot. In the reconstruction phase, the input optical waveforms for all sub-pixels are reconstructed. Through solving an inverse problem based on sparsity regularization, temporally sequential images or transient images of light are obtained. The depth calculation is then processed as in normal dToF imaging. The arrival times of the peaks of the light waveform are detected with quadratic curve fitting, and finally, they are converted to a depth image using the speed of light.
2.1. Multi-Tap Macro-Pixel CMOS Image Sensor
This sensor utilizes a charge modulator-based iToF image sensor, which enables a small pixel size or high-spatial-resolution imaging. Figure 2 shows the sensor block diagram and the pixel structure. The image sensor is composed of a pixel array, vertical and horizontal scanners, on-chip ADCs, and a readout circuit as a typical CMOS image sensor. The peculiar circuits in this sensor include the lab-designed macro-pixel array, the shutter controller, and the charge modulator drivers that control the signal accumulation in the macro-pixels. Every macro-pixel consists of four sub-pixels, each of which is a four-tap lateral electric field charge modulator (LEFM) [19]. The LEFM is composed of a photodiode, and four charge transfer gates and four storage diodes (), configured as a pair referred to as a tap. Here, the index identifies the tap. In the capture phase, the shutter controller outputs the pre-loaded exposure codes, which are applied to the transfer gates of each pixel via the driver circuit operating at a high-speed clock. All sub-pixels perform the charge modulation according to the exposure codes. When the exposure time is finished, the captured images are read out by the ADCs and the readout circuit.
Transfer of the generated photoelectrons depends on the exposure codes assigned to each transfer gate. These exposure codes, representing the time window function, are binary, as shown in Figure 3. A value of “1” indicates the transfer of photoelectrons, while “0” signifies no charge to be transferred. Through employing temporally randomized exposure codes, the incident light signals are temporally compressed in the charge domain. Considering that a time-variant optical signal is subjected to a designated time window function at tap , the output integrated charge value can be defined through correlation:
(1) |
where is the exposure time for each frame.
Table 1 shows the specifications of the prototype image sensor. In our previous research, we confirmed that the sensor was driven at a maximum modulation frequency of up to 303 MHz [19]. This high-speed operation of the LEFM pixels allows for high temporal resolution in ToF imaging. Also, no signal processing circuits are necessary for compressive sensing because compression processing is done through charge transfer. The only additional circuit is the shutter controller that generates exposure code, which can be implemented as a relatively small digital circuit, thus enabling high spatial resolution.
Table 1.
Technology | 0.11 μm FSI CIS |
Chip size | 7.0 mmH × 9.3 mmV |
Macro-pixel size | 22.4 μmH × 22.4 μmV |
Effective sub-pixel count | 212H × 188V |
Sub-pixel count per macro-pixel | 2H × 2V |
Maximum exposure code length | 256 bits by 8 bits |
Maximum modulation frequency | 303 MHz |
Maximum clock frequency | 607 MHz |
Maximum image readout frame rate 1 | 21 fps |
Power consumption | 2.8 W |
1 Readout frame rate when the exposure time is set to zero. The actual image readout frame rate is given through the reciprocal of the formula: (exposure code bit length × bit period) × number of light pulses + readout time, where the readout time is 47.6 ms (1/21 fps).
2.2. Compressive Sensing
Compressive sensing [20,21] is an efficient sampling method that reconstructs more data points from fewer samples when the original signal is sparse. We employ this principle to reconstruct the entire incident light waveform from the compressed output images. This allows pseudo-dToF imaging to obtain dToF-like signals that are robust against MPI, enabling high depth accuracy while being comparable with the conventional iToF image sensors in terms of their architecture and components. Note that compressed images are obtained in a single shot, which makes our sensor free of motion artifacts.
Time-compressive sensing and signal reconstruction are mathematically described as follows. Considering the characteristics of the imaging optics and charge modulations, Equation (1) can be expanded to:
(2) |
Here, represents the exposure code, is the sensor’s temporal IRF or sensor impulse response, is the spatial IRF of the imaging optics, shows incident light intensity at where indicates a two-dimensional spatial coordinate. The operator indicates convolution. The exposure codes are repeatedly applied during the exposure duration to obtain high photosensitivity through multiple exposures. Collectively, the exposure code, the sensor’s IRF, and the spatial IRF can be referred to as the spatio-temporal observation matrix . Equation (2) is then expressed in discrete values as follows:
(3) |
In cases where the detected signal is lower in dimensionality than the original input signal , the temporal information of is compressed. Retrieving the input signal from becomes an ill-posed problem. However, if is K-sparse, implying that only K elements have non-zero values, estimating the solution for becomes feasible through optimizing the L1 norm. In this research, we estimate the closest solution for through minimizing the total variation, as shown in:
(4) |
Here, represents a spatio-temporal differential operator. This process is performed by TVAL3 [22], a compressed sensing solver, using an iterative method. Matrix is selected as a random matrix to minimize the correlation between the row and column vectors of the observation matrix [20]. As matrix significantly influences the quality of the reconstruction, a random exposure pattern is generated and optimized in the pre-measurement phase. In this study, three different images were used for the reconstruction in order to mitigate the influence of the scene on the reconstruction results. For each image, we generated the exposure code at random, reconstructed the image, and iterated this process 30 times. The exposure code yielding the image with the highest average PSNR was selected. Note that due to the structure of the multi-tap macro-pixel, creating an ideal random exposure code with 16 taps may not be feasible. Consequently, it becomes necessary to create a hypothetical random pattern under specific generating conditions. For example, in a sub-pixel, only one tap can be turned on at any given time, and there must always be one active tap. Based on this principle of compressive sensing, the reflected light waveforms of all sub-pixels are reconstructed from the four compressed images and the pre-measured observation matrix .
Depth information is calculated after the reconstruction of transient images. To generate a depth map, the reconstructed waveform of each pixel is analyzed to identify the peak arrival time, which corresponds to the round-trip ToF. Finally, it is converted into a depth image using the speed of light. In pseudo-dToF imaging, the measurement range and depth resolution are dependent on the bit length of the exposure code and the frame rate of the reconstructed transient images. In the simplest situation, the temporal resolution is determined using the minimum time window width of the exposure code, which corresponds to the bin width of the dToF histogram. The measurement range is calculated through multiplying the depth resolution by the bit length. For example, if our sensor operates at 303 MHz and uses a 32-bit exposure code, the temporal resolution is 3.3 ns (or 0.5 m in depth), and the measurable distance is 16 m. Note that sub-time-window interpolation is often used to improve the depth resolution.
3. Improvement of Depth Resolution
So far, we have successfully performed ToF imaging at an operating clock frequency of 303 MHz [19]. This allowed a minimum time window with a duration of 3.3 ns, which corresponds to a 0.5 m resolution when only the frame position that gives the highest pixel value is considered. Generally, higher depth resolution can be achieved simply through increasing the operating frequency of the charge modulator. However, it was challenging to achieve a minimum time window duration below 3.3 ns due to the performance limitations of the charge modulators. Driving the charge modulation at 303 MHz is quite fast compared to the fastest modulation speed of 320 MHz [23], and realizing a higher frequency requires not only advanced fabrication technology but also an improved modulator and driver structures. To overcome these restrictions, the following two approaches were employed to obtain higher depth resolution without any hardware modifications.
3.1. Sub-Time Window
Here, a sub-time window of the exposure code is introduced to enhance the temporal resolution without affecting the minimum time window duration, which is determined by the modulator’s performance. To implement the sub-time window, we double the frequency of the on-chip phase locked loop (PLL) from 303 MHz to 607 MHz while maintaining the minimum time window duration at 3.3 ns (equivalent to two clocks). As shown in Figure 4, the rising and falling edges of the time window can be shifted by one clock unit. This shift is half of the minimal time window, resulting in halving of the temporal resolution to 1.65 ns.
3.2. Oversampling and Peak Detection with Fitting
Another approach to improve the temporal resolution is to oversample the IRF. When reconstructing the transient images, the quantity of reproduced frames is equal to the number of data points in the IRF or the observation matrix . Consequently, increasing the number of data points in the pre-measured observation matrix will subsequently increase the number of reconstructed frames, resulting in finer temporal resolution.
Normally, the IRF is sampled once per bit of the exposure code. However, in this study, we sub-sampled 10 points per bit (10× sampling), as depicted in Figure 5. A light waveform is then reconstructed with improved temporal (or depth) resolution. Since the signal waveforms can be obtained as in dToF imaging, we perform fitting to refine the depth. After the peak position in the waveform is detected, the five data points around the peak are considered in order to fit a quadratic curve to estimate the exact signal peak. Note that a similar depth calculation technique based on fitting or filtering is utilized in dToF to realize sub-bin depth resolution [17,18].
The depth accuracy and precision of the pseudo-dToF method are significantly improved because higher-density light waveforms through oversampling are utilized in depth estimation with fitting. This approach can distinguish and detect each peak separately even when there are multiple peaks due to MPI. It is expected that the pseudo-dToF method will become applicable even in challenging environments with MPI. Note that the depth resolution is improved without altering the image sensor hardware or the total measurement time, although the reconstruction process takes longer.
4. Experimental Results
For all experiments in this study, 32-bit long exposure codes corresponding to a range of 16 m at a clock frequency of 606 MHz were selected to accommodate the objects in the experimental scenes. The exposure code length determines the depth range and the compression ratio. As a result, a moderate compression ratio of 8× in the total spatio-temporal sampling points (2× in the temporal domain and 4× in the number of pixels) was achieved. In terms of the compression ratio, image reconstruction becomes more challenging at higher compression ratios, meaning longer exposure codes.
Firstly, we verified the feasibility of driving the sensor with doubled speed at 607 MHz while maintaining the minimum time window width. Figure 6 shows the temporal system response measured when this sensor was operated at 607 MHz. The exposure code was a 32-bit random binary code, and a short-pulse semiconductor laser (PicoQuant, Belin, Germany, LDH-IB-450-M-P, wavelength of 443 nm) with a pulse width of 228 ps was used. The pixel values were averaged for each tap, and dark frames were subtracted to reduce the dark noise level. It can be observed that charge modulation was successfully performed.
To evaluate the improvement of depth resolution with the sub-time window approach, we compared the signal peak separation performance under MPI conditions. In this experiment, interference light was introduced through placing a weakly diffusive plastic sheet in front of a panel serving as an object, with the black letters “SU” on it, as shown in Figure 7a. The letters look blurry due to the diffuser. The distance between the two objects was set to 0.65 m and 0.40 m, and experiments were conducted with the sensor operating at 303 MHz and 607 MHz. An ordinary single-aperture imaging lens (Edmund Optics, C series VIS-NIR fixed focal length lens, focal length of 16 mm, f/1.4) was used. The light source and exposure codes were the same as in the measurements shown in Figure 6. The actual image readout frame rates were 16.28 fps at 303 MHz and 18.34 fps at 607 MHz.
Next, a ToF imaging experiment was conducted to verify the effectiveness of oversampling for a ratio of 10. Note that the time/depth resolutions at 303 MHz and 607 MHz (with normal sampling), and 607 MHz (with 10× sampling) are 3.3 ns/0.5 m, 1.75 ns/0.25 m, and 0.175 ns/0.025 m, respectively. A pulsed semiconductor laser (Tama Electric Inc., Hamamatsu, Japan, LDB-300C) with a wavelength of 660 nm and an FWHM of 2.55 ns was used. One hundred images were averaged to improve the SNR. The actual image readout frame rates were 8.57 fps at 303 MHz and 12.13 fps at 607 MHz. The image processing was performed using MATLAB R2020b on an Intel® Core™ i7-9700 CPU (Intel, Santa Clara, CA, USA) running at 3.00 GHz equipped with 32 GB of memory. The image reconstruction time is approximately N times longer with N× sampling compared to without oversampling. This increase in time is primarily because the column dimension of the observation matrix A in Equation (3) becomes N times larger. Specifically, the reconstruction time for an image captured with 32-bit exposure codes is approximately 200 s without oversampling, and averages about 1800 s with 10× sampling. Notably, no acceleration through parallel processing was utilized in the reconstruction. Figure 8 shows the configuration of the targets.
Figure 9a,b compares the depth maps for each condition in frontal and right-side views. Figure 9c shows the reconstructed optical waveforms and their fitting curves with a quadratic approximation. The mean and standard deviation of the depths are quantitatively compared in Table 2 to demonstrate the effectiveness of the proposed approaches. In the result at 607 MHz, 10× sampling has the closest mean depths to the actual values and the smallest standard deviation of about 1 cm.
Table 2.
Real Depth (m) |
303 MHz, Normal Sampling |
607 MHz, Normal Sampling |
607 MHz, 10× Sampling |
||||
---|---|---|---|---|---|---|---|
Mean (m) | Std (cm) | Mean (m) | Std (cm) | Mean (m) | Std (cm) | ||
Cork | 1.00 | 1.04 | 2.53 | 1.01 | 1.17 | 1.00 | 0.94 |
Fan | 2.00 | 1.66 | 5.59 | 1.98 | 1.14 | 1.99 | 0.72 |
Panel | 3.00 | 3.03 | 3.67 | 3.08 | 5.60 | 3.00 | 0.57 |
ToF videos were captured using the proposed method. Two setups are shown in Figure 10a and Figure 11a, one without and the other with MPI. In the first acquisition, a man was observed walking toward the camera system in the background with obstacles. These shots were filmed in a dark room, using the same equipment as the experiment shown in Figure 8. The actual image readout frame rate was 7.68 fps. The videos in Figure 10b,c display the reconstructed 3D point cloud map without image averaging, at a PLL clock frequency of 303 MHz for normal sampling and 10× sampling. When we compare the two videos, oversampling and quadratic fitting yield cleaner, less noisy, and more defined outcomes. The object shapes in Figure 10c are more distinct and precise. These enhancements can significantly benefit subsequent depth-related tasks, like segmentation or gesture recognition, for improved quality. Furthermore, the superiority of pseudo-dToF imaging for dynamic scenes was confirmed since no motion artifacts were observed even for moving objects. Note that in this experiment, the man was wearing protective glasses, and the hair had low reflectivity. Hence, the head is difficult to capture at a long distance and only becomes clear when the man moves closer to the camera.
The setup for the video containing the MPI is the same as the experiment depicted in Figure 7. In this acquisition, the diffuser was positioned in front of the panel and moved vertically. The actual image readout frame rate was 17.24 fps. Both results, one with and one without oversampling, accurately captured the depth of the diffuser and the panel, even while the diffuser was moving. Figure 11c represents the result after applying oversampling and fitting, showing improved results with smoother and sharper objects. However, there is some crosstalk between the two objects, which is probably influenced by the difference between the measured and actual IRFs.
5. Conclusions
In this paper, we demonstrated the concept, benefits, and implementation of pseudo-dToF imaging. The iToF-based high-speed charge modulators are utilized to achieve high spatial resolution. The time-compressive sensing in the charge domain is performed through applying 32-bit random exposure codes in demodulation. Since the image reconstruction yields dToF-like output waveforms, pseudo-dToF is immune to MPI and motion artifacts. To improve the temporal resolution and estimated depth precision, two approaches are applied: a sub-time window and oversampling. While there are still some challenges, such as long processing time, vulnerability to ambient light, and so on, the proposed method remains a promising technique for applications like autonomous vehicles, and robotics in challenging environments with MPI. To further increase the compression ratio and to reduce the image reproduction time, longer exposure codes and image reproduction algorithms based on deep neural networks are being developed.
Acknowledgments
This work was also supported by the VLSI Design and Education Center (VDEC), The University of Tokyo, in collaboration with Cadence Corporation, Synopsys Corporation, and Mentor Graphics Corporation. The authors would like to thank DB HiTek for the chip fabrication. The authors are grateful to Futa Mochizuki, Yuto Sato, Tomoya Kokado, Michio Fukuda, Masaya Horio, Tatsuki Furuhashi, and Taishi Takasawa for their contributions to the development of the prototypes of the image sensor.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23239332/s1, Video S1: Figure 10b, Video S2: Figure 10c, Video S3: Figure 11b, Video S4: Figure 11c.
Author Contributions
Conceptualization, H.N. and K.K.; experiments and data processing, A.N.P. and T.I; algorithm and programming, T.I.; image sensor design, K.Y., S.K. and K.K.; supervision and project administration, K.K.; funding acquisition, K.K.; writing original draft, A.N.P.; writing review and editing, K.K. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data are contained within the article and Supplementary Materials.
Conflicts of Interest
The authors declare no conflict of interest.
Funding Statement
This research was funded by JST, CREST, JPMJCR22C1, and in part by Grants-in-Aid for Scientific Research (S), grant numbers 17H06102 and 18H05240.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Soutschek S., Penne J., Hornegger J., Kornhuber J. 3-D gesture-based scene navigation in medical imaging applications using Time-of-Flight cameras; Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops; Anchorage, AK, USA. 23–28 June 2008. [Google Scholar]
- 2.Uebersax D., Gall J., Van den Bergh M., Van Gool L. Real-time sign language letter and word recognition from depth data; Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops); Barcelona, Spain. 6–13 November 2011. [Google Scholar]
- 3.Van den Bergh M., Van Gool L. Combining RGB and ToF cameras for real-time 3D hand gesture interaction; Proceedings of the 2011 IEEE Workshop on Applications of Computer Vision (WACV); Kona, HI, USA. 5–7 January 2011. [Google Scholar]
- 4.Chakraborty B.K., Sarma D., Bhuyan M.K., MacDorman K.F. Review of constraints on vision-based gesture recognition for human–computer interaction. IET Comput. Vis. 2018;12:3–15. doi: 10.1049/iet-cvi.2017.0052. [DOI] [Google Scholar]
- 5.Jarabo A., Masia B., Marco J., Gutierrez D. Recent advances in transient imaging: A computer graphics and vision perspective. Vis. Inform. 2017;1:65–79. doi: 10.1016/j.visinf.2017.01.008. [DOI] [Google Scholar]
- 6.Bhandari A., Kadambi A., Whyte R., Barsi C., Feigin M., Dorrington A., Raskar R. Resolving multipath interference in time-of-flight imaging via modulation frequency diversity and sparse regularization. Opt. Lett. 2014;39:1705. doi: 10.1364/OL.39.001705. [DOI] [PubMed] [Google Scholar]
- 7.Niclass C., Rochas A., Besse P.A., Charbon E. Design and characterization of a CMOS 3-D image sensor based on single photon avalanche diodes. IEEE J. Solid-State Circuits. 2005;40:1847–1854. doi: 10.1109/JSSC.2005.848173. [DOI] [Google Scholar]
- 8.Ximenes A.R., Padmanabhan P., Lee M.-J., Yamashita Y., Yaung D.N., Charbon E. A 256 × 256 45/65 nm 3D-stacked SPAD-based direct TOF image sensor for LiDAR applications with optical polar modulation for up to 18.6 dB interference suppression; Proceedings of the 2018 IEEE International Solid—State Circuits Conference—(ISSCC); San Francisco, CA, USA. 11–15 February 2018. [Google Scholar]
- 9.Remondino F., Stoppa D. TOF Range-Imaging Cameras. Springer; Berlin/Heidelberg, Germany: 2013. [Google Scholar]
- 10.Lindner M., Kolb A. Dynamic 3D Imaging. Springer; Berlin/Heidelberg, Germany: 2009. Compensation of Motion Artifacts for Time-of-Flight Cameras. [Google Scholar]
- 11.Pham A.N., Thoriq I., Yasutomi K., Kawahito S., Nagahara H., Kagawa K. A 607 MHz time-compressive computational pseudo-dToF CMOS image sensor; Proceedings of the 2023 International Image Sensor Workshop; Edinburgh, UK. 21–25 May 2023. [Google Scholar]
- 12.Keel M.-S., Kim D., Kim Y., Bae M., Ki M., Chung B., Son S., Lee H., Jo H., Shin S.-C., et al. 7.1 A 4-tap 3.5 μm 1.2 Mpixel Indirect Time-of-Flight CMOS Image Sensor with Peak Current Mitigation and Multi-User Interference Cancellation; Proceedings of the 2021 IEEE International Solid-State Circuits Conference (ISSCC); San Francisco, CA, USA. 13–22 February 2021. [Google Scholar]
- 13.Stoppa D., Massari N., Pancheri L., Malfatti M., Perenzoni M., Gonzo L. An 80 × 60 range image sensor based on 10 μm 50 MHz lock-in pixels in 0.18 μm CMOS; Proceedings of the 2010 IEEE International Solid-State Circuits Conference—(ISSCC); San Francisco, CA, USA. 7–11 February 2010. [Google Scholar]
- 14.Mochizuki F., Kagawa K., Okihara S-i Seo M.-W., Zhang B., Takasawa T., Yasutomi K., Kawahito S. 6.4 Single-shot 200 Mfps 5 × 3-aperture compressive CMOS imager; Proceedings of the 2015 IEEE International Solid-State Circuits Conference—(ISSCC) Digest of Technical Papers; San Francisco, CA, USA. 2–26 February 2015. [Google Scholar]
- 15.Mochizuki F., Kagawa K., Okihara S-i Seo M.-W., Zhang B., Takasawa T., Yasutomi K., Kawahito S. Single-event transient imaging with an ultra-high-speed temporally compressive multi-aperture CMOS image sensor. Opt. Express. 2016;24:4155. doi: 10.1364/OE.24.004155. [DOI] [PubMed] [Google Scholar]
- 16.Kagawa K., Kokado T., Sato Y., Mochizuki F., Nagahara H., Takasawa T., Yasutomi K., Kawahito S. Multi-tap macro-pixel based compressive ultra-high-speed CMOS image sensor; Proceedings of the 2019 International Image Sensor Workshop; Snowbird, UT, USA. 24–27 June 2019. [Google Scholar]
- 17.Chen G., Wiede C., Kokozinski R. Data Processing Approaches on SPAD-Based d-TOF LiDAR Systems: A Review. IEEE Sens. J. 2021;21:5656–5667. doi: 10.1109/JSEN.2020.3038487. [DOI] [Google Scholar]
- 18.Kumagai O., Ohmachi J., Matsumura M., Yagi S., Tayu K., Amagawa K., Matsukawa T., Ozawa O., Hirono D., Shinozuka Y., et al. 7.3 A 189 × 600 Back-Illuminated Stacked SPAD Direct Time-of-Flight Depth Sensor for Automotive LiDAR Systems; Proceedings of the 2021 IEEE International Solid-State Circuits Conference (ISSCC); San Francisco, CA, USA. 13–22 February 2021. [Google Scholar]
- 19.Kagawa K., Horio M., Pham A.N., Ibrahim T., Okihara S-i Furuhashi T., Takasawa T., Yasutomi K., Kawahito S., Nagahara H. A Dual-Mode 303-Megaframes-per-Second Charge-Domain Time-Compressive Computational CMOS Image Sensor. Sensors. 2022;22:1953. doi: 10.3390/s22051953. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Baraniuk R. Compressive Sensing [Lecture Notes] IEEE Signal Process. Mag. 2007;24:118–121. doi: 10.1109/MSP.2007.4286571. [DOI] [Google Scholar]
- 21.Qaisar S., Bilal R.M., Iqbal W., Naureen M., Lee S. Compressive sensing: From theory to applications, a survey. J. Commun. Netw. 2013;15:443–456. doi: 10.1109/JCN.2013.000083. [DOI] [Google Scholar]
- 22.Li C., Yin W., Jiang H., Zhang Y. An efficient augmented Lagrangian method with applications to total variation minimization. Comput. Optim. Appl. 2013;56:507–530. doi: 10.1007/s10589-013-9576-1. [DOI] [Google Scholar]
- 23.Bamji C.S., Mehta S., Thompson B., Elkhatib T., Wurster S., Akkaya O., Payne A., Godbaz J., Fenton M., Rajasekaran V., et al. IMpixel 65 nm BSI 320 MHz demodulated TOF Image sensor with 3 μm global shutter pixels and analog binning; Proceedings of the 2018 IEEE International Solid—State Circuits Conference—(ISSCC); San Francisco, CA, USA. 11–15 February 2018. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data are contained within the article and Supplementary Materials.