Skip to main content
Photoacoustics logoLink to Photoacoustics
. 2020 Mar 11;18:100173. doi: 10.1016/j.pacs.2020.100173

Three-dimensional clinical handheld photoacoustic/ultrasound scanner

Changyeop Lee a,1, Wonseok Choi b,1, Jeesu Kim c, Chulhong Kim d,*
PMCID: PMC7090348  PMID: 32215250

Abstract

Clinical 2D photoacoustic (PA) imaging can be easily implemented in a traditional ultrasound (US) system. However, 3D PA imaging is still preferable because 2D B-mode PA/US imaging suffers from low reproducibility and high-operator dependency. Here, we demonstrate a compact clinical handheld 3D PA/US scanner using an 1D linear array US transducer combined with a mechanical scanning stage working via a Scotch yoke mechanism. The entire scanner measures just 100 × 80 × 100 mm3 and weighs only 950 g, so it can easily be operated by hand. Blood vessels and hemoglobin oxygen saturation images of different parts of the human body (e.g., neck, wrist, thigh, and instep) have been successfully acquired. The system can potentially be used for clinical applications in fields such as oncology, dermatology, nephrology, and internal medicine.

Keywords: Clinical translation, Handheld scanner, Photoacoustic, Scotch yoke mechanism, Three-dimensional imaging

1. Introduction

Photoacoustic (PA) imaging, which utilize ultrasound (US) signals from objects that have absorbed light, has become a popular medical imaging technique. When a tissue region of interest (ROI) is illuminated with short-pulsed laser light, US signals are generated by the slight thermoelastic expansions resulting from the absorption of each pulse. PA images are created by receiving the US signals with a US transducer and using beamforming techniques to reconstruct the signals into images.

Photoacoustic imaging (PAI, also called optoacoustic imaging) can map the distribution of endogenous chromophores such as oxy-hemoglobin, deoxy-hemoglobin, lipids, water, and melanin [1,2]. Furthermore, by providing functional information (e.g., total hemoglobin concentration and hemoglobin oxygen saturation (sO2)), PAI can aid in diagnosing diseases such as cancers and peripheral diseases [[3], [4], [5], [6], [7]]. In addition, PAI can be readily combined with widely-used US imaging on a single platform [[8], [9], [10]], creating a dual-modality PA/US imaging system that can simultaneously supply multi-parametric morphological, physiological, and molecular information [[11], [12], [13], [14], [15], [16], [17]].

In most cases, PAI is performed in 2D to provide a cross-sectional image, but the accompanying low reproducibility and high operator dependence are major problems with this approach [18]. For these reasons, automatic 3D imaging can be a promising alternative, and it is typically conducted with either a 2D array of US transducers (e.g., matrix or hemispherical) or a 1D array of US transducers (e.g., linear, convex, phased, curved, or circular) with mechanical scanning [[19], [20], [21]]. While 2D array US transducers offer volume images directly, their increased number of channels and associated electronics make the overall system very complex and expensive [22,23]. On the other hand, 1D array US transducers with mechanical scanning require fewer channels and are less complex than 2D array systems, and still provide acceptable volumetric image quality [24]. However, most mechanical scanners for 3D PA imaging are heavy and bulky, and hence not suitable for handheld operation [25,26]. One commercial 3D mechanical handheld US probe, as known as a “wobbler”, has been considered as a solution for developing 3D PA handheld scanners [27]. However, it is ultimately not suitable for PAI because the transducer rotates in a convex manner about the center axis, and thus co-axiality between the laser incidence and image planes is not ensured, resulting in low signal-to-noise ratios (SNRs). Furthermore, PA signals propagate spherically in a 3D space, and the convex scanning pattern is not adequate for receiving them.

In this paper, we demonstrate a new type of 3D clinical handheld PA/US scanner that uses a 1D linear array US transducer and a mechanical scanner. The handheld scanner is compact (100 × 80 × 100 mm3) and light enough (950 g) for handheld image formation, because it uses a Scotch yoke mechanism to convert rotation into linear reciprocating movement. To confirm the device's suitability for clinical environments, we conducted multi-wavelength PAI and successfully obtained PA images of various body parts such as the neck, wrist, thigh, and instep. In addition, we performed PA sO2 imaging on a foot, using a spectral unmixing technique. Based on the results, we expect that the newly developed 3D handheld PA/US scanning system can be greatly useful for diagnosis and post-operative examinations of various clinical conditions of interest such as melanomas, free flaps, kidney failures, and diabetic feet.

2. Materials and methods

2.1. 3D clinical handheld PA/US imaging system and scanner

The clinical 3D PA/US imaging system is represented in Fig. 1a. The system consists of the programmable PA/US imaging system [28] and the 3D clinical handheld PA/US scanner. The US system is composed of a US machine (EC-12R, Alpinion Medical Systems, Republic of Korea) with 64 receiving channels and a 128-element linear array US transducer (L3-12, Alpinion Medical Systems, Republic of Korea) with a center frequency of 8.5 MHz and a fractional bandwidth of 62 % at -6 dB. While it is helpful to use broadband transducer in the context of PA imaging, we used a conventional US platform and a commercialized linear array US transducer to develop a clinically-translatable PA/US scanner. The tunable pulsed laser system (Phocus Mobile, OPOTEK, USA) is used at a pulse repetition frequency (PRF) of 10 Hz, where the wavelengths can be tuned between every pulse by the fast tuning function, and the pulsed light beam is delivered through fiber bundles (TFO-VIS100SL46-2000-F, TAIHAN FIBEROPTICS, Republic of Korea), which is linearly distributed at the output. Fig. 1b is a drawing of the handheld scanner, containing a guide rod, a holder, an arm, a geared motor, a water tank, a membrane, connectors, a plastic cap, and a handle, respectively. The guide rod minimizes shaking and tilting of the parts linked to the motor when the holder moves. The holder integrates the US transducer and the fiber bundles at 15° to optimize light transmission. The arm is connected to the holder so that the holder is driven in the elevational direction with cosine component of the angular motion by the Scotch yoke mechanism, which is a reciprocating mechanism that converts rotational motion into linear motion. Thus, the trajectory of the end of the motor arm with respect to time directly follows a simple harmonic motion because of the rotational movement of the motor with a constant angular velocity. The motor (CRK523PBK-T30, INA ORIENTAL MOTOR, Japan) weighs 170 g, and produces a 0.5 N m torque that rotates the arm. The water in the tank efficiently transmits US signals, and the 3 cm thick tank body serves as a stand-off that allows direct laser illumination to the imaging plane of the transducer [29]. At the top and bottom of the water tank, two 0.2 mm thick polyvinyl chloride (PVC) membranes seal in the water tank and provide acoustic coupling with the imaged object. The connectors facilitate convenient attachment and detachment of the water tank. The plastic cap protects imaging volunteer's skin from any sharp edge of the scanner. The handle allows an operator to hold and position the scanner conveniently. The entire handheld scanner can be positioned in any direction without spilling the water. The handheld scanner measures 100 , 80 , and 100 mm along the x, y, and z axes, respectively. It weighs 950 g, and its field of view (FOV) is 38.4 ×40.0 mm2 along the x and y axes, respectively. Fig. 1c shows an operational procedure of an in vivo human imaging, which is also shown in Supplementary Video S1.

Fig. 1.

Fig. 1

(a) Photograph of the clinical 3D PA/US imaging system. (b) Schematic of the 3D handheld PA/US scanner. (c) Photograph of the in-vivo imaging (see Supplementary Video S1). PA, photoacoustic; US, ultrasound; PC, personal computer; TR, ultrasound transducer; and FB, fiber bundle.

Fig. 2 illustrates the PA/US image acquisition procedure and operating sequences. The scanner is positioned over a ROI and then moves in the elevational direction (Fig. 2a) to reconstruct a 3D PA/US image, following the operating sequences shown in Fig. 2b. After the US machine receives the trigger signal from the laser system, PA/US imaging begins. Because the number of channels is one-half of the number of US transducer elements, one B-mode PA/US image is acquired with two laser shots, and thus the PA imaging rate is 5 frames per second. At exactly the same time, when the scanner system receives the first trigger signal from the laser system, the scanner starts moving in the elevational direction, traversing the full scanning range yScan during TScan:

TScan = TImage × nFrame
yScan=R×1-cosθScan

Where TScan is the full scanning time, TImage is the imaging time per frame (0.2 s), nFrame is the number of scanned frames, R is the length of the arm, and θScan is the full scanning angle between the arm and origin. Tscan's time range can be from milliseconds to seconds, depending on nFrame. The scanning range is maximized when the arm moves from the origin to the maximum angle of 180 degrees. Because the movement of the scanner is managed by the Scotch yoke mechanism, the distance of each frame depends on the angle θ:

θ=θScan×nnFrame

Where n is the current frame index. When each B-mode PA/US image is obtained with TScan, TImage, θ Scan, the scanning index numbers of the previous frame ni and the subsequent frame nf, the scanner is moved by yImage (a one frame scanning range) according to the scanning angle of the previous frame θi and the subsequent frame θf as follows:

yImage=R×|(1-cosθi)-(1-cosθf)| (2)
=R×cosθScan×nfnFrame-cosθScan×ninFrame

Fig. 2.

Fig. 2

Schematics of (a) PA/US image acquisition procedure and (b) operating sequences. PA, photoacoustic; US, ultrasound; PA/USI, photoacoustic and ultrasound imaging; yImage, one frame scanning range; yScan, full scanning range; TScan, full scanning time; and TImage, imaging time.

To make the image dense, the scanning step size dy is adjusted through the step angle θs_i and θs_f as follows:

dy=R×cosθs_f-cosθs_i

For spectral unmixing of multi-wavelength PA images, the scanning step size dy is divided by the number of laser wavelengths to be used. As the number of wavelengths increases, the scanning step size, dy, becomes dense and the full scanning time TScan increases. To minimize motion artifacts while the transducer is being translated, we set dy (at most 0.63 mm) to be much less than the elevational beam width of the linear array US transducer (about 1 mm at the depth of 30 mm), which allows a single PA image to be safely acquired within the elevational resolution of the transducer. When the scanning is completed, the imaging probe automatically returns to the origin position (see Supplementary Video S2).

2.2. 3D PA/US image processing

3D PA/US image processing is performed by online and offline processes. The online processes include real-time reconstruction and data storage. After receiving signals from the US transducer, real-time B-mode PA/US imaging is performed using a reconstruction algorithm that is accelerated by a graphics processing unit (GPU), and at the same time PA raw data and US beam-formed image data are stored [28,30,31]. The offline processes are carried out in three steps: (1) PA beamforming, (2) B-mode image stacking, and (3) volume data interpolation. The PA raw data is normalized by the laser output power and reconstructed into beam-formed image data by Fourier-domain beamforming [[32], [33], [34]]. The PA/US beam-formed image data for each frame are stacked in the elevational direction to create PA/US volume image data. Because dy is dependent on cosθ, the volume image is relatively dense at the beginning and the end of the scan range, as shown in Fig. 2b. Until this step, the stacked images are distorted by the non-uniform scanning because the position of each B-mode frame is represented as a digital index instead of the actual physical position. To reshape the images regarding the actual positions, we applied interpolation on the stack of images into a uniform grid where the grid index directly represents the actual position. This process enables representation of the images in the actual scales without being affected by the non-uniform acquisition. To obtain a uniform volume image, interpolation is performed in the elevational direction, with a uniform scanning step size Δy=yScan/nFrame.

2.3. Performance evaluation via phantom experiments

The system’s performance was evaluated through phantom experiments (Fig. 3). The scanner was positioned on a phantom (μeff:0.9 cm−1) consisting of water, gelatin, and intralipid where μeff is the effective attenuation coefficient, and the scanner was scanned along the elevation direction (i.e., y direction) (Fig. 3a) [18]. To identify the SNRs ​​along the depth, six black threads having a thickness of 90 μm were positioned as light absorbers at 5 mm intervals in the lateral and axial directions [9]. US gel (Ecosonic, SANIPIA, Republic of Korea) was applied to the scanner and phantom for acoustic coupling. The laser wavelength was 797 nm, and the pulse energy at the surface of the phantom was 7.5 mJ/cm2, which was below the American National Standards Institute (ANSI) safety limit of 31.3 mJ/cm2 at this wavelength. The yScan, FOV, and TScan, were 31.4 mm, 31.4 × 38.4 mm2, and 11.4 s, respectively.

Fig. 3.

Fig. 3

3D PA/US imaging of a tissue-mimicking phantom. (a) Photograph of the experimental setup. (b) Overlaid B-mode PA/US image. (c) Volume rendered PA/US image (see Supplementary Video S3). (d) Quantified SNRs of black threads at various depths. (e) Quantified axial thickness of threads at different depths, measured by FWHM. PA, photoacoustic; US, ultrasound; SNR, signal-to-noise ratio; and FWHM, full width at half maximum.

2.4. In vivo 3D PA/US imaging of various human body parts

The suitability of the system for clinical translation was evaluated through in vivo human imaging experiments by imaging various body parts of four healthy volunteers, such as the neck, wrist, and thigh. The experiments were conducted according to protocols of the Institutional Review Board of Pohang University of Science and Technology (POSTECH). For acoustic coupling, US gel was used at the top and bottom of the water tank and examined body parts. We used the laser wavelength of 850 nm and the pulse energy of 6.1 mJ/cm2 at the skin, which was less than the ANSI safety limit of 39.9 mJ/cm2 at this wavelength. To prevent eye damage from the laser exposure, the volunteers and examiner wore laser safety goggles. The yScan, FOV, and TScan were 40.0 mm, 40.0 × 38.4 mm2, and 20.0 s, respectively. To enhance the visualization of vessels, we compensated the PA raw data using time gain compensation with an exponential function. After then, we performed depthwise sectioning of the PA volume and removed underlying noise signals by thresholding relatively to the maximum amplitude of each section. Further, we differentiated vessels located at shallow depths and deep depths using the MAP method. For shallow depth PA MAP images, we used 3.0–6.8, 2.0–6.8, and 3.6–9.4 mm depth data from the skin for the neck, wrist, and thigh, respectively. For deep depth PA MAP images, we used 6.8–11.6, 6.8–10.0, and 9.4–11.6 mm depth data from the skin for the neck, wrist, and thigh, respectively. For validating the PA vascular images, we obtained US Doppler images of the corresponding body parts using the same US transducer. Because our scanner holds the US transducer orthogonally to the skin, it is not appropriate for Doppler imaging which needs manual or electronic tilting of the transducer for better sensitivity. Thus, we detached the transducer from the scanner and obtained the Doppler images from the similar position in the scanned region.

2.5. Multi-wavelength PAI of a human instep

Multi-wavelength PAI was conducted using wavelengths of 775, 797, and 825 nm, which were chosen to be as evenly distributed as possible with respect to the isosbestic point (797 nm) of oxy-hemoglobin and deoxy-hemoglobin. While using more wavelengths would enhance the accuracy of spectral unmixing, only three wavelengths were selected to perform fast imaging acquisitions in clinical environments. We imaged the instep of a volunteer’s foot, by placing the scanner carefully on the instep to avoid any unwanted signals from hair or abnormal skin and applied US gel for acoustic coupling. Based on the fluence compensation method in [4], we estimated the light fluence by the mean background intensity at each depth in each reconstructed PA image, and normalized each PA image accordingly. To apply the method in our results, we first detected the contour of the skin using the 3D PHOVIS software [35], and segmented the depth at each B-mode PA image relatively to the detected skin position. To identify spectral trends at each wavelength, PA amplitude analysis of imaged areas #1, #2, #3 and #4 shown in Fig. 5b was performed. Each PA amplitude was normalized by calculating mean values of the imaged areas at each wavelength and dividing the values by the maximum mean value of them. Further, we applied the spectral unmixing on the PA volume data, and used maximum amplitude projection to map the corresponding relative sO2 values to the instep's vessel image. PA sO2 estimation method is described in Supplementary Note S1. The yScan, FOV, and TScan, were 31.4 mm, 31.4 × 38.4 mm2, and 34.2 s, respectively.

Fig. 5.

Fig. 5

Multi-wavelength PA imaging of a human instep. (a) Photograph of the imaged human instep. (b) PA MAP images acquired at wavelengths of 775, 797, and 825 nm. (c) Normalized PA amplitudes of imaged areas represented by #1, #2, #3, and #4. (d) PA sO2 image. PA, photoacoustic; MAP, maximum amplitude projections; sO2, oxygen saturation.

3. Results and discussion

3.1. Performance evaluation via phantom experiments

Fig. 3b and c show the overlaid B-mode PA/US image and volume rendered PA/US image (Supplementary Video S3) of the phantom, respectively, where the PA images are in a pseudo color and the US images in gray. All six black threads, located at depths up to 30 mm, are well visible, and US reverberation artifacts are also identifiable at the bottom of the images. PA and US thread images may look different in the axial and lateral direction, because log compression parameters are separately adjusted for each modality. Fig. 3d shows the linear decrease (10.1 dB / cm) in SNR (dB) along the depth direction, from which the 1/e decay penetration depth is calculated as 1.98 cm. Fig. 3e shows the PA axial profiles of the six threads at the different depths, with the calculated full width at half maximum (FWHM) values. The axial profiles are almost identical with the depth, and the FWHMs are 191 ± 8 μm. PA lateral, US axial and US lateral profiles are represented in Supplementary Fig. S1.

3.2. In vivo 3D PA/US imaging of various human body parts

The photograph in Fig. 4a shows imaged regions of the human body (e.g., neck, wrist, and thigh), with the imaged areas indicated by red dashed rectangles. Fig. 4b and c display the corresponding PA maximum amplitude projection (MAP) images, which show blood vessels located at shallow depths and deep depths (e.g., common carotid artery (CCA), internal jugular vein (IJV), radial artery (RA) and great saphenous vein (GSV)) of the neck, wrist, and thigh. The overlaid PA/US images are shown in Fig. 4d, cut along the white dashed lines in Fig. 4c. The US image acquired from the neck area shows the sternocleidomastoid muscle (SCM) and the CCA and IJV, which are located below the SCM, and the B-mode PA image shows the upper boundaries of the CCA and IJV. The RA is displayed in the both US and PA images. The GSV is shown in the thigh image, and it is surrounded by the saphenous fascia (SF) and muscular fascia (MF), which are represented in the US image, and the upper boundary of the GSV is shown in the PA image. We can confirm that the PA/US images of the CCA, IJV, RA, and GSV match well with the US color Doppler images (Fig. 4e). While US Doppler can better show the major vessels in cross-sectional views, the advantage of PA imaging comes from providing 3D PA vascular images using a linear array transducer and a mechanical scanner, which is difficult to obtain with the US Doppler imaging due to the scanner motion. As can be seen in Fig. 4b and c, PA images can well visualize the major vessels in a 3D view and also the small vessels which were not visible from the US Doppler images. The maximum PA imaging depths in the neck, wrist, and thigh were 9.8, 9.7, and 14.2 mm, respectively, found at the CCA, RA, and GSV, and its distributions in 3D space are shown in Fig. 4f, which are PA MAP depth encoded-images. Imaging depths diverse at each imaged part because of the different organization of body components, such as lipids, muscles, and melanin, that affects the speed of sound and optical/acoustic attenuation. Heterogeneous distributions of these body components affect the speed-of-sound distribution, which could degrade the beamforming process which typically assumes a constant speed of sound, and the acoustic/optical attenuation of the components might affect the detectable depth of the PA signals. Further, the presence of bone structures can block the acoustic wave propagation and light penetration. For instance, in our results in Fig. 4, the thigh had the best penetration depth of 14.2 mm because it consisted of mostly muscles, and it contained less melanin component on the skin. The wrist had the worst penetration depth of 9.7 mm because of the presence of bones such as radius and ulna [36]. The neck had relatively more lipid component and more melanin on the skin, which would have resulted in the penetration depth of 9.8 mm. Conservatively, we used thick PVC membranes to seal the system's water tank and prevent water leakage during the in vivo experiments, but acoustic attenuation in this material is relatively high. Low-density polyethylene (LDPE) may be a good alternative to increase the imaging depth, because its optical and acoustic attenuation is relatively small. In addition, we used 6.5 times less energy than the ANSI limit. We are using the maximum energy from the laser (about 120 mJ), which is delivered by fiber bundle with an efficiency of about 70 %. Then, the end of the fiber bundle is spread in a linear shape to make an illumination area of about 10 cm2. Furthermore, the PVC membrane layers attenuate the light delivery to the skin. As a result, our light fluence to the skin is far smaller than the limit value. As our next step, we are planning to enhance the light delivery by optimizing the beam shape and using a more transparent material for the membrane to enhance the penetration depth. Other than that, we used a linear array transducer having relatively high center frequency (∼ 8 MHz), thus using imaging probes with a lower frequency might increase the imaging depth. Fig. 4h shows volume-rendered PA images, in which blood vessels such as CCA, IJV, RA and GSV are clearly identified, as are vessels located at shallow depths (Supplementary Videos S4, S5, and S6) [35]. To improve the image quality, using the synthetic aperture focusing technique (SAFT) with a slit might improve the elevational resolution [37,38]. Motion artifacts can occur mainly due to the low frame rate, especially in the areas near the main body having many pulsational motions. To the best of our knowledge, there are two techniques to overcome this limitation. One is to use an electrocardiogram (ECG) sensor to compensate motion artifacts induced from heartbeats. By using ECG sensor, laser and non-heartbeat zone can be synchronized to remove motion artifacts. The other is to use cross-correlation methods between each image frame for compensating distance errors. As the next steps, we will develop algorithms using cross-correlation methods to reduce motion artifacts. The in vivo imaging results imply a versatile clinical translation of our handheld scanner. One promising clinical application is monitoring free flap transplantation, which embeds whole tissues that contain blood vessels in other areas with poor blood circulation. The most important aspect of this operation is the anastomosis of the blood vessels, so it is important to know the exact extent of the defect site [39]. The scanner can be conveniently used to image a variety of body parts and can view blood vessel distribution in 3D space, which can be useful for pre-operative and post-operative monitoring of the free flap.

Fig. 4.

Fig. 4

In vivo 3D PA/US imaging at various parts of a human body. (a) Photograph of the human imaging parts. (b) PA MAP images (shallow). (c) PA MAP images (deep). (d) Overlaid B-mode PA/US images. (e) US Color Doppler images. (f) PA MAP Depth-encoded images. (g) Volume rendered PA images (see Supplementary Video S4, S5, S6). PA, photoacoustic; US, ultrasound; and MAP, maximum amplitude projections; S, skin; SCM, sternocleidomastoid muscle; CCA, common carotid artery; IJV, internal jugular vein; RA, radial artery; SF, saphenous fascia; GSV, great saphenous vein, MF, muscular fascia.

3.3. Multi-wavelength PA imaging of the human instep

The photograph of the human instep is shown in Fig. 5a, where the red dashed rectangle represents the imaged area. Fig. 5b shows PA MAP images for 775, 797, and 825 nm, respectively. Fig. 5c shows that the normalized PA amplitudes of imaged areas #1, #2, #3, and #4 in Fig. 5b. As the wavelength increases, the normalized PA amplitudes in the imaged areas #1 and #2 increase, while those in #3 and #4 decrease. Therefore, the blood vessels indicated by the green arrows can be regarded as containing relatively more oxy-hemoglobin than deoxy-hemoglobin, and the vessels indicated by the yellow arrows can be considered the opposite. These spectral trends obviously follow the molar extinction coefficients of oxy- and deoxy-hemoglobin at those wavelengths. The depths of the vessels can be confirmed via Supplementary Fig. S2, which shows PA MAP images in the XZ and YZ planes. Fig. 5d shows a multi-wavelength PA sO2 image, in which relatively higher and lower oxygenated blood vessels, indicated by the green and yellow arrows, are shown in red and blue. This sO2 information provided by the 3D PA/US imaging system can be useful in diagnosing various types of malignant cancer, because a malignant tumor is characterized by hypoxia, and 3D imaging can also be advantageous in distinguishing the margins of cancer areas. The scanner system can be used to further distinguish endogenous contrast agents such as melanin, which can be particularly beneficial in the diagnosis of the melanomas [40].

4. Conclusion

We presented a 3D clinical handheld PA/US scanner that uses a Scotch yoke mechanism, and we investigated its potential clinical use by performing in-vivo human imaging experiments. Acceptable image qualities were confirmed to the depth of 14.2 mm, and the convenience of the scanner was also proved by imaging various human body parts (e.g., the neck, wrist, thigh, and instep) that were difficult to image with previous imaging systems. Spectral unmixing was applied to provide sO2 information. Although technical developments such as membrane optimization and the use of a synthetic aperture with a slit are planned as next steps for better PAI, we believe that the currently implemented scanner can be applied to image such conditions of clinical interest as the free flaps, kidney failures, and diabetic feet. The scanner may be particularly advantageous for cancer imaging, where sO2 information is important. Other types of US transducers (e.g., convex or phased) can be also combined with the scanner to greatly enhance its clinical translatability.

Declaration of Competing Interest

Chulhong Kim has financial interests in OPTICHO, however, did not support this research.

Acknowledgments

This research was supported in part by the National Research Foundation of the Republic of Korea (NRF) grant (No. NRF-2019R1A2C2006269) funded by the MSIT (Ministry of Science and ICT), Republic of Korea Government, in part by the Korea institute for Advancement of Technology (KIAT) grant (P0008763, The Competency Development Program for Industry Specialist) funded by the MOTIE (Ministry of Trade, Industry and Energy), Republic of Korea Government.

Biographies

graphic file with name fx1.jpg

Changyeop Lee is a Ph.D. candidate in Mechanical Engineering at POSTECH. He completed his B.S. in Mechanical Engineering at Ajou University. His current research interests include 3D photoacoustic imaging systems and their clinical applications.

graphic file with name fx2.jpg

Wonseok Choi is a Ph.D. candidate in Electrical Engineering at POSTECH. He completed his B.S. in Electrical Engineering at POSTECH in 2012. His research interests include biomedical photoacoustic and ultrasound imaging techniques, therapeutic ultrasound, and their clinical applications.

graphic file with name fx3.jpg

Dr. Jeesu Kim received his B.Sc., M.Sc., and Ph.D. degrees from the Department of Electrical Engineering at Pohang University of Science and Technology (POSTECH). He is working as a post-doctoral researcher in the Department of Creative IT Engineering at POSTECH. His research interests focus on the development of non-ionizing and non-invasive novel biomedical imaging techniques, including photoacoustic imaging and ultrasound imaging, and their applications to clinical studies.

graphic file with name fx4.jpg

Dr. Chulhong Kim earned his Ph.D. degree and did postdoctoral training at Washington University in St. Louis, Missouri. He currently holds the Mueunjae Chaired Professorship and is an Associate Professor of Creative IT Engineering at Pohang University of Science and Technology in the Republic of Korea. He has received a 2017 IEEE EMBS Early Career Achievement Award, the 2017 Korean Academy of Science and Technology Young Scientist Award, the 2016 Nightingale Award from IFMBE, and the 2017 KOSOMBE Young Investigator Award for “Contributions to multi-scale photoacoustic imaging, ranging from super-resolution atomic force photoactivated microscopy for research to systems for clinical applications.

Footnotes

Appendix A

Supplementary material related to this article can be found, in the online version, at doi:https://doi.org/10.1016/j.pacs.2020.100173.

Contributor Information

Changyeop Lee, Email: ckdduq0801@postech.ac.kr.

Wonseok Choi, Email: zoids1215@postech.ac.kr.

Jeesu Kim, Email: hybridjs@postech.ac.kr.

Chulhong Kim, Email: chulhong@postech.edu.

Appendix A. Supplementary data

The following are Supplementary data to this article:

mmc1.docx (369.7KB, docx)
Download video file (7.2MB, mp4)
Download video file (4.2MB, mp4)
Download video file (1.9MB, mp4)
Download video file (2.1MB, mp4)
Download video file (2.1MB, mp4)
Download video file (2.1MB, mp4)

References

  • 1.Steinberg I., Huland D.M., Vermesh O. Photoacoustic clinical imaging. Photoacoustics. 2019 doi: 10.1016/j.pacs.2019.05.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Taruttis A., Ntziachristos V. Advances in real-time multispectral optoacoustic imaging and its applications. Nat. Photonics. 2015;9(4):219. [Google Scholar]
  • 3.Wang L.V., Hu S. Photoacoustic tomography: in vivo imaging from organelles to organs. Science. 2012;335(6075):1458–1462. doi: 10.1126/science.1216210. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Oraevsky A., Clingman B., Zalev J. Clinical optoacoustic imaging combined with ultrasound for coregistered functional and anatomical mapping of breast tumors. Photoacoustics. 2018;12:30–45. doi: 10.1016/j.pacs.2018.08.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Choi W., Park E.Y., Jeon S. Clinical photoacoustic imaging platforms. Biomed. Eng. Lett. 2018;8(2):139–155. doi: 10.1007/s13534-018-0062-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Shiina T., Toi M., Yagi T. Development and clinical translation of photoacoustic mammography. Biomed. Eng. Lett. 2018;8(2):157–165. doi: 10.1007/s13534-018-0070-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Manohar S., Dantuma M. Current and future trends in photoacoustic breast imaging. Photoacoustics. 2019 doi: 10.1016/j.pacs.2019.04.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kim C., Erpelding T.N., Jankovic L. Deeply penetrating in vivo photoacoustic imaging using a clinical ultrasound array system. Biomed. Opt. Express. 2010;1(1):278–284. doi: 10.1364/BOE.1.000278. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Erpelding T.N., Kim C., Pramanik M. Sentinel lymph nodes in the rat: noninvasive photoacoustic and US imaging with a clinical US system. Radiology. 2010;256(1):102–110. doi: 10.1148/radiol.10091772. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Jeon M., Song W., Huynh E. Methylene blue microbubbles as a model dual-modality contrast agent for ultrasound and activatable photoacoustic imaging. J. Biomed. Opt. 2014;19(1) doi: 10.1117/1.JBO.19.1.016005. [DOI] [PubMed] [Google Scholar]
  • 11.Park S., Jang J., Kim J. Real-time triple-modal photoacoustic, ultrasound, and magnetic resonance fusion imaging of humans. IEEE Trans. Med. Imaging. 2017;36(9):1912–1921. doi: 10.1109/TMI.2017.2696038. [DOI] [PubMed] [Google Scholar]
  • 12.Kim J., Choi W., Park E.-Y. Real-time photoacoustic thermometry combined with clinical ultrasound imaging and high intensity focused ultrasound. IEEE Trans. Biomed. Eng. 2019 doi: 10.1109/TBME.2019.2904087. [DOI] [PubMed] [Google Scholar]
  • 13.Akers W.J., Edwards W.B., Kim C. Multimodal sentinel lymph node mapping with single-photon emission computed tomography (SPECT)/computed tomography (CT) and photoacoustic tomography. Transl. Res. 2012;159(3):175–181. doi: 10.1016/j.trsl.2011.09.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Zhang Y., Jeon M., Rich L.J. Non-invasive multimodal functional imaging of the intestine with frozen micellar naphthalocyanines. Nat. Nanotechnol. 2014;9(8):631. doi: 10.1038/nnano.2014.130. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Xia J., Kim C., Lovell J.F. Opportunities for photoacoustic-guided drug delivery. Curr. Drug Targets. 2015;16(6):571–581. doi: 10.2174/1389450116666150707100328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Kim C., Favazza C., Wang L.V. In vivo photoacoustic tomography of chemicals: high-resolution functional and molecular optical imaging at new depths. Chem. Rev. 2010;110(5):2756–2782. doi: 10.1021/cr900266s. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Upputuri P.K., Pramanik M. Recent advances toward preclinical and clinical translation of photoacoustic tomography: a review. J. Biomed. Opt. 2016;22(4) doi: 10.1117/1.JBO.22.4.041006. [DOI] [PubMed] [Google Scholar]
  • 18.Fenster A., Downey D.B. 3-D ultrasound imaging: a review. Ieee Eng. Med. Biol. Mag. 1996;15(6):41–51. [Google Scholar]
  • 19.Wang Y., Guo Z., Wang L.V. In vivo three-dimensional photoacoustic imaging based on a clinical matrix array ultrasound probe. J. Biomed. Opt. 2012;17(6) doi: 10.1117/1.JBO.17.6.061208. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Wang D., Wang Y., Wang W. Deep tissue photoacoustic computed tomography with a fast and compact laser system. Biomed. Opt. Express. 2017;8(1):112–123. doi: 10.1364/BOE.8.000112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Matsumoto Y., Asao Y., Sekiguchi H. Visualising peripheral arterioles and venules through high-resolution and large-area photoacoustic imaging. Sci. Rep. 2018;8(1):14930. doi: 10.1038/s41598-018-33255-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Nagae K., Asao Y., Sudo Y. Real-time 3D photoacoustic visualization system with a wide field of view for imaging human limbs. F1000Research. 2018;7 doi: 10.12688/f1000research.16743.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Dean-Ben X., Fehm T.F., Razansky D. Universal hand-held three-dimensional optoacoustic imaging probe for deep tissue human angiography and functional preclinical studies in real time. J. Vis. Exp. 2014;(93) doi: 10.3791/51864. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Karadayi K., Managuli R., Kim Y. Three-dimensional ultrasound: from acquisition to visualization and from algorithms to systems. IEEE Rev. Biomed. Eng. 2009;2:23–39. [Google Scholar]
  • 25.Wang Y., Zhan Y., Harris L.M. A portable three-dimensional photoacoustic tomography system for imaging of chronic foot ulcers. Quant. Imaging Med. Surg. 2019;9(5):799–806. doi: 10.21037/qims.2019.05.02. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Upputuri P.K., Pramanik M. Fast photoacoustic imaging systems using pulsed laser diodes: a review. Biomed. Eng. Lett. 2018;8(2):167–181. doi: 10.1007/s13534-018-0060-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Yang W., Ingle A., Varghese T. Comparison of three dimensional strain volume reconstructions using SOUPR and wobbler based acquisitions: a phantom study. Med. Phys. 2016;43(4):1615–1626. doi: 10.1118/1.4942814. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Kim J., Park S., Jung Y. Programmable real-time clinical photoacoustic and ultrasound imaging system. Sci. Rep. 2016;6:35137. doi: 10.1038/srep35137. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Schellenberg M.W., Hunt H.K. Hand-held optoacoustic imaging: a review. Photoacoustics. 2018;11:14–27. doi: 10.1016/j.pacs.2018.07.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Diot G., Metz S., Noske A. Multispectral optoacoustic tomography (MSOT) of human breast Cancer. Clin. Cancer Res. 2017;23(22):6912–6922. doi: 10.1158/1078-0432.CCR-16-3200. [DOI] [PubMed] [Google Scholar]
  • 31.Kim J., Park E., Park B. Towards clinical photoacoustic and ultrasound imaging: probe improvement and real-time graphical user interface. Exp. Biol. Med. 2019 doi: 10.1177/1535370219889968. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Köstli K.P., Frenz M., Bebie H. Temporal backward projection of optoacoustic pressure transients using Fourier transform methods. Phys. Med. Biol. 2001;46(7):1863. doi: 10.1088/0031-9155/46/7/309. [DOI] [PubMed] [Google Scholar]
  • 33.Rosenthal A., Ntziachristos V., Razansky D. Acoustic inversion in optoacoustic tomography: a review. Curr. Med. Imaging Rev. 2013;9(4):318–336. doi: 10.2174/15734056113096660006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Jeon S., Park E.-Y., Choi W. Real-time delay-multiply-and-sum beamforming with coherence factor for in vivo clinical photoacoustic imaging of humans. Photoacoustics. 2019;15 doi: 10.1016/j.pacs.2019.100136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Cho S., Baik J., Managuli R. 3D PHOVIS: 3D Photoacoustic Visualization Studio. Photoacoustics. 2020 doi: 10.1016/j.pacs.2020.100168. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Matsumoto Y., Asao Y., Yoshikawa A. Label-free photoacoustic imaging of human palmar vessels: a structural morphological analysis. Sci. Rep. 2018;8(1):786. doi: 10.1038/s41598-018-19161-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Wang Y., Wang D., Zhang Y. Slit-enabled linear-array photoacoustic tomography with near isotropic spatial resolution in three dimensions. Opt. Lett. 2016;41(1):127–130. doi: 10.1364/OL.41.000127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Jeon S., Park J., Managuli R. A novel 2-D synthetic aperture focusing technique for acoustic-resolution photoacoustic microscopy. IEEE Trans. Med. Imaging. 2019;38(1):250–260. doi: 10.1109/TMI.2018.2861400. [DOI] [PubMed] [Google Scholar]
  • 39.Lee K., Lee D., Roh S. Reconstruction of hand. J. Korean Med. Assoc. 2016;59(2):127–135. [Google Scholar]
  • 40.Kim J., Kim Y., Park B. Multispectral ex vivo photoacoustic imaging of cutaneous melanoma for better selection of the excision margin. Br. J. Dermatol. 2018;179(3):780–782. doi: 10.1111/bjd.16677. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

mmc1.docx (369.7KB, docx)
Download video file (7.2MB, mp4)
Download video file (4.2MB, mp4)
Download video file (1.9MB, mp4)
Download video file (2.1MB, mp4)
Download video file (2.1MB, mp4)
Download video file (2.1MB, mp4)

Articles from Photoacoustics are provided here courtesy of Elsevier

RESOURCES