Abstract
Acoustic-resolution photoacoustic microscopy (AR-PAM) is a promising imaging modality that renders images with ultrasound resolution and extends the imaging depth beyond the optical ballistic regime. To achieve a high lateral resolution, a large numerical aperture (NA) of a focused transducer is usually applied for AR-PAM. However, AR-PAM fails to hold its performance in the out-of-focus region. The lateral resolution and signal-to-noise ratio (SNR) degrade substantially, thereby leading to a significantly deteriorated image quality outside the focal area. Based on the concept of the synthetic-aperture focusing technique (SAFT), various strategies have been developed to address this challenge. These include 1D-SAFT, 2D-SAFT, adaptive-SAFT, spatial impulse response (SIR)-based schemes, and delay-multiply-and-sum (DMAS) strategies. These techniques have shown progress in achieving depth-independent lateral resolution, while several challenges remain. This review aims to introduce these developments in SAFT-based approaches, highlight their fundamental mechanisms, underline the advantages and limitations of each approach, and discuss the outlook of the remaining challenges for future advances.
Keywords: AR-PAM, Lateral resolution, SNR, SAFT, Coherent factor, DAS, DMAS, Artifacts, Side lobes
1. Introduction
The photoacoustic (PA) technique [1], [2] is a hybrid imaging method that incorporates optical contrast and ultrasound wave detection. In PA imaging, light absorptive samples such as biological tissues are irradiated by a short-pulse laser beam, during which the absorbed photons are partially converted to heat, thereby inducing thermo-elastic expansion [3], [4]. This leads to a pressure rise that propagates in the form of wideband frequency ultrasonic waves that can be detected by an ultrasound transducer [5], [6], [7], [8]. Given that ultrasonic scattering is two to three orders of magnitude less than optical scattering, the transformation from photon to acoustic energy in the PA technique breaks the optical diffusion limit and significantly extends the imaging depth to ranges that are inaccessible by conventional optical imaging [9], [10], [11], [12]. Meanwhile, PA imaging preserves the rich optical absorption contrast, which is derived from photon absorption, as opposed to conventional ultrasound imaging techniques, where the imaging contrast is a result of mechanical properties. These features, i.e., a large penetration depth and rich optical contrast [13], [14], [15], [16], [17], [18], make PA imaging a revolutionary modality suitable for wide applications in biomedical research [19], [20], [21], [22], [23], [24], [25].
Photoacoustic microscopy (PAM) is one major implementation of PA imaging [26], [27]. There are two sub-categories of PAM: optical-resolution PAM (OR-PAM) and acoustic-resolution PAM (AR-PAM). OR-PAM achieves optical lateral resolution (micron or even submicron scale) at superficial depths using tight optical focusing [28], [29], [30], [31], [32]. AR-PAM renders images with acoustic lateral resolution (tens to hundreds of microns) at a much greater penetration depth using wide-beam optical illumination but tight acoustic focusing [33], [34], [35], [36], [37], [38], [39], [40], [41], [42], [43], [44], [45]. This paper focuses on the second category (AR-PAM). In AR-PAM, the lateral resolution is determined by the acoustic numerical aperture, NA, and the PA center frequency, [2], as , where c denotes the speed of sound. To achieve high spatial resolution, a spherically focused transducer with a high center frequency, and a large numerical aperture, , is usually employed for signal detection. Unfortunately, the lateral resolution and imaging depth are in a trade-off relationship. High spatial resolution limits the acoustic focal zone, thereby leading to a significantly deteriorated image quality (i.e., lateral resolution and signal-to-noise ratio (SNR)) in the out-of-focus regions [26], [46]. Specifically, the PA signals of targets outside the focal zone are distorted and curved away from the focus, as depicted in Fig. 1. The distorted signals generally exhibit a concave shape above the focus and a convex shape below (Fig. 1a); the mechanisms for these phenomena are described below.
The two scenarios described above are illustrated in Fig. 1b and c, where the imaging target is represented by the red dot, and the ultrasound transducer (UST) scans horizontally from position 1 (P1) to position 3 (P3). In the case of concavely distorted PA signals (Fig. 1b), the PA waves generated by the target are depicted by the red solid lines. In PAM, the detected A-lines are combined next to each other to form a 2D image (B-scan). When the UST locates at P1, the acoustic waves (red solid lines) of the target are first detected by the right edge of the UST. PAM assumes that the target locates right below the UST. Thus, based on the time of flight, the signal of the target is projected as PA1. The distance between PA1 and the center of the UST detection surface (D1) equals the distance between the target and the right edge of the UST (D2). When the UST moves to P2, the acoustic wave arrives at the center of UST detection region first because (1) the transducer applied in AR-PAM is generally spherically focused, and (2) the target locates above the focus (the center point of a sphere). Based on the time of flight, the PA image of the target superposes the position of the target, i.e., PA2 reflects the actual position of the target. When the UST moves to P3, the acoustic waves are first detected by the left edge of the UST. Again, according to the time of flight, the signal of the target is projected as PA3. Connecting all of the PA signal projected positions during the UST scanning, a concave imaging shape of the target is achieved (dashed blue line in Fig. 1b). On the other hand, Fig. 1c demonstrates the case in which the target (red dot) locates below the focal plane. The hyperbola curves (dashed and solid black lines) indicate the acoustic detection regions of the UST at different depths. In contrast to the case where the target locates above the focus, the acoustic signals below the UST focus need to propagate through the middle region of the hyperbola for the PA signal to be detected by the UST, as imaging targets located outside the detection area generally cannot be imaged. When the UST locates at P1, the PA waves generated by the target first propagate through the middle of the hyperbola and then arrive at the detection surface of the UST. This takes more time than when the UST locates at P2, in which PA waves directly propagate through the detection region without a detour. For the former case (the UST at P1), the imaged signal of the target is projected as PA1, and the distance between PA1 and the focal point (D1) equals the distance between the target and the focal point (D2). For the latter case (the UST at P2), the imaged signal of the target superposes the position of the target, i.e., PA2 reflects the actual position of the target. When the UST moves to P3, similar to when it is located at P1, the pressure rise generated by the target (red circle) needs to travel through the middle region of the hyperbola and then arrive at the detection surface of the UST. Again, according to the time of flight, the imaged signal of the target is projected as PA3. By connecting all of the detected signals along the scanning direction of the UST, a convex shape of the distorted signals is formed (dashed blue line in Fig. 1c).
Consequently, AR-PAM exhibits inherent artifacts with distorted shapes (concave or convex) as a function of the imaging depth (Fig. 1d). Efforts have been made in the past to address this challenge, including model-based approaches [47], [48], [49], [50], time reversal strategies [51], [52], [53], [54], [55], and the synthetic aperture focusing technique (SAFT) [56], [57], [58], [59], [60], [61]. Compared with other schemes, SAFT is a far more commonly used approach to solve the blurring problem in the out-of-focus region. As a conventional post-processing method for image reconstruction, SAFT has been widely used in radar and ultrasound imaging [62], [63], [64], [65]. By employing the virtual detector concept [57], SAFT-based solutions can be extended to the focused transducer applications in AR-PAM. The SAFT-based approach along with coherent factor (CF) weighting has raised a lot of interest since their initial applications to AR-PAM by Liao et al. [56] and Li et al. [57]. By properly delaying and summing the signals received at adjacent scan lines (i.e., A-lines), SAFT improves the spatial resolution and SNR in the out-of-focus region. However, the effectiveness of SAFT is sensitive to the synthesis direction. Synthesizing in one dimension leads to anisotropy in the lateral resolution and SNR. To overcome this limitation, Deng et al. [66] developed a 2D-SAFT approach in which the synthesis of correlated signals was processed in both the x-z and y-z planes. Although a homogenous lateral resolution could be achieved, the resolution and SNR enhancement of the 2D-SAFT were worse than the best performance of 1D-SAFT when imaging vasculatures [66], owing to the mismatch between the spherical wavefront concept of the 2D algorithm and the cylindrical wavefronts of the targets [66], [67]. To address this challenge, an adaptive SAFT solution [68] that synthesizes signals perpendicular to an arbitrary direction of the targets has been proposed. Another approach, the frequency-domain SAFT [69], has also shown the potential to tackle such problems. Furthermore, to remove invalid signals involved in SAFT, the spatial impulse response (SIR) [70] of the transducer has been introduced to weigh the contributions of the SAFT at different depths.
Another research interest in SAFT-based strategies lies in the development of beamformers. Generally, the delay-and-sum (DAS)-based algorithm is employed for the synthesis owing to its simple implementation. However, DAS is a blind beamformer that treats all of the input A-lines equally, thereby resulting in a high level of noises and artifacts. To compensate for this limitation, DAS is usually applied in conjunction with CF weighting. Recent studies have shown that other approaches, such as the delay-multiply-and-sum algorithm (DMAS) [71] and double-state (DS) DMAS [72], can be applied for this purpose. With a correlation operation of the PA signals at each scanline, DMAS-based SAFT has shown an improved ability to reduce incoherent components such as noises and artifacts.
In brief, the above SAFT-based approaches have made progress in expanding the depth of focus and decreasing the dependence of lateral resolution and SNR on the imaging depth. However, these strategies have limitations and challenges. In this review, we introduce recent advances and outline their fundamental algorithms. The merits and limitations of different strategies are compared and analyzed. Furthermore, this review discusses the prospects and challenges of future advances in SAFT-based strategies for AR-PAM.
2. Basic principle of VD-based SAFT strategies
The schematic illustration of SAFT using the virtual detector (VD) concept is depicted in Fig. 2a, where the focus of the UST is usually considered as the VD. The red dots represent the imaging targets, and the blue dots are the artifacts of the targets, the connection of which forms the convex/concave shape of the distorted signals, as detailed illustrated by Fig. 1 in the Introduction. To restore the authentic target shape, the PA signals detected at adjacent scanning points are synthesized by considering their acoustic propagation delays, [56], [57]:
(1) |
On the left side of Eq. (1), (i, t) denotes the signal after SAFT processing, where i and t are the two coordinates (the scanning direction and the depth direction (propagation time ), respectively) of a two-dimensional space (B-scan). On the right side of Eq. (1), a delay and sum (DAS) process is performed for the raw data (RF, i.e., A-line). More specifically, an appropriate time delay, , is applied to the detected scanlines in the vicinity of the ith line, such that the correlated signals are adjusted to be in-phase, as illustrated in Fig. 2b. After this procedure, the adjacent A-lines can be correctly synthesized (Fig. 2b, right). In Eq. (1), defines the adjacent scanlines involved in the summation, the total number (N) of which is determined by the opening angle of the UST—a large NA of the focused transducer leads to a larger number of summed scanlines (as described in Section 3.4). The time delay, , is quantified based on the difference in the detection time. For scanline (i + 1) shown in Fig. 2a, the time delay () relative to scanline (i) is calculated as , where is the distance between the synthesized point (target location) and the focal point, is the distance between the detected signal of the target (artifact) and the focal point when the UST moves to position i + 1, and c is the speed of sound. The quantity represents the difference in the detection time between lines i and i + 1. Note that generally cannot be known a priori. Fortunately, is equal to , as shown in Fig. 2a. The latter () is the distance between the VD of scanline (i + 1) and the synthesized point, which can be expressed as , where is the moving step of the UST during scanning between adjacent A-lines. Therefore, for an arbitrary scanline , the time delay () applied to the received signal is given by:
(2) |
where sign is the signum function; > 0 if the target locates below the focus, < 0 if the target locates below the focus, and = 0 if the target locates at the focus; z and signify the depth of the synthesized point and focal point, respectively. With the implementation of (1), (2), a set of new A-lines are synthetically generated by combining the delayed versions of the detected A-lines (Fig. 2b). This reconstruction solution is also referred to as VD-based SAFT because the A-lines can be considered to be detected at the VD location [56], [57].
Note that SAFT is performed in a pixel-by-pixel manner, as illustrated in Fig. 3a. Thus, not only the main lobes (such as point 2) are processed with DAS, but also the artifacts (such as points 1 and 3). For the real target location (such as point 2), DAS synthesizes the signals of the target and neighboring points (i.e., the summation of PA1, PA2, and PA3). For artifacts (such as points 1 or 3), DAS synthesizes the artifact itself and adjacent noises (i.e., PA1+ noises, PA3+ noises). Generally, the image noises generated in PAM are randomly positive or negative. Therefore, the sum of noises becomes negligible when sufficient pixels are involved in the reconstruction, and the PA signals for artifacts (such as points 1 and 3) after SAFT are close to their original value (PA1 or PA3). In other words, the image after processing by DAS-based SAFT (Eq. (1)) still preserves the signals of artifacts (PA1 and PA3), as illustrated in Fig. 3b, although the main lobe signal (point 2) is greater than that of adjacent artifacts (points 1 and 3). The high level of noises and artifacts in DAS-based SAFT restricts its application. To compensate for this limitation, CF weighting is generally applied to the output of DAS to suppress the artifacts and noises [56], [57], [58], [59].
(3) |
The denominator of CF is the sum of the squares of the delayed versions of the A-lines (Fig. 2b, middle) used in SAFT, and the numerator is the square of the sum of the aforementioned A-lines. The CF value, which can be derived from Eq. (3), ranges from 0 to 1. To provide a straightforward understanding, Fig. 3(d and e) demonstrates the CF concept from a signal coherence perspective. It can be inferred from Fig. 3d and e and Eq. (3) that CF can be considered as the ratio of the coherent energy to the total energy of the delayed signals [73], [74]. To be more specific, CF reaches maximal (close to 1) when the delayed version of the signals is perfectly coherent (Fig. 3d) along the synthesis direction. In contrast, CF is minimized when the delayed signals are incoherent (Fig. 3e). The former case (CF close to 1) corresponds to the main lobes, where the adjacent delayed A-lines are highly coherent. The latter case (CF close to 0) corresponds to the side lobes and noises where the adjacent delayed A-lines are generally incoherent. By applying CF to the output of the SAFT, the CF-weighted SAFT is given by the following [56], [57], [58], [59], [66], [67]:
(4) |
CF weighting is applied to the SAFT on a pixel-by-pixel basis, after which the side lobes and noises in the SAFT are suppressed, whereas the main lobe signals of the SAFT are reserved, as illustrated in Fig. 3c. This further increases the beam quality [66], [67], [68], [70]. It should be noted that this section discusses the CF-weighed SAFT based on the DAS beamformer, which is a commonly used strategy in SAFT. In addition to the DAS beamformer, the DMAS-based beamforming technique [71], [72], which has recently been employed in AR-PAM, is discussed in Section 3.5.
3. SAFT-based strategies in AR-PAM
3.1. 1D-SAFT
1D-SAFT indicates that SAFT is applied in one direction—normally along the B-scan direction of the UST, such that the distorted signals generated in the B-scan images are synthesized. 1D-SAFT is a commonly used strategy because of its straightforward nature. The first demonstration of this implementation was presented in Ref. [56], where the effectiveness of SAFT in conjunction with CF was validated using phantom experiments (Fig. 4a). Compared with the original image (Fig. 4a, left), in which the image quality is poor, the SAFT image (Fig. 4a, middle) exhibits enhanced resolution and SNR but still with large side lobes and high background noises, the reason of which has been elaborated in Section 2 (Fig. 3a, b). With the application of CF weighting, these artifacts and noises are significantly suppressed (Fig. 4a, right). The above results (Fig. 4a) are similar to the results of carbon fiber imaging [57] (Fig. 4b), in which the lateral resolution of the original image decreases rapidly when the carbon fibers are further from the focus (Fig. 4b, left). Consequently, the post-processing (SAFT+CF) expands the focal region and reduces the dependence of image quality on the focal depth (Fig. 4b, right). It is noteworthy that the distorted shapes of the signals above and below the focus (concave and convex, respectively) obtained from the experiments are consistent with our theoretical analysis (Fig. 1).
The in vivo application of CF-weighted SAFT in AR-PAM was demonstrated for the first time in Ref. [57], where the vasculature of a rat’s scalp was placed out of focus. As expected, compared with the original image, the lateral resolution of the PA images (Fig. 4c, MAP and B-scan) improved considerably after SAFT+CF was performed. However, it should be noted that 1D-SAFT causes the anisotropy in the lateral resolution (between x and y directions) because only the vessels perpendicular to the synthetic direction (usually the scanning direction) can be well reconstructed. The inhomogeneous image quality will be showcased by Fig. 5b (middle, 1D-SAFT), in which the SNR and resolution of vessels along the x-axis are much worse than those of the vessels located along the y-axis. To summarize, the dependence on target orientation poses limitations for the use of 1D-SAFT in vivo, given that biological targets generally exhibit various directions that are not in a controllable manner.
3.2. 2D-SAFT
To diminish the anisotropy in the lateral resolution caused by 1D-SAFT, Deng et al. [66] developed a 2D-based SAFT scheme (Fig. 5a) in which the image construction was performed in both the x-z and y-z planes for each pixel in the raw data.
(5) |
where i, j, and t are the coordinates in three-dimensional space; i is the coordinate in the direction between adjacent A-lines, j is the coordinate in the direction between adjacent B-scans, and t is the coordinate in the depth direction (propagation time ). and are the time delays of the adjacent scanlines in the x-z and y-z planes, respectively. Eq. (5) is analogous to Eq. (1) (1D-SAFT) only with extra dimension included. Similarly, by considering another dimension, the CF weighting is given as
(6) |
The CF value also ranges from 0 to 1. With the combination of CF, the output of the 2D-SAFT is expressed as . Note that all of the following 2D-SAFT image results refer to the outcomes of . To validate the performance of the 2D-SAFT strategy, PA imaging of two carbon fibers was conducted at various imaging depths (Fig. 5b). In the original image (Fig. 5b, left column), the signal blurs considerably when the carbon fibers are out of focus. The 1D-SAFT images (Fig. 5b, middle column) show enhanced image quality along the y-axis direction outside the focal region, but no obvious improvement is observed along the x-axis because SAFT is conducted in the x-z plane. Note that the fibers along the x-axis outside the focal region seem to be diminished in the 1D-SAFT. This is because the improvement in the SNR along the y-axis is much greater than that along the x-axis. With 2D-SAFT adopted, the images become more homogenous compared with the 1D-SAFT (Fig. 5b, right column). The isotropic resolution is particularly essential when imaging point-like targets. For example, as shown in Fig. 5c, the dust of the carbon fiber (indicated by the white arrow), which can be approximated as a spherical target, displays a distorted shape (Fig. 5c, middle) in the 1D-SAFT image owing to the anisotropic lateral resolution of 1D-SAFT. As opposed, 2D-SAFT overcomes this limitation and retains the original shape of the target (Fig. 5c, right). However, 2D-SAFT has its limitation—the lateral resolution in 2D-SAFT is worse than the best performance of 1D-SAFT when the synthesis is performed perpendicular to the carbon located outside the focal region, as shown in Fig. 5b. This owes to the mismatch between the spherical wavefront concept of the 2D algorithm and the cylindrical wavefront of the carbon fiber, as described below.
To elaborate on the origin of this mechanism, Fig. 6 demonstrates the difference that 1D- and 2D-SAFT made on the lateral resolution of a cylindrical target placed perpendicular to the x-z plane. In Fig. 6a, Point B indicates the authentic signal of the cylindrical target’s edge, and A and C are the artifacts of B, generated as a result of the inherent signal blurring in AR-PAM. For the 1D-SAFT, SAFT was performed in the x-z plane on a pixel-by-pixel basis, and the artifacts of A and C were significantly suppressed after the application of SAFT+CF (see Figs. 3c, e and 6b, first row). For 2D-SAFT, SAFT+CW was performed in both the x-z and y-z planes for each pixel. Although the reconstruction results of the side lobes (such as A and C) were low in the x-z plane owing to the incoherence of the delayed signals (see Fig. 3c and e), the signals of the side lobes (A and C) increase after the signal summing in the y-z plane. This is due to the existence of a number of correlated signals of side lobes in the y-z plane generated by tubular-structured targets with a cylindrical wavefront. To be more specific, the distorted signals, indicated by the blue dashed line (including A and C) in Fig. 6a, exhibits an extension in the y-axis direction because the tubular target is placed along the y-axis. Therefore, the synthesis of artifacts (such as point A) in the y-z plane, as shown in Fig. 6c, leads to a constructive summation because the adjacent signals are highly in-phase. As a result, 2D-SAFT manifests points A and C with much higher values than those in 1D-SAFT, as depicted in Fig. 6b. This results in a wider full width at half maximum of the point spread function (PSF) for the target at point B in 2D-SAFT than in the 1D solution. The above description explains why 2D-SAFT shows a lower lateral image resolution along the B-scan direction than 1D-SAFT (Fig. 5b).
Essentially, the occurrence of low lateral resolution in 2D-SAFT originates from the inconsistent wavefront between the 2D-SAFT concept (spherical wavefront) and the tubular object (cylindrical wavefront). As opposed, 1D-SAFT algorithm implies a spherical wavefronts concept and matches with tubular object. To summarize, the proposed 2D-based strategy helps minimize the anisotropic resolution feature exhibited in the 1D solution, but demonstrates a lower resolution for cylindrical targets such as vascular architecture, as summarized in Table 1. Interestingly, when the tubular targets are arranged along the diagonal directions (Fig. 5c), the resolution enhancement for carbon fibers is approximately equivalent between 1D- and 2D-SAFT because the SAFT is not optimal for either case. For the former, the synthesis direction is not perpendicular to the target direction. For the latter, as discussed previously, the cylindrical wavefront of the targets does not match the spherical wavefront in the 2D-based solution.
Table 1.
Algorithm description | Dimensions of improved lateral resolution | Achievable resolution (a) |
Wave-shape concept | Processing speed | Suppress SNR at focus (b) | ||
---|---|---|---|---|---|---|---|
1D-SAFT+CF [56], [57] | Signals synthesized in one dimension (x-z or y-z) | 1 (Anisotropic lateral resolution) |
High (Comparable with resolution at focus) |
Cylindrical wave | Fast | Yes (if without SIR) | |
2D-SAFT+CF [66] | Signals synthesized in two dimensions (x-z and y-z) | 2 (Isotropic lateral resolution) |
Medium (Less than resolution at focus) |
Spherical wave | Yes (if without SIR) | ||
A-SAFT +CF [68] |
Tubular targets | Signals synthesized perpendicularly to each target’s orientation | 2 (Isotropic lateral resolution) |
High (Comparable with resolution at focus) |
Cylindrical wave | Medium | Yes (if without SIR) |
Point-like (or circular) target |
1(c) (Anisotropic lateral resolution) |
High (Comparable with resolution at focus) |
Yes (if without SIR) | ||||
SIR-based SAFT [70] | Transducer’s SIR used to weight contributions | Depends on its combination with 1D-, 2D-, or Adaptive-SAFT | No (Homogenous SNR achieved) | ||||
DMAS-SAFT [71] | Delay-multiply-and-sum used for beamforming | Depends on its implementation (in 1D-, 2D-, or Adaptive-SAFT manner) | Slow | Yes (if without SIR) |
(a)The best performance that could be achieved with each SAFT scheme. For 1D- and A-SAFT, the best resolution after reconstruction in the out-of-focus region is theoretically comparable with the original resolution at focus, thus achieving the depth-independent lateral resolution. 2D-SAFT shows a poorer performance compared to the 1D solution. (b) The SNR at the focal region is suppressed relative to the outside focal region if the SIR weight is not considered during the SAFT reconstruction. The weighing of SIR helps achieve a homogeneous SNR both far from and at the focus of the ultrasonic detector. (c) For point-like (or circular) targets, A-SAFT may lead to anisotropic lateral resolution, given that only one dimension synthesis is performed for a whole target’s shape when the target’s aspect ratio ≈ 1, as opposed to circular structured targets like vasculatures, where the synthesizing direction changes for each pixel based on the orientation of each branch of their network.
3.3. Adaptive-SAFT
In the complex in vivo environment, vascular architecture with different branches generally exhibits various directions. In such cases, 1D-SAFT leads to anisotropy in the lateral resolution. The 2D-based strategy helps minimize the resolution anisotropy, but provides a poorer performance in terms of resolution enhancement compared to the 1D solution. To address these challenges, the adaptive-SAFT (A-SAFT) [68] was developed, which allows SAFT to be automatically applied to each branch of the vascular network based on its orientation (Fig. 7a). Based on MAP images of the vasculature, a binary skeleton image (such as Fig. 7b) can be extracted, and the direction of each pixel is determined based on the linear fitting of the point and its consecutive surrounding points. According to the pre-determined directions, SAFT is performed in a plane perpendicular to the vessel branch (Fig. 7a),
(7) |
where , and ; denotes the angle between the x-z plane and the direction of the vessel branch. The notation [□] signifies the nearest integer in the direction of negative infinity. Eq. (7) is analogous to Eq. (1) (1D-SAFT)—the synthesis is only performed in one dimension, as opposed to 2D-SAFT (Eq. (5)) where the signal summing is conducted in two dimensions (two summation notations) for each pixel. The difference between 1D- and A-SAFT is that the former is conducted in the x-z plane; thus, only two coordinates (i, t) are required to localize the pixel for synthesis, as shown in Eq. (1). In contrast, A-SAFT is performed in a plane perpendicular to the direction of the cylindrical targets; therefore, three coordinates (i, j, t) are needed to index each pixel.
To estimate the efficiency of this strategy, in vivo PA imaging of a vascular network was performed using different SAFT strategies for comparison, including 1D-, 2D-, and A-SAFT schemes. These solutions were performed in conjunction with CF weighting. Compared to the original image (Fig. 8a), the above three SAFT solutions exhibited noticeable enhancements of the lateral resolution and SNR (Fig. 8b–d). For a detailed analysis of the image quality improvements, three typical vessel orientations were particularly examined, with the outcomes interpreted as follows. For the direction along the x-axis, such as the vessel marked with the green rectangle, the above three strategies exhibit different efficacies in improving the lateral resolution. The decreasing order of their performances is evaluated as follows: Adaptive > 2D > 1D. Note that the 1D solution exhibits the worst performance because the SAFT was not conducted in that direction. For the direction along the y-axis, such as the vessel marked with the blue rectangle, Adaptive ≈ 1D > 2D. The A-SAFT and 1D-SAFT schemes show equivalent efficacies because the SAFT is implemented approximately in the same plane for both solutions. For the diagonal direction, such as the vessel marked with the red rectangle, Adaptive > 2D ≈ 1D. The 2D and 1D schemes are inferior to A-SAFT because neither is suitable for this situation (owing to the tubular-structured targets and diagonal direction, respectively), as demonstrated in Section 3.2. In summary, compared to 1D- and 2D-SAFT, A-SAFT enables images of blood vessels with better lateral resolution and retains the isotropic lateral resolution and SNR, overcoming the limitations of both 1D- and 2D-SAFT, as has been summarized in Table 1. Thus, A-SAFT is well suited for processing microvasculature imaging in AR-PAM. However, it should be mentioned that A-SAFT is developed based on cylindrical wavefront targets, and thus it may lead to image distortion when reconstructing spherical wavefront objects, given that only one-dimensional synthesis is applied for each pixel in A-SAFT. Note that in a complex in vivo environment, spherical wavefronts may coexist with cylindrical wavefronts, such as tumors, which are usually accompanied by vascular architectures. Therefore, more efforts should be made in the future to explore new methods to account for the diversity of target wavefronts, such that SAFT can be robustly applied to complex in vivo environments.
3.4. SIR-based SAFT
Another consideration of the SAFT algorithm lies in the weighting of the signal contribution at different depths. In SAFT, PA signals at each point are summed with those at neighboring points, but the number of adjacent scanlines (N) involved in the summation changes considerably with the depth. More specifically, N is minimized when the synthesized point locates at the focus. In contrast, N is large when the synthesized point is far from the focus of the UST, given that the angular extent of the UST leads to a large set of contributions to the summation, as depicted in Fig. 9. If signal summing is performed for a fixed region (i.e., N is independent of the depth in SAFT), a large number of invalid signals outside the detection zone are included in the summation, particularly in the region near the focal point. This decreases the SNR in the near-focus region. To avoid this phenomenon, the spatial impulse response (SIR) of a transducer was introduced to weight the spatial contributions in the SAFT algorithm [70]. The SIR defines the size of the sensitivity distribution of the focused ultrasound sensor in the horizontal direction at different depths, as depicted by the focal-cone geometric profiles in Fig. 9. To implement such spatial weighting in SAFT, a binary weighting mask () is applied to multiply the output of the conventional SAFT:
(8) |
where the notation indicates a binary contribution map of SIR, which equals 1 when the pixel () is within the focal cone (Fig. 9) and equals 0 elsewhere. Similarly, after considering the SIR weighting, the CF is modified as follows:
(9) |
(8), (9) indicate that SAFT and CF are performed for pixels located only within the SIR region, such that the final output signal () should be correctly rectified in the summing of adjacent A-lines. Note that the above solution portrays the SIR-based SAFT in one dimension for simplicity. For multi-dimensional SAFT accounting for SIR weighting, the algorithm can be easily extrapolated from (5), (6). More details can be found in Refs. [69], [70]. To demonstrate the efficacy of SIR weighing, Fig. 10a shows the simulated B-scan results of SAFT with and without considering the weighting function. For the case in which the SIR weighting is not accounted for, as displayed in Fig. 10a (2) and (4), the signal near the focus is lower than that outside the focus. As mentioned previously, this is because when spatial weighting is not considered (N is independent of the depth in SAFT), the number of invalid signals included in the region near the focus is greater than that outside the focal region. The invalid signals, which mostly come from background noises, are generally out-of-phase, thereby leading to a destructive summation in the calculation of the CF. According to Eq. (3) and Fig. 3e, this brings about a lower value of CF at the focus than outside the focal region, consequently giving rise to the mismatch in the SNR and resolution between locations outside and inside the focal region. As opposed, SIR-SAFT, as shown in Fig. 10a (3) and (5), can seamlessly rectify the signals at different depths, and provide a homogenous image in terms of lateral resolution and SNR at different depths.
In Ref. [70], the zone applied for SAFT was determined by simulating the detector SIR using a software package (Field II). For simplification, other approaches have been proposed [72], [75] to delineate the SIR region (Fig. 10b and c). As shown in Fig. 10b, the contour of the angular sensitivity distribution is depicted using a hyperbolic function [75]:
(10) |
where and denote the half widths of the minimized and maximized acoustic spreading functions, respectively; and are the depth of the synthesized imaging point and the depth of the focal point, respectively. Eq. (10) provides a criterion to determine whether an arbitrary point (, ) should be involved in the SAFT, i.e., . In addition, a further-simplified solution [72] uses a linear function (Fig. 10c) to account for N (the number of scanlines used for the summation) as a function of depth:
(11) |
where w is the transducer width, and represents the scanning step of the UST between adjacent A-lines. Eq. (11) indicates that pixels that should be involved in the signal summation are encompassed by the crossed area of the two solid black lines in Fig. 10c. The above approaches provide straightforward methods to account for SIR weighing, based on which the involvement of neighboring A-lines is tailored to reflect the SAFT results. However, more phantom and in vivo validations should be performed to examine these schemes with both computational efficiency and image quality taken into consideration. These efforts can help achieve homogeneous signals in terms of SNR and resolution both far from and at the focus of the ultrasonic detector, such that SAFT solutions can ultimately be employed in full three dimensions.
3.5. DMAS-based SAFT
For the synthesis of new A-lines, a beamforming algorithm is required. Generally, DAS is commonly used in SAFT because of its straightforward nature. However, as described in Section 3.1, DAS is a blind beamformer; it linearly combines the delayed A-lines for either the main lobes or side lobes [76], [77], [78], [79], [80], [81]. This linear treatment results in a high noise level and wide beam. To compensate for this blindness, various beamforming solutions and weighting mechanisms have been developed [82], [83], [84], [85], [86], [87], [88], [89], [90], [91], [92], [93], [94], [95], [96], [97]. One approach is the DMAS, which was first used for radar imaging applications [98], and then applied for PA tomography imaging [99], [100], [101], [102], [103], [104], [105], [106]. Recently, DMAS was introduced in the AR-PAM system [71], [72] to address the resolution degradation outside the focus. In contrast to the DAS algorithm, the delayed PA signals from the neighboring scanlines in the DMAS are combinatorially coupled and multiplied before summation. Therefore, the output of the DMAS is expressed as follows:
(12) |
where indicates the delayed RF signals for an arbitrary scanline, which is a simplified version of in Eq. (1). The coupling operation in Eq. (12) can be interpreted as the extraction of spatial coherence information, which means that at each synthesized imaging point, the spatial cross-correlation is computed and emphasized along all of the scanlines. The nonlinear nature of DMAS enables the signals at the main lobes and side lobes to be weighted nonlinearly—signals at the main lobes are highly emphasized by Eq. (12) owing to their strong coherence, while signals at the side lobes are minimized because the artifacts and noises are generally out-of-phase. Consequently, a larger gap in PA amplitude between the main lobes and side lobes (or artifacts) is generated in the DMAS compared with the DAS; thus, better image contrast and resolution can be achieved. Note that Eq. (12) introduces dimensional mismatch because the PA signals are dimensionally squared. To compensate for this inconsistence, extra processing steps—a signed geometric mean to the coupled scanline signals—are added to the DMAS algorithm:
(13) |
With the application of these additional procedures (Eq. (13)), the dimensionality of can be made the same as that of without losing its sign. Then, the output of the generated A-lines for DMAS is given as . A schematic of the DMAS algorithm is presented in Fig. 11a. Note that direct current (DC) and harmonic components generally occur in the spectrum of , thus a bandpass (BP) filter needs to be applied to the output of the DMAS to attenuate the DC component and maintain the harmonic signals. Detailed information on the BP filter can be found in Ref. [71]. The proposed DMAS-based solution was validated experimentally through in vivo imaging of the vascular architecture (Fig. 11b–d). Compared with the original image (Fig. 11b) and DAS-based SAFT (DAS-SAFT) image (Fig. 11c), the image reconstructed using DMAS-based SAFT (DAMS-SAFT) showed better image quality in terms of the resolution and SNR (Fig. 11d), owing to the emphasis of the spatially coherent information. These outcomes verify the effectiveness of DMAS-SAFT for enhancing the coherent components along the scanlines by employing a spatial cross-correlation operation. Note that the algorithms (DMAS-SAFT and DAS-SAFT) shown in Fig. 11c and d are only carried out in one direction, thereby resulting in anisotropy in the lateral resolution. As shown in Fig. 11b–d, the vessel branch inclined to the x-axis (indicated by the blue dashed rectangle) exhibits a noticeably lower lateral resolution compared with the vessel branch along the y-axis (indicated by the green dashed rectangle).
Although DMAS-SAFT provides higher SNR and lateral resolution than DAS-SAFT, it suffers a greater computational burden [71], [72] (Table 1). For example, to process one pixel with N scanlines, the DAS-SAFT requires only N summations (Fig. 11a, second row). In contrast, DMAS-SAFT requires multiplications and summations (Fig. 11a, first row). With the application of graphics processing unit (GPU) hardware [107], [108], the required processing time can be reduced. However, DMAS-SAFT still poses a high burden when applied in a multi-dimensional manner (Section 3.2) to address the anisotropy of the lateral resolution, or implemented in an A-SAFT fashion (Section 3.3) to tackle the wavefront mismatch. Therefore, computationally efficient implementations are required to reduce the computational cost and complexity and facilitate the robust application of DMAS-based SAFT.
3.6. Other SAFT-related and focal-depth expanding approaches
Other SAFT-related strategies, including the double stage (DS)-DMAS [72], the weighted-SAFT (W-SAFT) [60], SAFT-based 3D-deconvolution [67], and the frequency domain SAFT [58], [61], [69], have also been proposed to tackle the low image quality out-of-focus. The DS-DMAS [102], [109] is an extended version of the DMAS algorithm which weighs the calculations by using two stages of DMAS. More specifically, DS-DMAS firstly computes the column sum of the auto-covariance matrix and then employs the DMAS processing in the second stage, further emphasizing on the spatial coherence of received channel data [72]. Though DS-DMAS outperforms DMAS in terms of the resolution improvement and the artifact attenuation [72], [102], [109], DS-DMAS requires a higher computational cost owning to the double-stage procedure [72], [102]. More efforts should be made in future to boost the computational efficiency of the DS-DMAS solution.
W-SAFT [60] is an image formation approach specifically designed for the multiscale PAM—a transition from OR-PAM to AR-PAM. During the transition, the results obtained from a typical scanning PAM are not only influenced by the spatially-varying sensitivity field, but also the inhomogeneous light distribution of the imaged volume, as opposed to a pure AR-PAM system where the illumination was assumed homogenous. To tackle this challenge, W-SAFT was proposed to (1) restrict the voxel’s projection based on the light distribution as well as the FOV of the transducer, and (2) compensate for a moving and heterogeneous illumination distribution. Consequently, W-SAFT helps overcome the resolution degradation out-of-focus in the AR regime without influencing the lateral resolution in the OR regime [60].
The SAFT scheme has been incorporated with 3D-deconvolution for improving the spatial resolution in three dimensions [67]. In earlier studies, 3D-deconvolution algorithms [110] using depth-dependent PSFs were applied to improve both the lateral and axial resolution of AR-PAM, but the performance of the resolution enhancement was conducted only in the focal zone [110]. This restriction stem from the fact that the measured PSFs used for deconvolution at the focal zone differ from those outside the focus. Fortunately, SAFT approaches provide depth-independent PSFs. The combination of the SAFT algorithm helps 3D deconvolution achieve enhanced lateral and axial resolutions both in and outside the focal region [67].
Frequency-domain solutions using the Fourier reconstruction algorithm based on the SAFT concept have been proposed in previous studies [58], [61], [111], in which the ‘delay and sum’ algorithm was conducted in the frequency domain by compensating the phase shift (i.e., time-flight) between the transducer and the virtual detector. Though SAFT in the frequency-domain is principally mathematically equivalent to the time-domain, it has its own merits and limitations. Owing to the fast Fourier transform applied in this strategy, the frequency-domain SAFT shows potentials to reduce the computational cost and burden [58], [61], [111]. However, the frequency-domain SAFT has been reported to generate more reconstruction artifacts than time-domain solutions [112], [113]. Recently, a few studies [58], [61] has been proposed to deal with this specific issue.
To extend the focal region, low NA transducers could be employed but they would compromise the lateral resolution, owing to the tradeoff correlation between the focal depth and the spatial resolution. An alternative to SAFT solutions is to use the deep learning method [114], [115] for the improvement of the image resolution out-of-focus. Deep learning approaches use the results of high-resolution images at the focal region as the ground truth to enhance the out-of-focus results based on the neural network training. Compared with SAFT, deep learning approaches have both merits and disadvantages. The advantages stem from its independence of resolution improvement on the orientation of the sample [115], in contrast with 1D-SAFT in which only targets with orientations perpendicular to the synthetic direction can be well reconstructed. However, in the learning process, a large dataset (acquired images and their corresponding ground truth) is needed for both training, validation and testing [114], [115], as opposed to SAFT strategies where only the measured raw data is required.
4. Conclusions and outlook
AR-PAM is an important imaging modality that renders images with ultrasound resolution and extends the imaging depth beyond the optical ballistic regime. To achieve a high lateral resolution, a large NA of a focused transducer is usually applied for AR-PAM. However, AR-PAM fails to maintain its performance outside the focal region. Based on the concept of the synthetic aperture focusing technique (SAFT), a number of approaches have emerged to address this challenge, including 1D-SAFT, 2D-SAFT, A-SAFT, SIR-based solutions, and DMAS-based strategies. These schemes have shown progress in achieving depth-independent lateral resolution but have various limitations and challenges. This review systematically examines these SAFT-based mechanisms, highlights their essential principles, and underlines the merits and disadvantages of each approach. The origins of various SAFT-based solutions for dealing with focusing-related distortions are geometrically illustrated with detailed mathematical reasoning (2, 3). These analyses can help improve the understanding of the essence of different SAFT-based algorithms for AR-PAM.
The advantages and limitations of each SAFT-based approach originate from their corresponding reconstruction algorithms. 1D-SAFT is considered as a time-efficient approach but causes anisotropy in the lateral resolution. Multi-dimensional SAFT can achieve isotropic lateral resolution but shows inferior performance to the 1D scheme for reconstructing tubular targets such as vascular architecture, owing to the wavefront mismatch mechanism elaborated in Section 3. To date, the most promising algorithm for reconstructing microvasculature is A-SAFT, yet this method is only suitable for tubular-structured targets (cylindrical wavefronts). Given that cylindrical wavefronts may coexist with spherical wavefronts in in vivo environments, such as tumors that are usually accompanied by angiogenesis, effort should be made in the future to incorporate various structured imaging objects in the development of reconstruction algorithms. In addition, further studies are also required to boost the advancements in beamformers. Nonlinear beamforming strategies (e.g., DMAS and DS-DMAS) outperform linear strategies (e.g., DAS) in reducing artifacts and side lobes, but require advanced computational hardware and large image acquisition times owing to the computational complexity arising from the cross-correlation operation. The computational burdens become even greater if nonlinear approaches are used in a multi-dimensional manner to address the wavefront mismatch. Computationally efficient implementations are needed to allow nonlinear SAFT to be applied in a robust and time-efficient manner.
In addition to the aforementioned considerations, future improvements can be made to explore the advantages of both the adaptive feature (Section 3.3) and SIR weighting (Section 3.4). Combining these two approaches may enable SAFT solutions to be employed in a full three-dimensional manner, given that both isotropic resolution in the lateral plane (A-SAFT) and homogeneous resolution and SNR in the axial direction (both far from and at the UST focus (SIR-SAFT)) can be achieved simultaneously. To further improve the imaging quality, SAFT-based schemes can be implemented by incorporating existing approaches of resolution enhancement, such as structured illumination [116], [117], model-based approaches [47], [48], [49], [50], and frequency band filtering [118], [119]. While some of the combined merits have been explored in previous studies [59], [67], [75], further model validations and comprehensive mathematical analyses are required to reinforce these methods. More phantom and in vivo validations should be performed to examine these schemes with both computational efficiency and image quality taken into consideration. This will facilitate a robust performance of the SAFT-based reconstruction methodology and ultimately promote the in vivo applications of AR-APM.
Declaration of Competing Interest
The authors declare that there are no conflicts of interest.
Acknowledgments
This work was supported by National Key Research and Development Program of China (2021YFE0202200 and 2020YFA0908800); National Natural Science Foundation of China (NSFC) Grants (82122034, 92059108, 81927807, 62105355 and 91739117); Chinese Academy of Sciences Grant (YJKYYQ20190078, 2019352 and GJJSTD20210003); CAS Key Laboratory of Health Informatics (2011DP173015); Guangdong Provincial Key Laboratory of Biomedical Optical Imaging (2020B121201010); Shenzhen Key Laboratory for Molecular Imaging (ZDSY20130401165820357); Shenzhen Science and Technology Innovation Grant (JCYJ20210324101403010 and JCYJ20170413153129570); Guangdong Basic and Applied Basic Research Foundation (2020A1515010978); Natural Science Foundation of Shenzhen City (JCYJ20190806150001764).
Biographies
Rongkang Gao is an Associate Professor at SIAT, CAS. She received her Ph.D. degree from the University of New South Wales, Australia, 2018. She joined SIAT in November, 2018. Dr. Gao’s research focuses on nonlinear photoacoustic techniques and algorithm in photoacoustic imaging.
Qiang Xue is a guest student at SIAT, CAS. He received his bachelor degree from Northwest A&F University, 2020. His research interests focus on photoacoustic molecular imaging and the nano-drug delivery system.
Yaguang Ren is an Assistant Professor at SIAT, CAS. In 2018, she got the Ph.D. degree in bioengineering at the Hong Kong University of Science and Technology, Hong Kong, China. Her research interest includes the development of photoacoustic imaging, fluorescent microscope imaging and image processing.
Hai Zhang is a Chief Doctor at Ultrasound Image Department of Shenzhen People’s Hospital and a professor at the First Affiliated Hospital of Southern University of Science and Technology. He received his Master’s degree from Tongji Medical University, 1992. His research interests include the development of multifunctional photoacoustic nanoprobes and their applications in biomedical fields.
Liang Song is a Professor at SIAT, CAS and founding directors of The Research Lab for Biomedical Optics, and the Shenzhen Key Lab for Molecular Imaging. Prior to joining SIAT, he studied at Washington University, St. Louis and received his Ph.D. in Biomedical Engineering. His research focuses on multiple novel photoacoustic imaging technologies.
Chengbo Liu is a Professor at SIAT, CAS. He received both his Ph.D. and Bachelor degree from Xi'an Jiaotong University, each in 2012 in Biophysics and 2007 in Biomedical Engineering. During his Ph.D. training, he spent two years doing tissue spectroscopy research at Duke University as a visiting scholar. Now he is working on multi-scale photoacoustic imaging and its translational research.
References
- 1.Attia A.B.E., Balasundaram G., Moothanchery M., Dinish U.S., Bi R., Ntziachristos V., Olivo M. A review of clinical photoacoustic imaging: current and future trends. Photoacoustics. 2019;16 doi: 10.1016/j.pacs.2019.100144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Wang L.V., Hu S. Photoacoustic tomography: in vivo imaging from organelles to organs. Science. 2012;335(6075):1458–1462. doi: 10.1126/science.1216210. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Gao R., Xu Z., Song L., Liu C. Breaking acoustic limit of optical focusing using photoacoustic-guided wavefront shaping. Laser Photonics Rev. 2021 [Google Scholar]
- 4.Wang L.V., Wu H.-i. John Wiley & Sons; 2012. Biomedical Optics: Principles and Imaging. [Google Scholar]
- 5.Xu M., Wang L.V. Photoacoustic imaging in biomedicine. Rev. Sci. Instrum. 2006;77(4):305–598. [Google Scholar]
- 6.Yong Zhou, Junjie Yao, Lihong, Wang Tutorial on photoacoustic tomography. J. Biomed. Opt. 2016;21:6. doi: 10.1117/1.JBO.21.6.061007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Wang L.V., Yao J. A practical guide to photoacoustic tomography in the life sciences. Nat. Methods. 2016;13(8):627. doi: 10.1038/nmeth.3925. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Jeon S., Kim J., Lee D., Baik J.W., Kim C. Review on practical photoacoustic microscopy. Photoacoustics. 2019;15 doi: 10.1016/j.pacs.2019.100141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Song H., Gonzales E., Soetikno B., Gong E., Wang L.V. Optical-resolution photoacoustic microscopy of ischemic stroke. Proc. SPIE. 2011;7899:789906–789906–5. [Google Scholar]
- 10.Wang X., Pang Y., Ku G., Xie X., Stoica G., Wang L.V. Noninvasive laser-induced photoacoustic tomography for structural and functional in vivo imaging of the brain. Nat. Biotechnol. 2003;21:7. doi: 10.1038/nbt839. [DOI] [PubMed] [Google Scholar]
- 11.Zhang H.F., Maslov K., Wang L.V. In vivo imaging of subcutaneous structures using functional photoacoustic microscopy. Nat. Protoc. 2007;2(4):797–804. doi: 10.1038/nprot.2007.108. [DOI] [PubMed] [Google Scholar]
- 12.Gao R., Xu Z., Ren Y., Song L., Liu C. Nonlinear mechanisms in photoacoustics—powerful tools in photoacoustic imaging. Photoacoustics. 2021;22 doi: 10.1016/j.pacs.2021.100243. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Zhang C., Gao R., Zhang L., Liu C., Yang Z., Zhao S. Design and synthesis of a ratiometric photoacoustic probe for in situ imaging of zinc ions in deep tissue in vivo. Anal. Chem. 2020;92(9):6382–6390. doi: 10.1021/acs.analchem.9b05431. [DOI] [PubMed] [Google Scholar]
- 14.Ning B., Kennedy M.J., Dixon A.J., Sun N., Cao R., Soetikno B.T., Chen R., Zhou Q., Shung K.K., Hossack J.A. Simultaneous photoacoustic microscopy of microvascular anatomy, oxygen saturation, and blood flow. Opt. Lett. 2015;40(6):910–913. doi: 10.1364/OL.40.000910. [DOI] [PubMed] [Google Scholar]
- 15.Schwarz M., Buehler A., Aguirre J., Ntziachristos V. Three‐dimensional multispectral optoacoustic mesoscopy reveals melanin and blood oxygenation in human skin in vivo. J. Biophotonics. 2016;9(1–2):55–60. doi: 10.1002/jbio.201500247. [DOI] [PubMed] [Google Scholar]
- 16.Yao J., Maslov K.I., Zhang Y., Xia Y., Wang L.V. Label-free oxygen-metabolic photoacoustic microscopy in vivo. J. Biomed. Opt. 2011;16(7) doi: 10.1117/1.3594786. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Chen Q., Yin T., Bai Y., Miao X., Gao R., Zhou H., Jie R., Song L., Liu C., Zheng H., Zheng R. Targeted imaging of orthotopic prostate cancer by using clinical transformable photoacoustic molecular probe. BMC Cancer. 2020;20:1–10. doi: 10.1186/s12885-020-06801-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Kim C., Cho E.C., Chen J., Song K.H., Au L., Favazza C., Zhang Q., Cobley C.M., Gao F., Xia Y. In vivo molecular photoacoustic tomography of melanomas targeted by bioconjugated gold nanocages. ACS Nano. 2010;4(8):4559–4564. doi: 10.1021/nn100736c. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Zhou H.-C., Chen N., Zhao H., Yin T., Zhang J., Zheng W., Song L., Liu C., Zheng R. Optical-resolution photoacoustic microscopy for monitoring vascular normalization during anti-angiogenic therapy. Photoacoustics. 2019;15 doi: 10.1016/j.pacs.2019.100143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Jansen K., van der Steen A.F.W., van Beusekom H.M.M., Oosterhuis J.W., van Soest G. Intravascular photoacoustic imaging of human coronary atherosclerosis. Opt. Lett. 2011;36(5):597–599. doi: 10.1364/OL.36.000597. [DOI] [PubMed] [Google Scholar]
- 21.Hu S., Wang L. Neurovascular photoacoustic tomography. Front. Neuroenerg. 2010 doi: 10.3389/fnene.2010.00010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Hu S., Yan P., Maslov K., Lee J.-M., Wang L.V. Intravital imaging of amyloid plaques in a transgenic mouse model using optical-resolution photoacoustic microscopy. Opt. Lett. 2009;34(24):3899–3901. doi: 10.1364/OL.34.003899. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Wang X., Pang Y., Ku G., Xie X., Stoica G., Wang L.V. Noninvasive laser-induced photoacoustic tomography for structural and functional in vivo imaging of the brain. Nat. Biotechnol. 2003;21(7):803–806. doi: 10.1038/nbt839. [DOI] [PubMed] [Google Scholar]
- 24.Li M., Tang Y., Yao J. Photoacoustic tomography of blood oxygenation: a mini review. Photoacoustics. 2018;10:65–73. doi: 10.1016/j.pacs.2018.05.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Gao R., Xu H., Liu L., Zhang Y., Yin T., Zhou H., Sun M., Chen N., Ren Y., Chen T., Pan Y., Zheng M., Ohulchanskyy T.Y., Zheng R., Cai L., Song L., Qu J., Liu C. Photoacoustic visualization of the fluence rate dependence of photodynamic therapy. Biomed. Opt. Express. 2020;11(8):4203–4223. doi: 10.1364/BOE.395562. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Yao J., Wang L.V. Photoacoustic microscopy. Laser Photonics Rev. 2013;7(5):758–778. doi: 10.1002/lpor.201200060. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Wang L.V. Multiscale photoacoustic microscopy and computed tomography. Nat. Photonics. 2009;3(9):503–509. doi: 10.1038/nphoton.2009.157. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Maslov K., Zhang H.F., Hu S., Wang L.V. Optical-resolution photoacoustic microscopy for in vivo imaging of single capillaries. Opt. Lett. 2008;33(9):929–931. doi: 10.1364/ol.33.000929. [DOI] [PubMed] [Google Scholar]
- 29.Zhang H.F., Maslov K., Stoica G., Wang L.V. Functional photoacoustic microscopy for high-resolution and noninvasive in vivo imaging. Nat. Biotechnol. 2006;24(7):848–851. doi: 10.1038/nbt1220. [DOI] [PubMed] [Google Scholar]
- 30.Zhang C., Maslov K., Wang L.V. Subwavelength-resolution label-free photoacoustic microscopy of optical absorption in vivo. Opt. Lett. 2010;35(19):3195–3197. doi: 10.1364/OL.35.003195. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Yao J., Maslov K.I., Puckett E.R., Rowland K.J., Warner B.W., Wang L.V. Double-illumination photoacoustic microscopy. Opt. Lett. 2012;37(4):659–661. doi: 10.1364/OL.37.000659. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Hu S., Maslov K., Wang L.V. Second-generation optical-resolution photoacoustic microscopy with improved sensitivity and speed. Opt. Lett. 2011;36(7):1134–1136. doi: 10.1364/OL.36.001134. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Ma R., Söntges S., Shoham S., Ntziachristos V., Razansky D. Fast scanning coaxial optoacoustic microscopy. Biomed. Opt. Express. 2012;3(7):1724–1731. doi: 10.1364/BOE.3.001724. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Maslov K., Stoica G., Wang L.V. In vivo dark-field reflection-mode photoacoustic microscopy. Opt. Lett. 2005;30(6):625–627. doi: 10.1364/ol.30.000625. [DOI] [PubMed] [Google Scholar]
- 35.Song K.H., Stein E., Margenthaler J., Wang L. Noninvasive photoacoustic identification of sentinel lymph nodes containing methylene blue in vivo in a rat model. J. Biomed. Opt. 2008;13(5) doi: 10.1117/1.2976427. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Wang L., Maslov K., Xing W., Garcia-Uribe A., Wang L. Video-rate functional photoacoustic microscopy at depths. J. Biomed. Opt. 2012;17(10) doi: 10.1117/1.JBO.17.10.106007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Harrison T., Ranasinghesagara J.C., Lu H., Mathewson K., Walsh A., Zemp R.J. Combined photoacoustic and ultrasound biomicroscopy. Opt. Express. 2009;17(24):22041–22046. doi: 10.1364/OE.17.022041. [DOI] [PubMed] [Google Scholar]
- 38.Wang X., Ku G., Wegiel M.A., Bornhop D.J., Stoica G., Wang L.V. Noninvasive photoacoustic angiography of animal brains in vivo with near-infrared light and an optical contrast agent. Opt. Lett. 2004;29(7):730–732. doi: 10.1364/ol.29.000730. [DOI] [PubMed] [Google Scholar]
- 39.Liao L.-D., Li M.-L., Lai H.-Y., Shih Y.-Y.I., Lo Y.-C., Tsang S., Chao P.C.-P., Lin C.-T., Jaw F.-S., Chen Y.-Y. Imaging brain hemodynamic changes during rat forepaw electrical stimulation using functional photoacoustic microscopy. NeuroImage. 2010;52(2):562–570. doi: 10.1016/j.neuroimage.2010.03.065. [DOI] [PubMed] [Google Scholar]
- 40.Zerda Adl, Liu Z., Bodapati S., Teed R., Vaithilingam S., Khuri-Yakub B.T., Chen X., Dai H., Gambhir S.S. Ultrahigh sensitivity carbon nanotube agents for photoacoustic molecular imaging in living mice. Nano Lett. 2010;10(6):2168–2172. doi: 10.1021/nl100890d. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Maslov K., Wang L. Photoacoustic imaging of biological tissue with intensity-modulated continuous-wave laser. J. Biomed. Opt. 2008;13(2) doi: 10.1117/1.2904965. [DOI] [PubMed] [Google Scholar]
- 42.Song K., Wang L. Deep reflection-mode photoacoustic imaging of biological tissue. J. Biomed. Opt. 2007;12(6) doi: 10.1117/1.2818045. [DOI] [PubMed] [Google Scholar]
- 43.Song K.H., Wang L.V. Noninvasive photoacoustic imaging of the thoracic cavity and the kidney in small and large animals. Med. Phys. 2008;35(10):4524–4529. doi: 10.1118/1.2977534. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Zhang H.F., Maslov K., Li M.-L., Stoica G., Wang L.V. In vivo volumetric imaging of subcutaneous microvasculature by photoacoustic microscopy. Opt. Express. 2006;14(20):9317–9323. doi: 10.1364/oe.14.009317. [DOI] [PubMed] [Google Scholar]
- 45.Stein E., Maslov K., Wang L. Noninvasive, in vivo imaging of blood-oxygenation dynamics within the mouse brain using photoacoustic microscopy. J. Biomed. Opt. 2009;14(2) doi: 10.1117/1.3095799. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Park S., Lee C., Kim J., Kim C. Acoustic resolution photoacoustic microscopy. Biomed. Eng. Lett. 2014;4(3):213–222. [Google Scholar]
- 47.Araque Caballero M.Á., Rosenthal A., Gateau J., Razansky D., Ntziachristos V. Model-based optoacoustic imaging using focused detector scanning. Opt. Lett. 2012;37(19):4080–4082. doi: 10.1364/OL.37.004080. [DOI] [PubMed] [Google Scholar]
- 48.X.L. Deán-Ben, H. Estrada, M. Kneipp, J. Turner, D. Razansky, Three-dimensional modeling of the transducer shape in acoustic resolution optoacoustic microscopy, in: Proceedings of the SPIE BiOS, Vol. 8943, SPIE, 2014.
- 49.Aguirre J., Giannoula A., Minagawa T., Funk L., Turon P., Durduran T. A low memory cost model based reconstruction algorithm exploiting translational symmetry for photoacoustic microscopy. Biomed. Opt. Express. 2013;4(12):2813–2827. doi: 10.1364/BOE.4.002813. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Jin H., Zhang R., Liu S., Zheng Y. Fast and high-resolution three-dimensional hybrid-domain photoacoustic imaging incorporating analytical-focused transducer beam amplitude. IEEE Trans. Med. Imaging. 2019;38(12):2926–2936. doi: 10.1109/TMI.2019.2917688. [DOI] [PubMed] [Google Scholar]
- 51.T. Berer, A. Hochreiner, H. Roitner, P. Burgholzer, Reconstruction algorithms for remote photoacoustic imaging, in: Proceedings of the 2012 IEEE International Ultrasonics Symposium, 2012.
- 52.Cox B.T., Treeby B.E. Artifact trapping during time reversal photoacoustic imaging for acoustically heterogeneous media. IEEE Trans. Med. Imaging. 2010;29(2):387–396. doi: 10.1109/TMI.2009.2032358. [DOI] [PubMed] [Google Scholar]
- 53.Cox B.T., Kara S., Arridge S.R., Beard P.C. k-space propagation models for acoustically heterogeneous media: application to biomedical photoacoustics. J. Acoust. Soc. Am. 2007;121(6):3453–3464. doi: 10.1121/1.2717409. [DOI] [PubMed] [Google Scholar]
- 54.Farnia P., Mohammadi M., Najafzadeh E., Alimohamadi M., Makkiabadi B., Ahmadian A. High-quality photoacoustic image reconstruction based on deep convolutional neural network: towards intra-operative photoacoustic imaging. Biomed. Phys. Eng. Express. 2020;6(4) doi: 10.1088/2057-1976/ab9a10. [DOI] [PubMed] [Google Scholar]
- 55.Jin H., Zheng Z., Liu S., Zhang R., Liao X., Liu S., Zheng Y. Pre-migration: a general extension for photoacoustic imaging reconstruction. IEEE Trans. Comput. Imaging. 2020;6:1097–1105. [Google Scholar]
- 56.Liao C.K., Li M.L., Li P.C. Optoacoustic imaging with synthetic aperture focusing and coherence weighting. Opt. Lett. 2004;29(21):2506–2508. doi: 10.1364/ol.29.002506. [DOI] [PubMed] [Google Scholar]
- 57.Li M.-L., Zhang H.F., Maslov K., Stoica G., Wang L.V. Improved in vivo photoacoustic microscopy based on a virtual-detector concept. Opt. Lett. 2006;31(4):474–476. doi: 10.1364/ol.31.000474. [DOI] [PubMed] [Google Scholar]
- 58.Spadin F., Jaeger M., Nuster R., Subochev P., Frenz M. Quantitative comparison of frequency-domain and delay-and-sum optoacoustic image reconstruction including the effect of coherence factor weighting. Photoacoustics. 2020;17 doi: 10.1016/j.pacs.2019.100149. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Amjadian M., Mostafavi S.M., Chen J., Kavehvash Z., Zhu J., Wang L. Super-resolution photoacoustic microscopy using structured-illumination. IEEE Trans. Med. Imaging. 2021 doi: 10.1109/TMI.2021.3073555. [DOI] [PubMed] [Google Scholar]
- 60.Turner J., Estrada H., Kneipp M., Razansky D. Universal weighted synthetic aperture focusing technique (W-SAFT) for scanning optoacoustic microscopy. Optica. 2017;4(7):770–778. [Google Scholar]
- 61.Jin H., Liu S., Zhang R., Liu S., Zheng Y. Frequency domain based virtual detector for heterogeneous media in photoacoustic imaging. IEEE Trans. Comput. Imaging. 2020;6:569–578. [Google Scholar]
- 62.Meng-Lin L., Wei-Jung G., Pai-Chi L. Improved synthetic aperture focusing technique with applications in high-frequency ultrasound imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2004;51(1):63–70. doi: 10.1109/tuffc.2004.1268468. [DOI] [PubMed] [Google Scholar]
- 63.Schickert M., Krause M., Müller W. Ultrasonic imaging of concrete elements using reconstruction by synthetic aperture focusing technique. J. Mater. Civ. Eng. 2003;15(3):235–246. [Google Scholar]
- 64.Frazier C.H., Brien W.D.O. Synthetic aperture techniques with a virtual source element. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 1998;45(1):196–207. doi: 10.1109/58.646925. [DOI] [PubMed] [Google Scholar]
- 65.Jensen J.A., Nikolov S.I., Gammelmark K.L., Pedersen M.H. Synthetic aperture ultrasound imaging. Ultrasonics. 2006;44:e5–e15. doi: 10.1016/j.ultras.2006.07.017. [DOI] [PubMed] [Google Scholar]
- 66.Deng Z., Yang X., Gong H., Luo Q. Two-dimensional synthetic-aperture focusing technique in photoacoustic microscopy. J. Appl. Phys. 2011;109(10) [Google Scholar]
- 67.Cai D., Li Z., Li Y., Guo Z., Chen S.-L. Photoacoustic microscopy in vivo using synthetic-aperture focusing technique combined with three-dimensional deconvolution. Opt. Express. 2017;25(2):1421–1434. doi: 10.1364/OE.25.001421. [DOI] [PubMed] [Google Scholar]
- 68.Deng Z., Yang X., Gong H., Luo Q. Adaptive synthetic-aperture focusing technique for microvasculature imaging using photoacoustic microscopy. Opt. Express. 2012;20(7):7555–7563. doi: 10.1364/OE.20.007555. [DOI] [PubMed] [Google Scholar]
- 69.Jeon S., Park J., Managuli R., Kim C. A novel 2-D synthetic aperture focusing technique for acoustic-resolution photoacoustic microscopy. IEEE Trans. Med. Imaging. 2019;38(1):250–260. doi: 10.1109/TMI.2018.2861400. [DOI] [PubMed] [Google Scholar]
- 70.Turner J., Estrada H., Kneipp M., Razansky D. Improved optoacoustic microscopy through three-dimensional spatial impulse response synthetic aperture focusing technique. Opt. Lett. 2014;39(12):3390–3393. doi: 10.1364/OL.39.003390. [DOI] [PubMed] [Google Scholar]
- 71.Park J., Jeon S., Meng J., Song L., Lee J., Kim C. Delay-multiply-and-sum-based synthetic aperture focusing in photoacoustic microscopy. J. Biomed. Opt. 2016;21(3) doi: 10.1117/1.JBO.21.3.036010. [DOI] [PubMed] [Google Scholar]
- 72.Mozaffarzadeh M., Varnosfaderani M.H.H., Sharma A., Pramanik M., de Jong N., Verweij M.D. Enhanced contrast acoustic-resolution photoacoustic microscopy using double-stage delay-multiply-and-sum beamformer for vasculature imaging. J. Biophotonics. 2019;12(11) doi: 10.1002/jbio.201900133. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.K.W. Hollman, K.W. Rigby, M.O. Donnell, Coherence factor of speckle from a multi-row probe, in: Proceedings of the 1999 IEEE International Ultrasonics Symposium (Cat. No. 99CH37027), 1999.
- 74.Mozaffarzadeh M., Makkiabadi B., Basij M., Mehrmohammadi M. Image improvement in linear-array photoacoustic imaging using high resolution coherence factor weighting technique. BMC Biomed. Eng. 2019;1(1):10. doi: 10.1186/s42490-019-0009-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Tsunoi Y., Yoshimi K., Watanabe R., Kumai N., Terakawa M., Sato S. Quality improvement of acoustic-resolution photoacoustic imaging of skin vasculature based on practical synthetic-aperture focusing and bandpass filtering. Jpn. J. Appl. Phys. 2018;57(12) [Google Scholar]
- 76.K.E. Thomenius, Evolution of ultrasound beamformers, in: Proceedings of the 1996 IEEE Ultrasonics Symposium, 1996.
- 77.Hoelen C.G.A., de Mul F.F.M. Image reconstruction for photoacoustic scanning of tissue structures. Appl. Opt. 2000;39(31):5872–5883. doi: 10.1364/ao.39.005872. [DOI] [PubMed] [Google Scholar]
- 78.Pramanik M. Improving tangential resolution with a modified delay-and-sum reconstruction algorithm in photoacoustic and thermoacoustic tomography. J. Opt. Soc. Am. A. 2014;31(3):621–627. doi: 10.1364/JOSAA.31.000621. [DOI] [PubMed] [Google Scholar]
- 79.G. Matrone, A.S. Savoia, G. Caliano, G. Magenes, Ultrasound plane-wave imaging with delay multiply and sum beamforming and coherent compounding, in: Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 2016. [DOI] [PubMed]
- 80.Kim J., Park S., Jung Y., Chang S., Park J., Zhang Y., Lovell J.F., Kim C. Programmable real-time clinical photoacoustic and ultrasound imaging system. Sci. Rep. 2016;6(1):35137. doi: 10.1038/srep35137. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Park S., Karpiouk A.B., Aglyamov S.R., Emelianov S.Y. Adaptive beamforming for photoacoustic imaging. Opt. Lett. 2008;33(12):1291–1293. doi: 10.1364/ol.33.001291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Mozaffarzadeh M., Mahloojifar A., Orooji M., Kratkiewicz K., Adabi S., Nasiriavanaki M. Linear-array photoacoustic imaging using minimum variance-based delay multiply and sum adaptive beamforming algorithm. J. Biomed. Opt. 2018;23(2) doi: 10.1117/1.JBO.23.2.026002. [DOI] [PubMed] [Google Scholar]
- 83.Mozaffarzadeh M., Yan Y., Mehrmohammadi M., Makkiabadi B. Enhanced linear-array photoacoustic beamforming using modified coherence factor. J. Biomed. Opt. 2018;23(2) doi: 10.1117/1.JBO.23.2.026005. [DOI] [PubMed] [Google Scholar]
- 84.Mozaffarzadeh M., Mahloojifar A., Periyasamy V., Pramanik M., Orooji M. Eigenspace-based minimum variance combined with delay multiply and sum beamformer: application to linear-array photoacoustic imaging. IEEE J. Sel. Top. Quantum Electron. 2019;25(1):1–8. [Google Scholar]
- 85.Matrone G., Savoia A.S., Caliano G., Magenes G. The delay multiply and sum beamforming algorithm in ultrasound B-mode medical imaging. IEEE Trans. Med. Imaging. 2015;34(4):940–949. doi: 10.1109/TMI.2014.2371235. [DOI] [PubMed] [Google Scholar]
- 86.Paridar R., Mozaffarzadeh M., Mehrmohammadi M., Orooji M. Photoacoustic image formation based on sparse regularization of minimum variance beamformer. Biomed. Opt. Express. 2018;9(6):2544–2561. doi: 10.1364/BOE.9.002544. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Synnevag J., Austeng A., Holm S. Benefits of minimum-variance beamforming in medical ultrasound imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2009;56(9):1868–1879. doi: 10.1109/TUFFC.2009.1263. [DOI] [PubMed] [Google Scholar]
- 88.Wang S., Li P. MVDR-based coherence weighting for high-frame-rate adaptive imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2009;56(10):2097–2110. doi: 10.1109/TUFFC.2009.1293. [DOI] [PubMed] [Google Scholar]
- 89.J. Synnevåg, C.C. Nilsen, S. Holm, P2B-13 speckle statistics in adaptive beamforming, in: Proceedings of the 2007 IEEE Ultrasonics Symposium, 2007.
- 90.Mehdizadeh S., Austeng A., Johansen T.F., Holm S. Eigenspace based minimum variance beamforming applied to ultrasound imaging of acoustically hard tissues. IEEE Trans. Med. Imaging. 2012;31(10):1912–1921. doi: 10.1109/TMI.2012.2208469. [DOI] [PubMed] [Google Scholar]
- 91.M. Mozaffarzadeh, A. Mahloojifar, M. Orooji, Medical photoacoustic beamforming using minimum variance-based delay multiply and sum, in: Proceedings of the SPIE Digital Optical Technologies, Vol. 10335, SPIE, 2017.
- 92.Mozaffarzadeh M., Periyasamy V., Pramanik M., Makkiabadi B. Efficient nonlinear beamformer based on P’th root of detected signals for linear-array photoacoustic tomography: application to sentinel lymph node imaging. J. Biomed. Opt. 2018;23(12) doi: 10.1117/1.JBO.23.12.121604. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Synnevag J.F., Austeng A., Holm S. Adaptive beamforming applied to medical ultrasound imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2007;54(8):1606–1613. [PubMed] [Google Scholar]
- 94.Zhang H.K., Bell M.A.L., Guo X., Kang H.J., Boctor E.M. Synthetic-aperture based photoacoustic re-beamforming (SPARE) approach using beamformed ultrasound data. Biomed. Opt. Express. 2016;7(8):3056–3068. doi: 10.1364/BOE.7.003056. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Zeng X., Wang Y., Yu J., Guo Y. Correspondence – beam-domain eigenspace-based minimum variance beamformer for medical ultrasound imaging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2013;60(12):2670–2676. doi: 10.1109/TUFFC.2013.2866. [DOI] [PubMed] [Google Scholar]
- 96.M. Polichetti, F. Varray, G. Matrone, A.S. Savoia, J. Béra, C. Cachard, B. Nicolas, A computationally efficient nonlinear beamformer based on p-th root signal compression for enhanced ultrasound B-mode imaging, in: Proceedings of the IEEE International Ultrasonics Symposium (IUS), 2017.
- 97.Cui M., Zuo H., Wang X., Deng K., Luo J., Ma C. Adaptive photoacoustic computed tomography. Photoacoustics. 2021;21 doi: 10.1016/j.pacs.2020.100223. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Lim H.B., Nhung N.T.T., Li E.P., Thang N.D. Confocal microwave imaging for breast cancer detection: delay-multiply-and-sum image reconstruction algorithm. IEEE Trans. Biomed. Eng. 2008;55(6):1697–1704. doi: 10.1109/tbme.2008.919716. [DOI] [PubMed] [Google Scholar]
- 99.Mozaffarzadeh M., Mahloojifar A., Orooji M., Adabi S., Nasiriavanaki M. Double-stage delay multiply and sum beamforming algorithm: application to linear-array photoacoustic imaging. IEEE Trans. Biomed. Eng. 2018;65(1):31–42. doi: 10.1109/TBME.2017.2690959. [DOI] [PubMed] [Google Scholar]
- 100.Jeon S., Park E.-Y., Choi W., Managuli R., Lee Kj, Kim C. Real-time delay-multiply-and-sum beamforming with coherence factor for in vivo clinical photoacoustic imaging of humans. Photoacoustics. 2019;15 doi: 10.1016/j.pacs.2019.100136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Mozaffarzadeh M., Hariri A., Moore C., Jokerst J.V. The double-stage delay-multiply-and-sum image reconstruction method improves imaging quality in a LED-based photoacoustic array scanner. Photoacoustics. 2018;12:22–29. doi: 10.1016/j.pacs.2018.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Mozaffarzadeh M., Sadeghi M., Mahloojifar A., Orooji M. Double-stage delay multiply and sum beamforming algorithm applied to ultrasound medical imaging. Ultrasound Med. Biol. 2018;44(3):677–686. doi: 10.1016/j.ultrasmedbio.2017.10.020. [DOI] [PubMed] [Google Scholar]
- 103.A. Alshaya, S. Harput, A.M. Moubark, D.M.J. Cowell, J. McLaughlan, S. Freear, Spatial resolution and contrast enhancement in photoacoustic imaging with filter delay multiply and sum beamforming technique, in: Proceedings of the IEEE International Ultrasonics Symposium (IUS), 2016.
- 104.M. Mozaffarzadeh, A. Mahloojifar, M. Nasiriavanaki, M. Orooji, Eigenspace-based minimum variance adaptive beamformer combined with delay multiply and sum: experimental study, in: Proceedings of the SPIE BiOS, Vol. 10467, SPIE, 2018.
- 105.Choi W., Oh D., Kim C. Practical photoacoustic tomography: realistic limitations and technical solutions. J. Appl. Phys. 2020;127(23) [Google Scholar]
- 106.Cho S., Jeon S., Choi W., Managuli R., Kim C. Nonlinear pth root spectral magnitude scaling beamforming for clinical photoacoustic and ultrasound imaging. Opt. Lett. 2020;45(16):4575–4578. doi: 10.1364/OL.393315. [DOI] [PubMed] [Google Scholar]
- 107.Rostami S.R.M., Mozaffarzadeh M., Ghaffari-Miab M., Hariri A., Jokerst J.V. GPU-accelerated double-stage delay-multiply-and-sum algorithm for fast photoacoustic tomography using LED excitation and linear arrays. Ultrason. Imaging. 2019;41(5):301–316. doi: 10.1177/0161734619862488. [DOI] [PubMed] [Google Scholar]
- 108.S.R. Miri Rostami, M. Mozaffarzadeh, A. Hariri, J. Jokerst, M. Ghaffari-Miab, OpenACC GPU implementation of double-stage delay-multiply-and-sum algorithm: toward enhanced real-time linear-array photoacoustic tomography, in: Proceedings of the SPIE BiOS, Vol. 10878, SPIE, 2019.
- 109.Shen C.-C. A study of double-stage DMAS and p-DMAS for their relation in baseband ultrasound beamforming. Biomed. Signal Process. Control. 2020;60 [Google Scholar]
- 110.Cai D., Li Z., Chen S.-L. In vivo deconvolution acoustic-resolution photoacoustic microscopy in three dimensions. Biomed. Opt. Express. 2016;7(2):369–380. doi: 10.1364/BOE.7.000369. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111.Boonsang S., Zainal J., Dewhurst R. Synthetic aperture focusing techniques in time and frequency domains for photoacoustic imaging. Insight Non-Destr. Test. Cond. Monit. 2004;46(4):196–199. [Google Scholar]
- 112.Köstli K.P., Frenz M., Bebie H., Weber H.P. Temporal backward projection of optoacoustic pressure transients using Fourier transform methods. Phys. Med. Biol. 2001;46(7):1863. doi: 10.1088/0031-9155/46/7/309. [DOI] [PubMed] [Google Scholar]
- 113.Schulze R.J., Scherzer O., Zangerl G., Holotta M., Meyer D., Handle F., Nuster R., Paltauf G. On the use of frequency-domain reconstruction algorithms for photoacoustic imaging. J. Biomed. Opt. 2011;16(8) doi: 10.1117/1.3605696. [DOI] [PubMed] [Google Scholar]
- 114.Sharma A., Pramanik M. Convolutional neural network for resolution enhancement and noise reduction in acoustic resolution photoacoustic microscopy. Biomed. Opt. Express. 2020;11(12):6826–6839. doi: 10.1364/BOE.411257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Yang C., Lan H., Gao F., Gao F. Review of deep learning for photoacoustic imaging. Photoacoustics. 2021;21 doi: 10.1016/j.pacs.2020.100215. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 116.Gustafsson M.G.L. Surpassing the lateral resolution limit by a factor of two using structured illumination microscopy. J. Microsc. 2000;198(2):82–87. doi: 10.1046/j.1365-2818.2000.00710.x. [DOI] [PubMed] [Google Scholar]
- 117.Ilovitsh T., Ilovitsh A., Foiret J., Fite B.Z., Ferrara K.W. Acoustical structured illumination for super-resolution ultrasound imaging. Commun. Biol. 2018;1(1):3. doi: 10.1038/s42003-017-0003-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118.Thomas, Chaigne, Jérôme Gateau, Ori Katz, Claude Boccara, Sylvain, Gigan Improving photoacoustic-guided optical focusing in scattering media by spectrally filtered detection. Opt. Lett. 2014;39(20):6054–6057. doi: 10.1364/OL.39.006054. [DOI] [PubMed] [Google Scholar]
- 119.Najafzadeh E., Farnia P., Lavasani S., Basij M., Yan Y., Ghadiri H., Ahmadian A., Mehrmohammadi M. Photoacoustic image improvement based on a combination of sparse coding and filtering. J. Biomed. Opt. 2020;25(10) doi: 10.1117/1.JBO.25.10.106001. [DOI] [PMC free article] [PubMed] [Google Scholar]