Skip to main content
Photoacoustics logoLink to Photoacoustics
. 2019 Nov 6;16:100142. doi: 10.1016/j.pacs.2019.100142

Optoacoustic model-based inversion using anisotropic adaptive total-variation regularization

Shai Biton 1,1, Nadav Arbel 1,1, Gilad Drozdov 1, Guy Gilboa 1, Amir Rosenthal 1,*
PMCID: PMC6849433  PMID: 31737487

Abstract

In optoacoustic tomography, image reconstruction is often performed with incomplete or noisy data, leading to reconstruction errors. Significant improvement in reconstruction accuracy may be achieved in such cases by using nonlinear regularization schemes, such as total-variation minimization and L1-based sparsity-preserving schemes. In this paper, we introduce a new framework for optoacoustic image reconstruction based on adaptive anisotropic total-variation regularization, which is more capable of preserving complex boundaries than conventional total-variation regularization. The new scheme is demonstrated in numerical simulations on blood-vessel images as well as on experimental data and is shown to be more capable than the total-variation-L1 scheme in enhancing image contrast.

Keywords: Optoacoustic imaging, Total variation, Inversion algorithms, Model-based reconstruction

1. Introduction

Optoacoustic tomography (OAT) is a hybrid imaging modality capable of visualizing optically absorbing structures with ultrasound resolution at tissue depths in which light is fully diffused [1], [2], [3], [4], [5]. The excitation in OAT is most often performed via high-energy short pulses whose absorption in the tissue leads to the generation of acoustic sources via the process of thermal expansion. The acoustic signals from the sources are measured over a surface that partially or fully surrounds the imaged object and used to form an image of the acoustic sources, which generally represents the local energy absorption in the tissue [6]. Since hemoglobin is one of the strongest absorbing tissue constituents, optoacoustic images often depict blood vessels and blood-rich organs, such as the kidneys [7] and heart [8].

Numerous algorithms exist for reconstructing optoacoustic images from measured tomographic data [6]. In several imaging geometries, analytical formulae exist that may be applied directly on the measured data in either time [9], [10] or frequency domain [11] to yield an exact reconstruction. The popularity of analytical formulae may be attributed to the simplicity of their implementation and low computational burden [12] However, analytical algorithms are not exact for arbitrary detection surfaces or detector geometries and lack the possibility of regularizing the inversion in the case of noisy or incomplete data. In such cases, it is often preferable to use model-based algorithms, in which the relation between the image and measured data is represented by a matrix whose inversion is required to reconstruct the image.

In the last decade, numerous regularization approaches have been demonstrated for model-based image reconstruction. The most basic approach is based on energy minimization and includes techniques such as Tikhonov regularization [13] and truncated singular-value decomposition [14]. In these techniques, a cost function on the image or components thereof is used to avoid divergence of the solution in the case of missing data, generally without making any assumptions on the nature of the solution. More advanced approaches to regularization exploit the specific properties of the reconstructed image. Since natural images may be sparsely represented when transformed to an alternative basis, e.g. the wavelet basis, using nonlinear cost functions that promote sparsity in such bases may be used for denoising and image reconstruction from missing data [15], [16], [17], [18], [19], [20], [21]. In images of blood vessels, in which the boundaries of the imaged structures may be of higher importance than the texture in the image, total-variation (TV) minimization has been shown to enhance image contrast and reduce artifacts [22], [23], [24]

In TV regularization, the cost function regularizer is the L1 norm of the image gradient, which is generally lower for images with sharp, yet very localized, variations than for images in which small variations occur across the entire image. Therefore TV regularization enhances boundaries and reduces texture, where over-regularization may lead to almost piecewise constant images, which are often referred to as cartoon-like. While TV regularization is capable of accentuating the boundaries of imaged objects, it does not treat all boundaries the same. In particular, boundaries with short lengths will lead to a lower TV cost function than boundaries with long lengths. Thus, complex, non-convex boundaries may be rounded by TV regularization to the closest convex form. Recently, Wang et al. have shown that if the directionality of the TV functional is adapted to the image features, TV regularization may be applied for optoacoustic reconstruction of objects with non-convex boundaries without distorting the boundaries [25].

In this paper, we demonstrate a new regularization framework for model-based optoacoustic image reconstruction that overcomes the limitations of TV regularization and is compatible with objects with complex non-convex boundaries. In our scheme, an adaptive anisotropic total variation (A2TV) cost function is used, in which the cost function is determined by the specific geometry of the imaged objects [26]. In particular, the A2TV cost function wishes to minimize the total variation in directions that are orthogonal to the boundary of the object. In contrast to [25], where the boundaries were calculated using geometrical considerations limited to 2D images, the A2TV framework developed in this work is based on eigenvalue decomposition of the image structure tensor, which may be applied in higher dimensions. The proposed formalism in the current study is based on a recent work by part of the authors which is concerned with nonlinear spectral analysis of the A2TV functional [26]. The work of Ref. [26] examines shapes which are perfectly preserved under A2TV regularization. Earlier works concerning TV have shown that only convex rounded shapes of low curvature are preserved [27]. For A2TV, however, it is shown in [26] that a parameter controlling the local extent of directionality is directly related to the degree of convexity (in the sense of [28]) and to the curvature magnitude of structures which are preserved. Thus, with appropriate parameters, long vessels of complex-nonconvex structure can be better regularized, keeping the original structure intact.

The performance of A2TV regularization was tested numerically in this work in numerical simulations for complex images of blood vessels and experimentally on 2D phantoms. The simulations were performed for the cases of noisy data and missing data and compared to unregularized reconstructions as well as to TV-L1 reconstructions [15]. In both the numerical and experimental reconstructions, A2TV significantly increased the image contrast and was more capable than TV-L1 in preserving non-convex structures when strong regularization was performed. In the experimental reconstructions, A2TV achieved a higher level of contrast enhancement of weak structures than the one achieved by TV-L1.

The rest of the paper is organized as follows: in Section 2 we give the theoretical background for OAT image reconstruction. Section 3 introduces the framework of A2TV and the A2TV algorithm for OAT image reconstruction developed in this work. The simulation results are given in Section 4, while the experimental ones are given in Section 5. We conclude the paper with a Discussion in Section 6.

2. Optoacoustic image reconstruction

2.1. The forward problem

The acoustic waves in OAT are commonly described by a pressure field p(r, t) that fulfills the following wave equation [29]:

2p(r,t)t2c22p(r,t)=ΓHr(r)Ht(t)t (1)

where c is the speed of sound in the medium, t is time, r = (x, y, z) denotes position in 3D space, p(r, t) is the generated pressure, Γ is the Grüneisen parameter, and Hr(r)Ht(t) is the energy per unit volume and unit time. The spatial distribution function of energy deposited in the imaged object, Hr(r), is referred to in the rest of the paper as the optoacoustic image.

The analysis of (1) for an optoacoustic point source at r′, i.e. Hr(r) = δ(r − r′), may be performed in either time or frequency domain. In the time domain, a short-pulse excitation Ht(t) = δ(t) is used. In this case, the solution to Eq. (1) is given by [30]

pδ(r,t)=Γ4πctδ(|rr|ct)|rr|. (2)

For a general image Hr(r), the solution for p(r, t) may be obtained by convolving Hr(r) with the expression in Eq. (2), which yields:

p(r,t)=Γ4πct|rr|=ctHr(r)|rr|. (3)

Since the relation between the measured pressure signals, or projections, and the image is linear, it may be represented by a matrix relation in its discrete form:

p=Mu (4)

where p and u are vector representations of the acoustic signals and originating image respectively, and M is the model matrix that represents the operations in Eq. (3). In our work, the images is given on a two dimensional grid, and the measured pressure signals are also two dimensional, where one dimension represents the projection number, and the other time. An illustration of the image grid and projection and their respective mapping to the vectors u and p, is shown in Fig. 1. Briefly, the vector u is divided into sub-vectors, each representing the image values for a given column, whereas the vector p is divided to sub-vectors, each of which represents the time-domain pressure signal for a given location of the acoustic detector.

Fig. 1.

Fig. 1

An illustration of the structure of vectors p and u used in the matrix construction in Eq. (4).

The ith column of the matrix M represents the set of acoustic signals generated by a pixel corresponding to the location of the ith entry in the vector u. Accordingly, to calculate the matrix M one needs to define time domain signal for a given detector location expected for a discrete pixel. Since the operations in Eq. (3) relate to continuous, rather than discrete images, one first needs to define the continuous representation of a single pixel, and then calculate its respective time-domain signals. For example, in [31] it was assumed that the image value was constant over each of the square pixels, leading to an image Hr(r) that is piece-wise constant. While simple to implement, a simple piece-wise uniform model for Hr(r) includes discontinuities that lead to significant numerical errors owing to the derivative operation in Eq. (3). In the current work, we use the model of [32], in which the image Hr(r) represented by a linear interpolation between its grid points.

2.2. The inverse problem

While several approaches to OAT image reconstruction exist, we focus herein on image reconstruction within the discrete model-based framework described in the previous sub-section, which involves inverting the matrix relation in Eq. (4) to recover u from p. The most basic method to invert Eq. (4) is based on solving the following optimization problem:

u*=argminpMu22, (5)

where u* is the solution and ∥ · ∥ 2 is the L2 norm. A unique solution to Eq. (5) exists, which is given by the Moore–Penrose inverse:

u*=(MTM)1MTp. (6)

Alternatively, Eq. (5) may be solved via iterative optimization algorithms. In particular, since the matrix M is sparse, efficient inversion may be achieved by the LSQR algorithm [33].

In many cases, the measured projection data p is insufficient to achieve a high-quality reconstruction of u that accurately depicts its morphology. For example, when the density or coverage of the projections is too small, the matrix M may become ill-conditioned, leading to significant, possibly divergent, image artifact. In other cases, M may be well conditioned, but the measurement data may be too noisy to accurately recover u. In both these cases, regularization may be used to improve image quality by incorporating previous knowledge on the properties image in the inversion process.

One of the simplest forms of regularization is Tikhonov regularization, in which an additional cost function is added [13]:

u*=argminpMu22+λLu22 (7)

where L is weighting matrix. In the simplest form of Tikhonov regularization, L is equal to the identity matrix, i.e. L = I, thus putting a penalty on the energy of the image. The value of the regularization parameter λ > 0 controls the tradeoff between fidelity and smoothness, where over-regularization may lead to the smearing of edges and texture in the image.

An alternative to the energy-minimizing cost function of Tikhonov are sparsity-maximizing cost functions. Since natural images may be sparsely represented in an alternative basis, e.g. the wavelet basis, a cost function that promotes sparsity may reduce reconstruction errors. Denoting the transformation matrix by Φ, one wishes that Φu be sparse, i.e. that most of its entries be approximately zero. In practice, sparsity is often enforced by using the L1 norm because of its compatibility of optimization algorithms [16], [17], [18], [19], [20], [21]. Accordingly, the inversion is performed by solving the following optimization problem:

u*=argminpMu22+μΦu1, (8)

where μ > 0 is the regularization parameter, controlling the tradeoff between sparsity and signal fidelity. When over regularization is performed, compression artifacts may appear in the reconstructed image.

Sparsity may be enforced not only on alternative representations of the image, but also on image variations. The discrete TV cost function approximates the l1 norm on the image gradient, and is given by

uTV=n|ux,ynux1,yn|2+|ux,ynux,y1n|2, (9)

where ux,yn is the nth entry of the vector u, and the subtraction of “1” to x or y in the subscript corresponds to an entry of u that is respectively shifted in relation to ux,yn by one pixel in the x or y direction of the 2D image. The inversion using TV is thus given by [22], [23], [24]

u*=argminpMu22+αuTV (10)

where α > 0 is the regularization parameter. In the case of TV minimization, over-regularization may lead to cartoon-like images and rounding of complex boundaries into convex shapes.

In some cases, it is beneficial to promote both sparsity of the image in an alternative basis and TV minimization. In such cases, the optimization problem is as follows [15]:

u*=argminpMu22+μΦu1+αuTV. (11)

We will refer to the regularization described in Eq. (11), as TV-L1 regularization.

3. Adaptive anisotropic total variation

3.1. The functional

We would like to use a regularizer that is adapted to the image in such a way that it regularizes more along edges (level-lines of the image) and less across edges (in the direction of the gradient). This idea has been introduced for nonlinear scale-space flows by Weickert [34] in the anisotropic diffusion formulation. However there is no known functional associated with anisotropic diffusion, and it is therefore not trivial to include an anisotropic-diffusion operation in our inverse-problem formulation. A more recent study of Grasmair et al. [35] uses a similar adaptive scheme within a TV-type formulation. In the study of [26], a comprehensive theoretical and numerical analysis was performed for adaptive-anisotropic TV (A2TV). It was shown that stable structures can be non-convex and in addition can have high curvature on the boundaries. Illustrations of the stable sets characterize the TV and A2TV regularizers are shown in Fig. 2. The degree of anisotropy directly controls the degree of allowed nonconvexity and the upper bound on the curvature. We adopt the formulation of [26], in which the mathematical underpinnings of A2TV are described in detail.

Fig. 2.

Fig. 2

An illustration of sets which are stable for TV and A2TV – notice A2TV admits non-convex and highly curved functions, including ones which resemble arteries.

Let the A2TV functional be defined by

JA2TV(u)=ΩA(x)u(x)2dx=ΩAu(x)2dx (12)

where x = (x, y) ∈ Ω denotes position in 2D space, A(x)2×2 is a tensor or a spatially adaptive matrix and ∇A = A(x)∇ is an “adaptive gradient”. This functional is convex and can be optimized by standard convex solvers given the tensor A(x). We now turn to the issue of how this tensor is constructed to allow a good regularization of vessel-like structures.

3.1.1. Constructing the tensor A(x)

We assume to have a rough approximation of the image to be reconstructed. This can be done, for instance, by having an initial least-square non-regularized approximation or a standard TV reconstruction, as in Eq. (10). We thus have an initial estimation u0. The tensor A(x) determines the principle and the secondary directions of the regularization at each point and their magnitude. The construction of A(x) is performed using u0 according to the following principles: In regions in which u0 is relatively flat, i.e. ∇u0 almost vanishes, A(x) should resemble the identity matrix and have no preferred direction, thus leading to the conventional TV regularizer. In regions with dominant edges, A(x) should capture the principle axes of the edge. See Fig. 3 for an illustration of A(x).

Fig. 3.

Fig. 3

An illustration of the tensor A(x). At any point the tensor rotates and rescales the coordinate system in an image-driven manner. It assumes some approximation u0 of the data exists. The tensor is designed such that lower regularization is applied across edges (top left ellipse) whereas in flat regions regularization is applied in an isotropic manner (bottom right circle).

Mathematically, the tensor A(x) is defined by adapting the eigenvalues of a smoothed structure tensor of a smoothed image u0;σ (with a Gaussian kernel of standard deviation σ), defined by,

Jρ(u0;σ)=κρ*u0;σu0;σ, (13)

where κρ is a Gaussian kernel with a standard deviation of ρ, * denotes an element-wise convolution and ⊗ denotes an outer product. The Gaussian kernel's standard deviations σ and ρ are chosen in coordination with the noise level of the image and the smallest object resolution within the image. High standard deviation values might diminish the effect of small objects on the structure tensor while low values might lead to the unwanted scenario in which reconstruction errors have a significant effect on the structure tensor. The explicit expressions for Jρ in the 2D and 3D cases are respectively given by

Jρ(x)=κρ*ux;σ2(x)κρ*ux;σ(x)uy;σ(x)κρ*ux;σ(x)uy;σ(x)κρ*uy;σ2(x) (14)

and

Jρ(x)=κρ*ux;σ2(x)ux;σ(x)uy;σ(x)ux;σ(x)uz;σ(x)ux;σ(x)uy;σ(x)uy;σ2(x)uy;σ(x)uz;σ(x)ux;σ(x)uz;σ(x)uy;σ(x)uz;σ(x)uz;σ2(x). (15)

The structure tensor matrix has eigenvectors corresponding to the direction of the gradient and tangent at each point x; and eigenvalues corresponding to the magnitude of each direction. In order to preserve structure, we should change the relation between those eigenvalues so that for flat-like areas in the image we will smooth the image in an isotropic way, while for edge-like areas, we will perform more smoothing in the tangent direction rather the gradient one. For 2D, we begin by looking in the eigen-decomposition of the structure-tensor,

Jρ=VDV1 (16)

Where V is a matrix whose columns, v1,v22, are the eigenvectors of Jρ, and D is a diagonal matrix with eigenvalues in the diagonal,

D=μ100μ2;V=(v1|v2) (17)

Assuming μ1 ≥ μ2. The spatially adaptive matrix A(x), used in Eq. (12), is constructed from a modification of the eigenvalue matrix, denoted by D˜, as follows [34]:

A=VD˜V1, (18)

where D˜ is given by

D˜=c(μ1/μ1,avg;k)001. (19)

In Eq. (19), μ1,avg is the average value of μ1 across the image and c(· ; ·) is a function of two parameters defined as follows:

c(s;k)=1,s01expcm(sk)m,s>0 (20)

where the chosen values for the parameter are the ones recommended in Ref. [34]: cm = 3.31488 and m = 4. The parameter k ≤ 1 determines which regions in the image will be regularized anisotropically and is chosen based on the desired level of anisotropy in the reconstructed image, as shown in Sections 4 and 5. In pixels in which s ≪ k, i.e. μ1/μ1,avg ≪ k, we will obtain c ≈ 1, and D˜ will be reduced to the unitary matrix. Accordingly, for regions in which the image gradient is sufficiently small from the average image gradient, as regulated by k, the A2TV functional (Eq. (12)) is reduced to the standard, isotropic TV functional (Eq. (9)). In the rest of the image, where c is sufficiently smaller than 1, regularization is performed more strongly in the direction of the eigenvector v2, i.e. the direction in which the image gradient is smaller, thus enhancing the anisotropy in those regions.

It is worth noting that in the 3D case, assuming μ1 ≥ μ2 ≥ μ3, the only modification to the analysis above is that D˜ accepts the following form:

D˜=c(μ1/μ1,avg;k)000c(μ2/μ1,avg;k)0001. (21)

The structure of D˜ in Eq. (21) will be highly anisotropic for tube-like structures and will enforce variation-reducing regularization along the structure length (eigenvector v3), while maintaining low regularization in the cross-section plane (eigenvectors v1 and v2).

3.2. Reconstruction based on A2TV

The reconstruction based on the A2TV minimizes the following functional:

u*=argminuJA2TV(u)+λ2Mup22, (22)

where M is the model matrix and p is the acoustic pressure wave. The solution of which is done by the modified Chambolle-Pock projection algorithm [36] described in Appendix A.

In the process of minimization, the tensor A is initialized as the identity matrix for all x2, which reduces the A2TV energy to the TV one as it performs the diffusion isotropically. The tensor A is then updated according to the initial solution u0. This is repeated until numerical convergence is reached.

We note that while the energy of the A2TV is convex for a fixed tensor A(x), it is not convex when A(x) is adaptive and depends on the imaged object. Thereby, we do not have a mathematical proof of convergence. Nonetheless, it has been shown both in our work [26] and in Refs. [34], [35] that heuristically, both the image u and the tensor A(x) converge.

4. Numerical simulations

In this section, we demonstrate the performance of A2TV-based inversion for the circular detection geometry illustrated in Fig. 4. The simulations were performed on a 2D vascular image of a mouse retina, obtained via confocal microscopy. The vasculature image, shown in Fig. 5a, was represented over a square grid with a size of 256 × 256 pixels with pixel size of 0.1 × 0.1 mm. The projections were simulated over a 270-degree semi-circle with a radius of 4 cm that surrounded the object, in accordance with conventional OAT systems [4]. A magnification of four square regions of the image is shown in Fig. 5b.

Fig. 4.

Fig. 4

An illustration of the image and detector geometry used in the numerical simulations and experimental measurements in Sections 4 and 5.

Fig. 5.

Fig. 5

(a) The originating image on which all the reconstructions were performed and (b) a binary mask that was generated from it. The bottom panel (c and d) shows magnifications of 4 marked regions in the respective images in the top panel.

The reconstructions were performed using the conventional L2-based regularization-free approach (Eq. (5)) performed via LSQR, TV-L1 regularization (Eq. (11)), and the proposed A2TV regularization (Eq. (22)) method. Since the scaling of the model matrix M (Eq. (4)) depends on the exact implementation of its construction [32], we normalized M by 1160MM1 to assure that the regularization parameters are independent of the scaling of M. To assess the quality of the reconstructions, we used the mean absolute distance (MAD) given by the following equation:

MAD=1Nuorigu*1, (23)

where N is the number of pixels in the image.

Two cases were tested: In the first case, zero-mean Gaussian noise with a standard deviation of 0.6 times the maximum value of p was added to the projection. The number of projections was chosen to be 256, corresponding to the geometry found in state-of-the-art optoacoustic systems [37] and sufficient for the accurate reconstruction of the tested image in the noiseless case. In the second case, the number of projections was reduced to 32, which is half the number of projections used in low-end optoacoustic systems characterized by reduced lateral resolution [37]. Accordingly, 32 projections are insufficient for producing detailed optoacoustic images using conventional reconstruction techniques. In all the examples, the number of iterations was chosen to be sufficiently high to achieve convergence.

Fig. 6, Fig. 7 respectively show the reconstructions obtained using TV-L1 and A2TV, performed with 3000 iterations, for the cases of additive Gaussian noise. In both figures, 9 reconstructions are shown, corresponding to a scan in the regularization parameters. In the case of TV-L1, μ and α represent the strength of the L1 and TV regularization terms Eq. (11), whereas in the case of A2TV, λ represents the strength of the fidelity term in Eq. (22) with respect to the A2TV term and k determines the strength of the anisotropy, where lower values of k correspond to higher anisotropy. In the A2TV reconstructions, the standard deviations of the smoothing kernels (Eq. (14)) were σ = 1.5 pixels and ρ = 3 pixels. For all the reconstructions, the MAD values appear on the top-right corner of the image. In Fig. 8, we show a comparison between the regularization-free LSQR reconstruction (Fig. 8a) and the TV-L1 (Fig. 8b) and A2TV (Fig. 8c) of Figs. 6d and 7e, respectively, which correspond to the regularization parameters that achieved the lowest MAD values. The middle panel of the figure (Fig. 8d–f) shows a magnification of 4 patches taken from the reconstructions, whereas the bottom panel (Fig. 8g) presents a 1D slice taken over the yellow line in Fig. 8b and c. While both TV-L1 and A2TV significantly improved the reconstruction quality, in A2TV more of the noise-induced texture between the blood vessels could be removed without damaging the structure of the blood vessels, thus leading to a lower MAD.

Fig. 6.

Fig. 6

(a–i) The reconstruction of the image shown in Fig. 5a for the case of additive Gaussian noise using different parameters for the TV-L1 case. The reconstructions were performed with 3000 iterations.

Fig. 7.

Fig. 7

(a–i) The reconstruction of the image shown in Fig. 5a for the case of additive Gaussian noise using different parameters for the A2TV case. The reconstructions were performed with 3000 iterations.

Fig. 8.

Fig. 8

(a–c) The reconstruction of the image shown in Fig. 5a for the case of additive Gaussian noise using (a) LSQR, (b) TV-L1, and (c) A2TV. The bottom panel (d-f) shows magnifications of 4 marked regions in the respective images in the top panel. Both the TV-L1 and A2TV reconstructions were produced using 3000 iterations. (g) A 1D slice of the TV-L1 and A2TV reconstructions, corresponding to the vertical lines on the top panel, in comparison to a 1D slice taken from the original image.

The reconstructions for the case of 32 projections are presented in a similar manner to the case of noisy data. Fig. 9, obtained with 1000 iterations, and Fig. 10, obtained with 1500 iterations, respectively show the TV-L1 and A2TV reconstructions for a scan regularization parameters, whereas Fig. 11 shows a comparison between the LSQR and TV-L1 and A2TV reconstructions that achieved the lowest MAD (Figs. 9d and 10e, respectively). All the A2TV reconstructions were obtained with σ = 1.5 pixels and ρ = 1 pixel. Fig. 11 shows that both TV-L1 and A2TV eliminated the streak artifacts that appeared with in the LSQR reconstruction, where the lowest MAD was achieved by the TV-L1 reconstruction. Since both regularization methods eliminated the streak artifacts, the lower MAD achieved by TV-L1 is a result of its ability to better preserve texture within the blood vessels, whereas in the A2TV much of the blood-vessel texture was lost. Indeed, when examining the 1D slice in Fig. 11g, it is easy to see that the TV-L1 reconstruction captures the variations within the blood vessels better, whereas in the A2TV reconstruction, these variations are smoothed.

Fig. 9.

Fig. 9

(a–i) The reconstruction of the image shown in Fig. 5a for the case of under-sampled projection data using different parameters for the TV-L1 case. The reconstructions were performed with 1000 iterations.

Fig. 10.

Fig. 10

(a–i) The reconstruction of the image shown in Fig. 5a for the case of under-sampled projection data using different parameters for the A2TV case. The reconstructions were performed with 1500 iterations.

Fig. 11.

Fig. 11

(a–c) The reconstruction of the image shown in Fig. 5a for the case of under-sampled projection data using (a) LSQR, (b) TV-L1, and (c) A2TV. The bottom panel (d–f) is as in Fig. 8. The TV-L1 and A2TV reconstructions were produced using 1000 and 1500 iterations, respectively. (g) A 1D slice of the TV-L1 and A2TV reconstructions, corresponding to the vertical lines on the top panel, in comparison to a 1D slice taken from the original image.

In both the cases studied, A2TV exhibited a higher capability than TV-L1 to perform regularization without harming non-convex structures. Even in the case in which TV-L1 achieved a lower MAD, the higher ability of A2TV to preserve the fine details of the blood vessel morphology can be observed when comparing Fig. 11e and f. Additionally, in all the examples, one can observe that when TV-L1 was performed with a high level of TV regularization (bottom row in Fig. 6, Fig. 9) significant smearing of the blood vessels was observed. In contrast, in the A2TV reconstructions, the smearing owing to over-regularization (low values of λ) could be diminished by increasing the anisotropy in the regularization, i.e. reducing the value of k. For the lowest values of k, higher levels of regularization (low values of λ) created undesirable anisotropic vessel-like texture in the reconstructed images.

5. Experimental results

To further validate the suitability of A2TV regularization for OAT image reconstruction, we tested its performance on experimental data. The optoacoustic setup comprised an optical parametric oscillator (OPO), which produced nanosecond optical pulses with an energy 30 mJ at a repetition rate of 100 Hz and at a wavelength of λ = 680 nm. (SpitLight DPSS 100 OPO, InnoLas Laser GmbH, Krailling, Germany). The OPO pulses were delivered to the imaged object using a fiber bundle (CeramOptec GmbH, Bonn, Germany). Ultrasound detection was performed by a 256-element annular array (Imasonic SAS, Voray sur l’Ognon, France) with a radius of 4 cm, and an angular coverage of 270 degrees, comparable to the geometry shown in Fig. 4. The ultrasound detectors were cylindrically focused to a plane, approximating a 2D imaging scenario.

The imaged object was a transparent agar phantom which contained four intersecting hairs. A photo of the phantom is shown in Fig. 12a, where Fig. 12b shows 4 magnified parts of the phantom. The phantom preparation involved mixing 1.3% (by weight) agar powder (Sigma–Aldrich, St. Louis, MO) in boiling water and pouring the solution in a cylindrical mold until solidification. To assure that all four hair strands lie approximately in the same plane, we first prepared a clear cylindrical agar phantom, on which the hairs were placed; additional agar solution was then poured on the structure to seal the hairs.

Fig. 12.

Fig. 12

(a) The originating image on which all the reconstructions were performed and (b) a binary mask that was generated from it. The bottom panel (c and d) shows magnifications of 4 marked regions in the respective images in the top panel.

The optoacoustic image was reconstructed from the measured data using TV-L1 regularization with 3000 iterations and A2TV regularization with 6000 iterations, σ = 1.5 pixels, and ρ = 1 pixel. Fig. 13, Fig. 14 respectively show the images obtained via TV-L1 and A2TV reconstructions for a set of regularization parameters. As in the previous section, over-regularization in the TV-L1 led to significant loss of structure, whereas for the A2TV the ability to capture the image morphology under strong regularization (low λ) was improved when the anisotropy was increased via low values of k.

Fig. 13.

Fig. 13

(a–i) The reconstruction of the image shown in Fig. 12a for the case of experimental data using different parameters for the TV-L1 case. The reconstructions were performed with 3000 iterations.

Fig. 14.

Fig. 14

(a–i) The reconstruction of the image shown in Fig. 12a for the case of experimental data using different parameters for the A2TV case. The reconstructions were performed with 6000 iterations.

Fig. 15 compares between the unregularized LSQR reconstruction (Fig. 15a) and the TV-L1 (Fig. 15b) and A2TV (Fig. 15c) reconstructions, respectively taken from Figs. 13e and 14e . The second row in the figure (Fig. 15d–f) shows a magnification of 4 patches from the three reconstructions of the top row (Fig. 15a–c), whereas the bottom panel (Fig. 15g) shows a 1D slice from the vertical yellow line in Fig. 15b and c. To allow an easy comparison, the 1D slices were normalized by their maximum values. We note that the negative values in the reconstruction are a common result of the limited detection bandwidth of the ultrasound detector [6]. The figure clearly shows that the A2TV reconstruction obtained the highest image quality, in particular for the weak hair structure at the image bottom. In particular, the 1D slice shows that the bottom hairs that appear around the position of 1 cm achieve a peak-to-peak signal over 4-times higher in the A2TV reconstruction than in the TV-L1 reconstruction.

Fig. 15.

Fig. 15

(a–c) The reconstruction of the image shown in Fig. 12a for the case of experimental data using (a) LSQR, (b) TV-L1, and (c) A2TV. The TV-L1 and A2TV reconstructions were produced using 3000 and 6000 iterations, respectively. (g) A 1D slice of the TV-L1 and A2TV reconstructions, corresponding to the vertical lines on the top panel.

6. Discussion

In this work, a novel regularization framework was developed for OAT image reconstruction. The new framework is based on an A2TV cost function, which represents a generalization of the conventional TV functional, that is compatible with objects that possess complex, nonconvex boundaries. In contrast to TV, the A2TV cost function is an adaptive functional whose form depends on the characteristics of the image. When using A2TV, one first roughly defines the boundaries of the objects in the image. Then, one uses these boundaries to determine the directions in which the gradients are applied on the image. Similar to TV, A2TV is most appropriate as a regularizer in optoacoustic image reconstruction when the images are comprised of objects with well-defined boundaries. However, A2TV is useful also when these boundaries are incompatible with TV regularization due to their complexity. A common category of optoacoustic images that fits the above description is blood-vessel images. Since blood is a major source of contrast in optoacoustic imaging, OAT systems often produce images that are dominated by a complex structure of interwoven blood vessels. In particular, high-resolution images of the micro-vasculature are characterized by a complex network of arterioles, venues, and capillaries with extremely complex, nonconvex boundaries.

In our current implementation, A2TV required setting 4 parameters: σ, ρ, λ, and k. The first two parameters, σ and ρ, determined the image smoothing used in calculating the image gradients. While smoothing reduces the noise, thus limiting the effect of reconstruction errors on the detection of the principle axes, only moderate smoothing may be used without the risk of merging the boundaries of different objects in the image. Therefore, in all our examples, the smoothing was performed with Gaussian kernels whose standard deviations, σ and ρ, were 3 pixels or less. While the choice of σ and ρ depended on the level of noise and artifacts in the regularization-free reconstruction, the parameters λ and k were determined by the amount of regularization desired in the reconstructed image, where k determined the amount of anisotropy and λ determined the strength of the regularizer. In our examples, performing over-regularization (λ = 0.0001) led to image smearing in the isotropic case (k = 1) and vessel-like artifacts in the case of high anisotropy (k = 0.01). While such artifacts are undesirable, it is worth noting that their presence did not obscure the underlying image morphology, whereas over-regularization in TV-L1 led to loss of image details.

We compared the performance of the A2TV algorithm TV-L1 algorithm for both numerically simulated data and experimental data. In the numerical simulations, an image of blood vessels was reconstructed for the cases additive noise and sparse sampling of the projection data. In the experimental reconstructions, the imaged object was four intersecting hair strands whose structure emulated the morphology of blood vessels. In both the numerical and experimental examples, A2TV demonstrated a higher ability to preserve the blood-vessel morphology for high regularization parameters. Nonetheless, in the numerical example in which the reconstructions were performed with a low number of projections, TV-L1 regularization achieved a lower MAD owing to the loss of texture in the A2TV reconstruction. In the experimental example, A2TV led to a considerable improvement in the contrast in the weak structures of the image in comparison to the TV-L1 reconstruction.

The reconstruction performance demonstrated in this work suggests that A2TV may be a useful tool for improving the ability of optoacoustic systems to perform vasculature imaging, which is a major application in the field. Since the texture within the blood vessels is affected by the random distribution of the red-blood cells within the blood vessels, its elimination by A2TV may be considered as an acceptable price for better visualization of the blood-vessel morphology. In deep-tissue OAT systems, sub-millimeter vasculature imaging has been suggested as a potential diagnostic tool and has been demonstrated in the human extremities [38], [39] and breast [40]. We note that while some OAT systems can also produce images of the tissue bulk, characterized by low frequencies and representative of the density of the microvasculature and fluence map, such systems require transducers capable of detecting ultrasound frequencies considerably below 1 MHz [6]. In high-resolution OAT systems, operating at frequencies above 1 MHz and capable of reaching resolutions better than 100 μm [41], [42], only signals from blood vessels may be detected. Finally, when performing optoacoustic imaging at resolutions better than 10 μm, e.g. using raster-scan optoacoustic mesoscopy (RSOM) [43], the image is dominated by the microvasculare and generally lacks any bulk component associated with blood.

We note that the formalism of A2TV, in which the structure tensor matrix analysis is performed via eigenvalue decomposition Eq. (13) enables its adaptation to higher image dimensions. TV regularization has been recently performed in 4D optoacoustic reconstruction that included 3 spatial dimensions and time [44]. Since the variation of the pixels in time is generally different than the one space, it may be expected that A2TV can further improve image fidelity in such cases.

Conflicts of interest

The authors declare that there are no conflicts of interest related to this article.

Acknowledgement

We are thankful to Anna M. Randi for providing information on the image used in Section 4. This work was supported by the Technion Ollendorff Minerva Center.

Biographies

graphic file with name fx1.jpg

Shai Biton received his B.Sc. degree from the Department of Electrical Engineering from Ort Braude, Karmiel, in 2011. Currently he pursues his MSc degree in Electrical Engineering from the Technion – Israel Institute of Technology, Haifa. His current research interests are in medical imaging, thermal imaging, image reconstruction and deep learning.

graphic file with name fx2.jpg

Nadav Yehonatan Arbel is an M.Sc. Candidate in the Viterbi Faculty of Electrical Engineering at the Technion – Israel Institute of Technology, where he formerly obtained a double B.Sc. in Electrical Engineering and Physics. His research interests include Matching of 3D Shapes, Medical Signal Processing and Optics.

graphic file with name fx3.jpg

Gilad Drozdov received his B.Sc. degree from the Department of Electrical Engineering from the Technion-Israel Institute of Technology, Haifa, in 2012, where he currently pursues his MSc degree in Electrical Engineering. His current research interests are in optoacoustic imaging, acoustic transducers modeling and image reconstruction.

graphic file with name fx4.jpg

Guy Gilboa received his PhD from the Electrical Engineering Department, Technion – Israel Institute of Technology in 2004. He was a postdoctoral fellow at UCLA, hosted by Prof. Stanley Osher (2004–2007). He later joined 3DV, a start-up company developing 3D sensors, which was acquired by Microsoft. He worked at Microsoft (Kinect project) and at Philips Healthcare as a senior researcher in the field of medical imaging. In 2013 he joined the Electrical Engineering Dept., Technion, for a tenure-track position. He is the main author of some highly cited papers on topics such as image sharpening and denoising, nonlocal operators theory for image processing and nonlinear spectral theory. He received several prizes including the Eshkol prize by the Ministry of Science, Technion outstanding Ph.D. thesis, Vatat scholarship and Gutwirth prize. He serves at the editorial boards of the journals IEEE Signal Processing Letters, Journal of Mathematical Imaging and Vision and Journal of Computer Vision and Image Understanding.

graphic file with name fx5.jpg

Amir Rosenthal received his B.Sc. and Ph.D. degrees, both from the Department of Electrical Engineering, the Technion-Israel Institute of Technology, Haifa, in 2002 and 2006, respectively. From 2009 to 2010, he was a Marie Curie fellow at the Cardiovascular Research Center at Massachusetts General Hospital and Harvard Medical School, Boston, MA, and from 2010 to 2014 he was a group leader at the Institute of Biological and Medical Imaging (IBMI), Technische Universität München and Helmholtz Zentrum München, Germany. Since 2015 he has been an assistant professor at the Andrew and Erna Viterbi Faculty of Electrical Engineering at the Technion - Israel Institute of Technology. His research interests include optoacoustic imaging, interferometric sensing, intravascular imaging, inverse problems, and optical and acoustical modeling.

Appendix A. Solving the generalized ROF model

A2TV based reconstruction amounts to solving the following generalized ROF model,

u*=argminuXE(u)=argminuXJA2TV(u)+λ2Mup22. (A.1)

Where M is the model matrix, p is the acoustic pressure wave and u* is the reconstructed image. The algorithm used for inversion is Chambolle-Pock, which solves the following primal form [36]:

argminuXF(Ku)+G(u). (A.2)

Where, in our case,

F(y)=y1,G(u)=λ2Mup22K=A=gradA,K*=AT=divA. (A.3)

Let X and Y be two finite-dimensional real vector spaces and K, a linear operator K : X → Y and its hermitian adjoint K* : Y → X. The appropriate saddle point problem is,

argminuXargmaxzY<Ku,z>+G(u)F*(z). (A.4)

Where z ∈ Y is the dual variable. In our case,

argminuXargmaxzY<Au,z>+λ2Mup22δP(z), (A.5)

where

F*(z)=δP(z), (A.6)

and

δP(z)=0,zP,zP,P={zY   |   z1}. (A.7)

But, because of the appearance of the model matrix, we shall adjust the saddle point problem to

argminuXargmaxzY,qQ<Au,z>+<Mup,q>12λq22δP(z)=argminuXargmaxzY,qQ<Au,z>+<Mu,q><p,q>12λq22δP(z)=argminuXargmaxzY,qQ<MAu,qz><p,q>12λq22δP(z), (A.8)

using

λ2Mup2=argmaxqQ<Mup,q>12λq2. (A.9)

Here, the original saddle point problem is transformed into,

argminuXargmaxzY,qQ<Ku,qz>+G(u)F*qz. (A.10)

Where

F*qz=<p,q>+12λq22+δP(z),G(u)=0,K=KAKA=MAK*=KA*KA*=MTAT. (A.11)

Now, in order to use the algorithm, we shall use the proximal operator of F* and G,

Proxσ,F*q˜z˜=q˜σp1+σλ,z˜max1,|z˜|pointwise,Proxτ,Gu˜=u˜ (A.12)

Algorithm 1

The Chambolle-Pock algorithm.

1: procedure CHAMBOLLE-POCKProxσn,F*(·),Proxτn,G(·),K,K*
2:  Initialize γ, τ0
3:  σ0 ⟵ 1/(τ0L2)
4:  Initialize u0, q0, z0
5:  i ⟵ 0
6:  whilei < Ndo
7:    [qn+1,zn+1]Proxσn,F*(qn+σnKA(un),zn+xxxxxxσnK(un))
8:    [un+1,zn+1]Proxτn,G(unτnK*([qn+1,zn+1]))
9:    θn+11/1+2γτn
10:    τn+1 ⟵ θn+1τn
11:    σn+1 ⟵ σn/θn+1
12:    Update tensor A according to the new image un+1 and its dependents K* & K
13:  end while
14:  return uN
15: end procedure

The algorithm is given in Fig. 1, where in our case, we normalize the model matrix M by 1160MM1, where MM1 is an approximation to the model matrix M Lipchitz constant. L = LM + L = 160 + 8 =168, γ = 0.7λ and τ0 = 0.5

References

  • 1.Ntziachristos V. Going deeper than microscopy: the optical imaging frontier in biology. Nat. Methods. 2010;7:603. doi: 10.1038/nmeth.1483. [DOI] [PubMed] [Google Scholar]
  • 2.Wang L.V., Hu S. Photoacoustic tomography: in vivo imaging from organelles to organs. Science. 2012;335:1458–1462. doi: 10.1126/science.1216210. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Ntziachristos V., Razansky D. Molecular imaging by means of multispectral optoacoustic tomography (msot) Chem. Rev. 2010;110:2783–2794. doi: 10.1021/cr9002566. [DOI] [PubMed] [Google Scholar]
  • 4.Taruttis A., Ntziachristos V. Advances in real-time multispectral optoacoustic imaging and its applications. Nat. Photonics. 2015;9:219. [Google Scholar]
  • 5.Wang L.V. Multiscale photoacoustic microscopy and computed tomography. Nat. Photonics. 2009;3:503. doi: 10.1038/nphoton.2009.157. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Rosenthal A., Ntziachristos V., Razansky D. Acoustic inversion in optoacoustic tomography: a review. Curr. Med. Imaging Rev. 2013;9:318–336. doi: 10.2174/15734056113096660006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Buehler A., Herzog E., Razansky D., Ntziachristos V. Video rate optoacoustic tomography of mouse kidney perfusion. Optics Lett. 2010;35:2475–2477. doi: 10.1364/OL.35.002475. [DOI] [PubMed] [Google Scholar]
  • 8.Deán-Ben X.L., Ford S.J., Razansky D. High-frame rate four dimensional optoacoustic tomography enables visualization of cardiovascular dynamics and mouse heart perfusion. Sci. Rep. 2015;5:10133. doi: 10.1038/srep10133. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Xu Y., Wang L.V. Time reversal and its application to tomography with diffracting sources. Phys. Rev. Lett. 2004;92:033902. doi: 10.1103/PhysRevLett.92.033902. [DOI] [PubMed] [Google Scholar]
  • 10.Xu M., Wang L.V. Universal back-projection algorithm for photoacoustic computed tomography. Phys. Rev. E. 2005;71:016706. doi: 10.1103/PhysRevE.71.016706. [DOI] [PubMed] [Google Scholar]
  • 11.Schulze R.J., Scherzer O., Zangerl G., Holotta M., Meyer D., Handle F., Nuster R., Paltauf G. On the use of frequency-domain reconstruction algorithms for photoacoustic imaging. J. Biomed. Optics. 2011;16:086002. doi: 10.1117/1.3605696. [DOI] [PubMed] [Google Scholar]
  • 12.Rosenthal A. Algebraic determination of back-projection operators for optoacoustic tomography. Biomed. Opt. Express. 2018;9:5173–5193. doi: 10.1364/BOE.9.005173. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Calvetti D., Morigi S., Reichel L., Sgallari F. Tikhonov regularization and the l-curve for large discrete ill-posed problems. J. Comput. Appl. Math. 2000;123:423–446. [Google Scholar]
  • 14.Yuan Z., Wu C., Zhao H., Jiang H. Imaging of small nanoparticle-containing objects by finite-element-based photoacoustic tomography. Optics Lett. 2005;30:3054–3056. doi: 10.1364/ol.30.003054. [DOI] [PubMed] [Google Scholar]
  • 15.Han Y., Tzoumas S., Nunes A., Ntziachristos V., Rosenthal A. Sparsity-based acoustic inversion in cross-sectional multiscale optoacoustic imaging. Med. Phys. 2015;42:5444–5452. doi: 10.1118/1.4928596. [DOI] [PubMed] [Google Scholar]
  • 16.Provost J., Lesage F. The application of compressed sensing for photo-acoustic tomography. IEEE Trans. Med. Imaging. 2009;28:585–594. doi: 10.1109/TMI.2008.2007825. [DOI] [PubMed] [Google Scholar]
  • 17.Guo Z., Li C., Song L., Wang L.V. Compressed sensing in photoacoustic tomography in vivo. J. Biomed. Optics. 2010;15:021311. doi: 10.1117/1.3381187. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Liang D., Zhang H.F., Ying L. Compressed-sensing photoacoustic imaging based on random optical illumination. Int. J. Funct. Inform. Pers. Med. 2009;2:394–406. [Google Scholar]
  • 19.Sun M., Feng N., Shen Y., Shen X., Ma L., Li J., Wu Z. Photoacoustic imaging method based on arc-direction compressed sensing and multi-angle observation. Optics Express. 2011;19:14801–14806. doi: 10.1364/OE.19.014801. [DOI] [PubMed] [Google Scholar]
  • 20.Meng J., Wang L.V., Ying L., Liang D., Song L. Compressed-sensing photoacoustic computed tomography in vivo with partially known support. Optics Express. 2012;20:16510–16523. [Google Scholar]
  • 21.Liu X., Peng D., Guo W., Ma X., Yang X., Tian J. Compressed sensing photoacoustic imaging based on fast alternating direction algorithm. J. Biomed. Imaging. 2012;2012:12. doi: 10.1155/2012/206214. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Wang K., Sidky E.Y., Anastasio M.A., Oraevsky A.A., Pan X. Vol. 7899. International Society for Optics and Photonics; 2011. Limited data image reconstruction in optoacoustic tomography by constrained total variation minimization; p. 78993U. (Photons Plus Ultrasound: Imaging and Sensing 2011). [Google Scholar]
  • 23.Yao L., Jiang H. Enhancing finite element-based photoacoustic tomography using total variation minimization. Appl. Optics. 2011;50:5031–5041. [Google Scholar]
  • 24.Yao L., Jiang H. Photoacoustic image reconstruction from few-detector and limited-angle data. Biomed. Optics Express. 2011;2:2649–2654. doi: 10.1364/BOE.2.002649. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Wang J., Zhang C., Wang Y. A photoacoustic imaging reconstruction method based on directional total variation with adaptive directivity. Biomed. Eng. Online. 2017;16:64. doi: 10.1186/s12938-017-0366-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Biton S., Gilboa G. 2018. Adaptive Anisotropic Total Variation — A Nonlinear Spectral Analysis.arXiv:1811.11281 [Google Scholar]
  • 27.Andreu F., Ballester C., Caselles V., Mazón J.M. Minimizing total variation flow. Differ. Integral Equ. 2001;14:321–360. [Google Scholar]
  • 28.Peura M., Iivarinen J. Efficiency of simple shape descriptors. Proceedings of the Third International Workshop on Visual Form, vol. 443; Citeseer; 1997. p. 451. [Google Scholar]
  • 29.Kruger R.A., Liu P., Fang Y., Appledorn C.R. Photoacoustic ultrasound (paus)-reconstruction tomography. Med. Phys. 1995;22:1605–1609. doi: 10.1118/1.597429. [DOI] [PubMed] [Google Scholar]
  • 30.Blackstock D.T. John Wiley & Sons; 2000. Fundamentals of Physical Acoustics. [Google Scholar]
  • 31.Paltauf G., Viator J., Prahl S., Jacques S. Iterative reconstruction algorithm for optoacoustic imaging. J. Acoust. Soc. Am. 2002;112:1536–1544. doi: 10.1121/1.1501898. [DOI] [PubMed] [Google Scholar]
  • 32.Rosenthal A., Razansky D., Ntziachristos V. Fast semi-analytical model-based acoustic inversion for quantitative optoacoustic tomography. IEEE Trans. Med. Imaging. 2010;29:1275–1285. doi: 10.1109/TMI.2010.2044584. [DOI] [PubMed] [Google Scholar]
  • 33.Paige C.C., Saunders M.A. Lsqr: an algorithm for sparse linear equations and sparse least squares. ACM Trans. Math. Softw. 1982;8:43–71. [Google Scholar]
  • 34.Weickert J. vol. 1. Teubner Stuttgart; 1998. (Anisotropic Diffusion in Image Processing). [Google Scholar]
  • 35.Grasmair M., Lenzen F. Anisotropic total variation filtering. Appl. Math. Optim. 2010;62:323–339. [Google Scholar]
  • 36.Chambolle A., Pock T. A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vis. 2011;40:120–145. [Google Scholar]
  • 37.Dima A., Burton N.C., Ntziachristos V. Multispectral optoacoustic tomography at 64, 128, and 256 channels. J. Biomed. Optics. 2014;19:036021. doi: 10.1117/1.JBO.19.3.036021. [DOI] [PubMed] [Google Scholar]
  • 38.Wray P., Lin L., Hu P., Wang L.V. Photoacoustic computed tomography of human extremities. J. Biomed. Optics. 2019;24:026003. doi: 10.1117/1.JBO.24.2.026003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Matsumoto Y., Asao Y., Yoshikawa A., Sekiguchi H., Takada M., Furu M., Saito S., Kataoka M., Abe H., Yagi T. Label-free photoacoustic imaging of human palmar vessels: a structural morphological analysis. Sci. Rep. 2018;8:786. doi: 10.1038/s41598-018-19161-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Toi M., Asao Y., Matsumoto Y., Sekiguchi H., Yoshikawa A., Takada M., Kataoka M., Endo T., Kawaguchi-Sakita N., Kawashima M. Visualization of tumor-related blood vessels in human breast by photoacoustic imaging system with a hemispherical detector array. Sci. Rep. 2017;7:41970. doi: 10.1038/srep41970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Deán-Ben X.L., López-Schier H., Razansky D. Optoacoustic micro-tomography at 100 volumes per second. Sci. Rep. 2017;7:6850. doi: 10.1038/s41598-017-06554-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Gateau J., Chekkoury A., Ntziachristos V. Ultra-wideband three-dimensional optoacoustic tomography. Optics Lett. 2013;38:4671–4674. doi: 10.1364/OL.38.004671. [DOI] [PubMed] [Google Scholar]
  • 43.Omar M., Schwarz M., Soliman D., Symvoulidis P., Ntziachristos V. Pushing the optical imaging limits of cancer with multi-frequency-band raster-scan optoacoustic mesoscopy (rsom) Neoplasia. 2015;17:208–214. doi: 10.1016/j.neo.2014.12.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Özbek A., Deán-Ben X.L., Razansky D. Optoacoustic imaging at kilohertz volumetric frame rates. Optica. 2018;5:857–863. doi: 10.1364/OPTICA.5.000857. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Photoacoustics are provided here courtesy of Elsevier

RESOURCES