Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Feb 1.
Published in final edited form as: J Math Imaging Vis. 2019 Nov 19;62:189–205. doi: 10.1007/s10851-019-00929-5

Sparse reconstruction of log-conductivity in current density impedance tomography

Madhu Gupta *, Rohit Kumar Mishra , Souvik Roy
PMCID: PMC7347294  NIHMSID: NIHMS1544096  PMID: 32647406

Abstract

A new non-linear optimization approach is proposed for the sparse reconstruction of log-conductivities in current density impedance imaging. This framework comprises of minimizing an objective functional involving a least squares fit of the interior electric field data corresponding to two boundary voltage measurements, where the conductivity and the electric potential are related through an elliptic PDE arising in electrical impedance tomography. Further, the objective functional consists of a L1 regularization term that promotes sparsity patterns in the conductivity and a Perona-Malik anisotropic diffusion term that enhances the edges to facilitate high contrast and resolution. This framework is motivated by a similar recent approach to solve an inverse problem in acousto-electric tomography. Several numerical experiments and comparison with an existing method demonstrate the effectiveness of the proposed method for superior image reconstructions of a wide-variety of log-conductivity patterns.

Keywords: Inverse problems, PDE-constrained optimization, proximal methods, edge-enhancement, sparsity patterns, current density impedance imaging

MSC: 35R30, 49J20, 49K20, 65M08, 82C31

1. Introduction

Electrical impedance tomography (EIT) is an imaging modality, where one attempts to recover the conductivity of a body from the boundary measurement of current and voltage [14]. The underlying inverse problem is highly ill-posed and non-linear yet very important due to its wide range applications in the fields such as medical imaging [54] and engineering [27, 53]. The mathematical formulation of the EIT inverse problem is given by the following conductivity equation

(σ(x)u(x))=0xΩ,σ(x)uν(x)=f(x)xΓ, (1)

where Ωn is a convex and bounded domain with Lipschitz boundary and Γ is the boundary of Ω. In this model, σ is the electrical conductivity, u represents the electric potential and f is the current applied to the boundary.

The reconstructions obtained through the EIT setup usually have high contrast but limited spatial resolution [48]. On the other hand, reconstructions obtained through ultrasound imaging has very high resolution but limited contrast [3, 42]. In recent years, attempts have been made to combine multiple imaging modalities to obtain image reconstructions with both high contrast and high resolution. This led to the emergence of hybrid imaging methods that belong to class of coupled-physics imaging modalities to generate images of superior quality. One of such imaging methods, known as current density impedance imaging (CDII) combines the classical EIT setup with magnetic resonance (MR) scanning [19, 36]. It is alternatively known as magnetic resonance EIT (MREIT). Current or voltage is applied through the electrodes, which give rise to an interior electric field and the corresponding generated magnetic field, represented as B = (Bx, By, Bz), is measured by the MR scanner. The corresponding inverse problem is to solve for the conductivity σ from Bz using the well-known iterative Harmonic Bz-algorithm [17, 30]. Convergence of the harmonic Bz algorithm has been well-studied [17, 30, 31]. In particular, it has been shown that for small contrast values of the target conductivity, the harmonic Bz-algorithm is stable and convergent, provided we have a good initial guess [30]. Thus, it is not clear that one can recover good quality images for high contrast objects through Harmonic Bz-algorithm.

An alternate approach to solve the CDII inverse problem is to use the knowledge of interior electric field, which is obtained from the magnetic field. Correspondingly, the magnitude of the interior electric field is also determined [47, 49], which is given by

H(σ(x))=σ(x)|u(x)|,xΩ. (2)

The formulation of reconstruction problem is as follows: Given the boundary data f for, possibly, several choices of boundary patterns and the corresponding interior measurement data H, find the conductivity distribution σ. In this framework, we use the internal function H(σ) to replace σ in the EIT equation (1) to get the following nonlinear equation

(H|u|u)=0 in Ω,H|u|uν=f on Γ. (3)

For the CDII inverse problem, the solution to the boundary value problem (3) is crucial but it is difficult to use it in practice because of its highly nonlinear behaviour and also because the data represented by the measured values of H enter as a coefficient of the differential model [48]. Even with the additional measurements, analysis and application of the 1-Laplacian relies on an iterative localized algorithm, wherein one considers an approximation of the CDII problem. This subsequently led to several computational approaches in solving the CDII inverse problem. In [23], it was proved that the linearized problem is elliptic and hence solvable, if there are at least n set of measurements {Hi(σ)}i=1n and corresponding to n boundary data {fi}i=1n such that ∇ui and ∇uj are nowhere collinear for ij. It has been shown in [26] that the solution of the above 1-Laplacian equation with the Neumann boundary condition is non-existent unless additional measurements with different boundary current patterns are used. Recovery of isotropic conductivity in regions where the magnetic field is transversal using two internal current distributions was done using an explicit local formula [28]. Moreover, using the information of two internal current distributions, the authors in [26] uniquely determine the singular support of the conductivity function. In [34], the authors showed that the conductivity in the planer domain can be recovered from a single voltage-current on a part of boundary and the magnitude of one interior current density. In the same article, they also provide sufficient conditions on Dirichlet boundary data to guarantee unique recovery of conductivity. In [35], the recovery of Hölder continuous conductivities have been establised for domains with connected boundary from the interior measurement of the magnitude of one current density. Determination of isotropic conductivity variations from measurements of two current density vector fields was studied in [19]. In [51], authors showed the recovery of planar conductivities by solving the 1-Laplace equation with partial boundary data. The authors in [32, 8, 9] present explicit reconstruction formulae for recovering the conductivity distribution from multiple interior measurements in two and higher dimensions.

The well-known numerical reconstruction algorithm using the internal current distribution is an iterative J-substitution algorithm which was first introduced by [24] and subsequently considered in other works, see for e.g., in [21, 25, 35, 36]. It has been shown that the J-substitution algorithm is able to reconstruct the conductivity with high resolution [25, 26]. Another numerical reconstruction iterative method is the regularized D-bar method [22] that provides images with high resolution. In [33], the authors use an alternating split Bregman algorithm for solving a minimization problem related to the energy functional corresponding to the 1-Laplacian equation (3). Also, in [20], Picard and Newton type algorithms are implemented to solve the 1-Laplacian problem. But there is not enough evidence to suggest that these existing algorithms (linearized or localized iterative methods) can provide high contrast images, specially for objects with holes or inclusions, which are inherent to CDII reconstructions.

The CDII inversion problem can be viewed as an inverse problem of estimating the conductivity parameter from the conductivity partial differential equation (PDE) (1). A robust way of solving this inverse problem is to formulate it as an optimal control problem, where the condiuctivity parameter is the control variable that drives the interior electric field close to the measured value, with dynamics governed by the conductivity PDE. Such optimal control methods has been used previously in the context of ultrasonically-induced Lorentz force electrical impedance tomography [5], magnetoacoustic tomography [6, 7, 38] and acousto-electric tomography (AET) [1, 41]. In [5], the authors use an optimal control framework to recover the conductivity distribution from the measurements of current induced by static magnetic field through the Lorentz force. They solve the optimal control problem using an orthogonal field method. In [6, 7, 38], the authors use an optimal control approach to reconstruct conductivity distribution of biological tissue from measurements of the Lorentz force induced tissue vibration. In [1, 41], the authors reconstruct log-conductivity in acousto-electric tomography (AET) using an optimal control formulation based on the theory of anisotropic diffusion to potentially obtain reconstructions with high resolution and contrast. The results obtained in these papers demonstrated that such optimal control frameworks are robust and accurate for imaging modalities arising through a partial differential equation (PDE).

In this paper, we consider a similar optimization framework developed in [41] for reconstructing the log-conductivity in CDII. We formulate a minimization problem, where given interior electric field intensity data, we aim at determining the variation in conductivity from a known background conductivity. We, further, assume that this variation demonstrates a sparsity pattern. Such patterns arise frequently in several tomographic imaging scenarios, for e.g. in blood vessel tomographic reconstructions [29]. This is incorporated in our model through a L2L1 regularization term in our objective functional. To obtain sharp edges and, thus, improve spatial resolution of the reconstructed images, we use a Perona-Malik anisotropic diffusion filtering term in our functional. The resulting optimality system gives rise to an elliptic adjoint equation with a L2 source term. Classical cell-nodal finite difference schemes are not applicable for solving such equations. We, thus, use a averaged cell-nodal scheme to solve such equations. Finally, we solve the optimization problem using a variable inertial proximal scheme that efficiently handles the non-differentiable terms in the objective functional. We demonstrate through several examples that our method can be used to obtain superior quality reconstructions for objects with holes and inclusions. In this context we would like to remark that our framework can also be used for obtaining superior reconstruction of anisotropic conductivity distributions, in ultrasonically-induced Lorentz force electrical impedance tomography and magnetoacoustic tomography. Further, our proposed framework can also be used in other hybrid imaging modalities like quantitative photoacoustic tomography [16] and quantitative thermoacoustic tomography [4] to obtain better reconstructions.

The article is organized as follows: In the Section 2, we formulate the minimization problem for the CDII. In the Section 3, we present some theoretical results about our optimization problem. We also characterize the optimality system. The variable inertial proximal scheme and the averaged cell-nodal schemes to solve the optimization problem are discussed in Section 4. In the Section 5, we present simulation results of our CDII framework and compare them with the reconstructions obtained using the Picard scheme proposed in [20], which validate our framework for CDII and demonstrate the effectiveness of our method to reconstruct wide variety of objects with corners, holes and inclusions. A section on conclusions completes our work.

2. A minimization problem

We consider the conductivity equation in 2 arising in EIT

(eσ(x,y)u(x,y))=0 in Ω,u(x,y)|Γ=fD(x,y), (4)

where Ω2 is bounded, Γ is the boundary of Ω, eσ is the conductivity coefficient and uHfD1(Ω)={uH1(Ω):u=fD on Γ} is the electric potential.

We assume that σ is a sparse conductivity coefficient which we want to recover, given the fact that the conductivity of the background is 1. The conductivity equation (4) can also be written as

L(u,σ,fD)=0,

where σ(x,y)Lad={σH01(Ω):σlσ(x,y)σu,(x,y)Ω},σu>0 and σl=12σt.

We consider an optimization-based approach for reconstructing σ given H1(σ), H2(σ), where

H(σ)=eσ|u|

is the interior electric field corresponding to the voltage potential u. We consider the following cost functional

J(σ,u1,u2)=j=12αj2Ω(Hj(x,y)Hjδ(x,y))2dxdy+β2σL2(Ω)2+γσL1(Ω)+δ2Ωlog(1+|σ(x,y)|2)dxdy (5)

where u1, u2 satisfy (4) with boundary data fD1,fD2. We now consider the following minimization problem

minσJ(σ,u1,u2), s.t. L(u1,σ,fD1)=0,L(u2,σ,fD2)=0. (P)

The term γσL1(Ω),γ>0 in the functional, defined in (5), implements a L1 regularization of the minimization problem that promotes sparsity patterns in the reconstruction of conductivity. Such a regularization method mirrors the well known compressed-sensing technique; see [11]. In recent past, optimal control with L1 cost functionals has become a topic of major interest [50], because one obtains sparse controls through this procedure, which finds numerous applications. The motivation for sparse log-conductivity patterns is based on the assumption that the background conductivity is known to be 1 in a substantial part of the domain Ω after normalization and varies considerably from this value in correspondence to different kind of objects present within the domain.

The combined L2-L1 regularization allows for the reconstruction of conductivity, and thus the imaging of, possibly, irregular objects inside Ω. This does not serve the ultimate goal of reconstructing objects like tissues in medical imaging, which are more regular, save for the edges that eventually define them. We infuse this additional aprior knowledge into our model through the last term in our functional (5) that, commonly, appears in the field of anisotropic diffusion. Such a term plays an important role in dampening image noise while keeping significant parts of the image content such as edges and other anatomical details that are of utmost importance in the interpretation of the image. Anisotropic diffusion means non-uniform diffusion in different directions. The regions where |∇σ| is very small corresponds to noise and thus, the process of smoothening occurs. At the edges or singularities of an object, where the value of |∇σ| is large, there is a small amount of smoothening and this preserves the edges. A standard technique to implement anisotropic diffusion, in order to obtain a good contrast, is to use a total variation (TV) regularization [13, 43]. But, this regularization method gives rise to a non-differentiable term in the functional (5), thus requiring more sophisticated optimization algorithms. On the other hand, anisotropic diffusion is inherent to Perona-Malik (PM) filtering [37]. It is well-known that the diffusion process governed by the PM equation leads to a decrease in the total variation during its evolution [44]. We, thus, choose the energy functional of the Perona-Malik equation for anisotropic diffusion [37]. One can note that the PM regularization term is differentiable and, thus, easier to handle than the TV regularization term.

Mathematically, one can consider the PM filtering as the gradient flow generated by the non-convex and lower semi-continuous functional given by

JPM(σ)=Ωlog(1+|σ(x,y)|2)dxdy.

We refer to [12, 44] for a general introduction to anisotropic diffusion and a detailed discussion on the PM functional. Further, in [41], the PM model was used in the reconstruction of log-conductivities in AET and it was observed that the reconstructions obtained demonstrated superior contrast and resolution. Thus, for the current setup in CDII, we use a similar PM anisotropic diffusion filter to facilitate high contrast and high resolution images.

3. Theory of the minimization problem

In this section, we discuss the existence of solutions of the minimization problem (P) and its characterization through a first-order optimality system. We refer to this minimization problem as the CDII sparse reconstruction problem (CDII-SR). Our analysis of this problem begins with the discussion concerning the existence and uniqueness of weak solutions of L(u,σ,fD)=0, which can be proved by standard arguments of Riesz representation theorem [18, Chapter 8].

Proposition 1. Let σLad and fDH1/2(Ω). Then the problem (4) has a unique solution in HfD1(Ω).

The solvability of the CDII inversion problem depends on the type of Dirichlet boundary data fDj,j=1,2. In this, context, we have the following lemma from [2].

Lemma 3.1 (Boundary data). Let Ω2 be a bounded simply connected open set, whose boundary Γ is a simple closed curve. Let f = (f1, f2) be a mapping Γ2 which is a homeomorphism of Γ onto a convex closed curve C, and let D denote the bounded convex domain bounded by C. Let σL(Ω), and let U = (u1, u2) be the eσ-harmonic mapping whose components u1 and u2 are solutions to the Dirichlet problem (4) with fD=fD1 and fD=fD2, respectively, and fDJH1(Ω)C(Ω¯),J=1,2. Then U is a homeomorphism of Ω onto D. In particular, for all ω ⊂⊂ Ω we have either det(∇u1, ∇u1) > 0 or det(∇u1, ∇u1) < 0 almost everywhere in ω.

In [23], the authors have shown that in 2D, the boundary condition pair fD1=x and fD2=y satisfies the conditions of Lemma 3.1 and, thus, the corresponding solutions to (4) u1 and u2 have no critical points and ∇u1,∇u2 are not collinear in Ω¯. We will use these boundary conditions for our numerical experiments in Section 5.

Next, we consider the Fréchet differentiability of the mapping u(σ).

Lemma 3.2. The map u(σ) defined by (4) is Fréchet differentiable as a mapping from Lad to HfD1(Ω).

For the proof of this Lemma, we refer to [23]. Using Lemma 3.2, we introduce the reduced cost functional

J^(σ)=J(σ,u1(σ),u2(σ)), (6)

where ui(σ), i = 1,2 denotes the unique solution of (4) given σ and fDi,i=1,2. The constrained optimization problem (P) can be formulated as an unconstrained one as follows

minσLadJ^(σ). (7)

We next investigate the existence of a minimizer to the CDII-SR problem (P). We first consider the case when δ = 0, i.e., the Perona-Malik term in the functional J is absent.

Proposition 2. Let fD1,fD2H1/2(Ω) such that |∇u1| > 0, |∇u2| > 0 and let δ = 0. Then there exists a triplet (σ*,u1*,u2*)Lad×HfD11(Ω)×HfD21(Ω) such that ui*,i=1,2 are solutions to L(σ,ui,fDi)=0,i=1,2 and σ* minimizes J^ in Lad.

Proof. Boundedness from below of J^ guarantees the existence of a minimizing sequence (σm). Since Lad is reflexive and J^ is sequentially weakly lower semi-continuous, this sequence is bounded. Therefore it contains a weakly convergent subsequence (σml) in Lad, σmlσ*. Correspondingly, the sequence (u1ml,u2ml), where uiml=ui(σml), is bounded in HfD11(Ω)×HfD21(Ω). Therefore the sequence converges weakly to (u1*,u2*). Now, using the Rellich Kondrachev compactness theorem in 2, we have that Lad is compactly embedded in L2(Ω). This results in a strong convergence of the subsequence σml in L2(Ω) to σ*. We, now, consider the weak formulation of the solutions of the elliptic problem (4) and, thus, focus on (σmluiml),ψL2(Ω) for any ψH01(Ω). Using integration by parts, we have (σmluiml),ψL2(Ω)=σmluiml,ψL2(Ω). From the above discussion, the sequence of products σmluiml is weakly convergent in L2(Ω), that is, σmluiml,ψL2(Ω)σ*ui*,ψL2(Ω). With this preparation and using the continuity of the maps ui(σ), it follows that (u1*,u2*)=(u1(σ*),u2(σ*)), and the triplet (σ*,u1*,u2*) minimizes the objective J^. □

In the case δ ≠ ≠ 0, we first note that the function log(1 + z2) is not convex. Therefore the PM functional, and, hence, the functional J^ in (5) is not weakly lower semi-continuous on W1,p(Ω) for any 1 < p < ∞. Nevertheless, J^ is a bounded below, lower semi-continuous Lipschitz functional, for which a minimizer exists, provided that Lad is compact.

3.1. Characterization of local minima

To characterize the solution of our optimization problem through first-order optimality conditions, we write the reduced functional J^ as

J^=J^1+J^2,Ji:Lad+,i=1,2,

where

J^1(σ)=α12eσ|u1|g1δL2(Ω)2+α22eσ|u2|g2δL2(Ω)2+β2σL2(Ω)2+δ2Ωln(1+|σ(x,y)|2)dxdy,J^2(σ)=γσL1(Ω). (8)

Remark 3.1. The functional J^1 is smooth and possibly non-convex, while J^2 is non-smooth and convex.

We next state some properties of the reduced functional J^1(σ) which can be proved using the arguments in [23, Lemma 3.1].

Proposition 3. The reduced functional J^1(σ) is weakly lower semi-continuous, bounded below and Fréchet differentiable.

We now define the subdifferential of a non-smooth functional.

Definition 3.1 (Subdifferential). If J^ is finite at a point σ, the Fréchet subdifferential of J^ at σ is defined as follows [15]

J^(σ¯){ϕLad*:lim infσσ¯J^(σ)J^(σ¯)ϕ,σσ¯σ¯σ20}, (9)

Where Lad* is the dual space of Lad. An element ϕJ^(σ) is called a subdifferential of J^ at σ.

In our setting, we have the following

J^(σ)=J^1(σ)+J^2(σ),

since J^1 is Fréchet differentiable by Prop. 3. Moreover, for each α > 0, it holds that

(αJ^)=αJ^.

The following proposition gives a necessary condition for a local minimum of J^ (see [41]).

Proposition 4 (Necessary condition). If J^=J^1+J^2, with J^1, J^2 given by (8), attains a local minimum at σ* ∈ Lad, then

0J^(σ*),

or equivalently

J^1(σ*)J^2(σ*).

The following variational inequality holds for each λJ^2(σ*) (see [50]).

J^1(σ*),+λ,σσ*0,σLad. (10)

Using the definition of J^2 in (8) and the fact that Lad is reflexive, the inclusion λJ^2(σ*) gives the following characterization of space of λ

λΛad{λL2(Ω):0λγ, a.e. in Ω}.

A pointwise analysis of the variational inequality (10) leads to the existence of a non-negative functions λσl*,λσu*L2(Ω) that correspond to Lagrange multipliers for the inequality constraints in Lad. We, thus, have the following first-order optimality system.

Proposition 5 (First-order necessary conditions). The optimal solution of the minimization problem (7) can be characterized by the existence of (λ*,λσl*,λσu*)Λad×L2(Ω)×L2(Ω) such that

σJ^1(σ*)+λ*+λσu*λσl*=0, (11)
λσu*0,σuσ*0,λσu*,σuσ*=0, (12)
λσl*0,σ*σl0,λσl*,σ*σl=0, (13)
λ*=γ a.e. on {xΩ:σ*(x)>0}, (14)
0λ*γ a.e. on {xΩ:σ*(x)=0}. (15)

The conditions (12)–(15) are known as the complementarity conditions for (σ*, λ*).

To determine the gradient σJ^1, we use the adjoint approach (see for e.g., [39, 40]). This gives the following reduced gradient of J^1

σJ^1(σ*)=α1(eσ*|u1|g1δ)|u1|+α2(eσ*|u2|g2δ)|u2|+u1v1+u2v2+βσ*δ(σ*1+|σ*|2) (16)

where u1, u2 satisfy the forward equations L(u1,σ*,fD1)=0,L(u2,σ*,fD2)=0, respectively, and v1, v2 satisfy the adjoint equations

(eσ*v1)=α1[eσ*(eσ*|u1|g1δ)sign(u1)] in Ω,v1|Γ=0, (17)
(eσ*v2)=α2[eσ*(eσ*|u2|g2δ)sign(u2)] in Ω,v2|Γ=0. (18)

The complementarity conditions (12)–(15) can be rewritten in a compact form as follows. Define

μ*=λ*+λσu*λσl*. (19)

Then the triplet (λ*,λσl*,λσu*) is obtained by solving the following equations

λ*=min(γ,max(0,μ*)),λσl*=min(0,μ*+γ),λσu*=max(0,μ*γ), (20)

(see [50]). For each k+, define the following quantity

E(σ*,μ*)=σ*max{0,σ*+k(μ*γ)}+max{0,σ*σu+k(μ*γ)}min{0,σ*+k(μ*+γ)}+min{0,σ*σl+k(μ*+γ)}.

The following lemma determines the complementarity conditions (12)–(15) in terms of E (see [50, Lemma 2.2]).

Lemma 3.3. The complementarity conditions (12)–(15) are equivalent to the following

E(σ*,μ*)=0, (21)

where μ is defined in (19).

Using the gradients in (16) and Lemma 3.3, the optimality conditions (11)–(15) for the CDII-SR problem can be rewritten as follows

Proposition 6. A local minimizer (u1, u2, σ*) of the problem (P) can be characterized by the existence of (v1,v2,μ*)H01(Ω)×H01(Ω)×Lad, such that the following system is satisfied

(eσ*u1)=0 in Ω,u1|Γ=fD1,(eσ*v1)=α1[eσ*(eσ*|u1|g1δ)sign(u1)] in Ω,v1|Γ=0,(eσ*u2)=0 in Ω,u2|Γ=fD2,(eσ*v2)=α2[eσ*(eσ*|u2|g2δ)sign(u2)]  in Ω,v2|Γ=0,α1(eσ*|u1|g1δ)eσ*|u1|+α2(eσ*|u2|g2δ)eσ*|u2|+u1v1+u2v2+βσ*δ(σ*1+|σ*|2)+μ*=0, a.e. in Ω,E(σ*,μ*)=0,  a.e. in Ω, (22)

4. Numerical solution of the CDII-SR problem

In this section, we discuss numerical optimization and approximation schemes to solve the CDII-SR problem. We also describe the Picard-type iterative scheme used by the authors in [20] to compare the results of our scheme with the Picard scheme.

4.1. Variable inertial proximal method for CDII-SR

In this context, we first discuss proximal methods that consists of identifying a smooth and a non-smooth part in the reduced objective J^(σ). Thus, we consider the following optimization problem

minσLadJ^(σ)J^1(σ)+J^2(σ). (23)

We assume that σJ^1(σ), given in (16) is Lipschitz continuous and the upper bound for the Lipschitz constant is obtained using a backtracking search scheme, which will be discussed later. Also, from (8), we have that J^2(σ) is a continuous, convex, and nondifferentiable functional. The formulation of proximal methods depends, essentially, on the following lemma [41]

Lemma 4.1. Let J^1 be differentiable with a Lipschitz continuous gradient with Lipschitz constant L(J^1). Then the following holds

J^1(σ)J^1(σ˜)+J^1(σ˜),σσ˜+L2σσ˜2,σ,σ˜Lad, (24)

for all LL(J^1)>0.

We note that LL(J^1) represents the smallest value of L such that (24) holds true.

In a proximal scheme, one usually minimizes an upper bound of the objective functional at each iteration, instead of minimizing the functional directly. From Lemma 4.1, we obtain the following

minσLad{J^1(σ)+J^2(σ)}minσLad{J^1(σ˜)+J^1(y),σσ˜+L2σσ˜2+J^2(σ)},

where equality holds if σ=σ˜. Furthermore, we have the following equation

arg minσLad{J^1(σ˜)+J^1(σ˜),σσ˜+L2σσ˜2+J^2(σ)}=arg minσLad{L2σ(σ˜1LJ^1(σ˜))2+J^2(σ)}. (25)

Using the definition of J^2(σ)=γσL1(Ω), we have the following lemma from [45] that helps in characterizing the solution of (25).

Lemma 4.2. The following equation holds

arg minσLad{τσL1+12σσ˜2}=SτLad(σ˜) for any  σ˜L2(Ω),

where the left-hand side represents the proximal function and the projected soft thresholding function on the right-hand side is defined as follows

SτLad(σ˜){min{σ˜τ,σu} on {(x,y)Ω: σ˜(x,y)>τ}0 on {(x,y)Ω:  |σ˜(x,y)|τ}max{σ˜+τ,σl} on {(x,y)Ω: σ˜(x,y)<τ}. (26)

Using this lemma, the solution to (25) is given by

arg minσLad{J^2(σ)+L2σ(σ˜1LJ^1(σ˜))2}=SγLLad(σ˜1LJ^1(σ˜)).

This gives rise to the following iterative scheme

σk+1SγLLad(σk1LJ^1(σk)),

starting from a given σ0 and is known as the iterative shrinkage-thresholding algorithm (ISTA) scheme [45]. We note that the argument of SγLLad represents a gradient update in a steepest descent scheme with a fixed step size s = 1/L in conjunction with a regularized PM filter [41]. Further, to accelerate the ISTA scheme described above, one can consider a sequence {tk, vk} [45, 46] such that

t0=1,tk1+1+4tk12/2, (27)

and

v0σ0,vkσk+(tk11)tk(σkσk1). (28)

This gives us the following update for the optimization variable σk

σk+1SγLLad(vk1LJ^1(vk)). (29)

Replacing vk in (29) with (28), and assuming that J^1(σk)J^1(vk), we obtain the following iterative scheme [46]

σk+1SγskLad(σkskσJ^1(σk)+θk(σkσk1)), (30)

Where σ−1 = σ0.

The above discussion is valid for any LL(J^1). However, since the quantity s = 1/L represents the step size in a gradient update, we use a backtracking line search algorithm to determine the optimal step size in each iteration. This leads to the computation of an upper bound Lk that satisfies LkL(J^1) at each iteration step. Thus, we define our variable step size as sk = 1/Lk and substitute τ in (26) with γsk. The variable step size causes the factor (tk11)tk in (28) to be non-optimal and we replace it by the fixed inertial parameter θ. This leads to a variable inertial proximal (VIP) scheme, which is described in Algorithm 4.1.

With our VIP scheme, we aim at determining an optimal σLadH01(Ω). But in the update step of the algorithm, we have the argument of the thresholding function SγLLad as σkskσJ^1(σk). The term σJ^1(σk) is only in L2(Ω) and the resulting update gives us the argument of SγLLad in L2(Ω), which is not desired. We, thus, use the H1 gradient instead of the L2 gradient, which are related by the equation ((σJ^1)H1,v)H1(Ω)=(σJ^1,v)L2(Ω) for all vH1(Ω). But such a H1 gradient results in a highly diffused σ with blurred edges. We, instead, consider a weighted H1 product that represents a suitable denoising of the σJ^1(σ). We apply the denoising operator R(c) = (IcΔ)−1 with a small denoising parameter c and define (σJ^1)H1=R(c)σJ^1. Note that a higher value of c results in a greater blurring of the edges along with noise removal. On the other hand, since the PM term in the functional J promotes better resolution with edgeenhancement, we choose the value of c in proportion to the weight δ of the PM functional term.

We summarize the variable inertial proximal (VIP) scheme for our CDII-SR setup in Algorithm 4.1 below

Algorithm 4.1 (Variable inertial proximal (VIP) method).

  1. Input: β, J^1, σ0 = σ−1, Lad, TOL, n > 1, L0 > 0

    Initialize: E0 = 1, k = 0, choose θ ∈ (0, 1) and c1 < 2 and c2 > 0;

  2. WhileEk−1‖ > TOL do

  3. Compute σJ^1(σk)

  4. Backtracking: Find the smallest nonnegative integer i such that with L˜=niLk1
    J^1(σ˜)J^1(σk)+σJ^1(σk),σ˜σk+L˜2σ˜σk2
    where σ˜=SγsLad(σks(σJ^1)H1(σk)+θ(σkσk1)),s=c1(1θ)/(L˜+2c2),
  5. Set Lk=L˜ and sk=c1(1θ)/(Lk+2c2).

  6. σk+1=SγskLad(σksk(σJ1^)H1(σk)+θ(σkσk1))

  7. μk=ασk(σJ^1)H1(σk)

  8. Ek = E(σk, μk)

  9. k = k + 1

  10. end

In the VIP algorithm, we need to compute the reduced gradient σJ^1. This, in turn, requires an accurate numerical solution of the forward and the corresponding adjoint EIT problems as given in Proposition 6. For the forward EIT equation (4), we use the cell-nodal finite-difference approximation. We consider a sequence of uniform grids {Ωh}h>0 given by

Ωh={(xi,yj)2:(xi,yj)=(a+ih,a+jh),(i,j){0,,N}2}Ω,

where N represents the number of cells in each direction and h=(ba)N is the mesh size. The corresponding cell-nodal scheme for (4), at the grid point (xi, yj), is given as follows

1h2{(eσi+1/2,j+eσi1/2,j+eσi,j+1/2+eσi,j1/2)ui,jeσi+1/2,jui+1,jeσi1/2,jui1,jeσi,j+1/2ui,j+1eσi,j1/2ui,j1}=0,     1i,jN1, (31)

where σi±1,j = σ(xi ± h, yj), σi,j±1 = σ(xi, yj ± h). The required intermediate values of σ are computed as follows

σi±1/2,j=12(σi±1,j+σi,j)  and  σi,j±1/2=12(σi,j±1+σi,j).

The Dirichlet boundary data fD is included in the usual way in the right-hand side of the algebraic equation.

For the adjoint equations (17) and (18), we first note that the cell nodal finite difference scheme is not applicable to the right-hand side term in both the equations as they are of the form G = ∇·F, where F is in L2(Ω). We modify the cell nodal scheme by replacing the nodal value of G at (xi, yj) with a cell average of G given as follows

Ga=1h2CijG(x,y) dxdy,

where the cell Cij is defined by

Cij(xih2,xi+h2)×(yjh2,yj+h2),1i,jN1,

Since G = ∇ · F, using the divergence theorem we have

Ga=1h2CijF(x,y) dxdy=1h2CijF(x,y).n ds

The above integral can be approximated with a midpoint quadrature rule along each edge of Cij. This results in the following approximation

Ga=Fi+1/2,j1Fi1/2,j1h+Fi,j+1/22Fi,j1/22h,

where F = (F1, F2).

4.2. Picard-type algorithm

To solve the CDII inverse problem, the authors in [20] use a Picard-type iterative algorithm. This algorithm was the fundamental approach that was used to solve the CDII inverse problem expressed as the 1-Laplacian equation given in (3) [35, 36] was known as the J-substitution algorithm. To start the Picard algorithm, we use an initial guess for σ as σ0. We then solve (4) for u1 in the first iteration and then obtain σ1 from the interior data (2) corresponding to this value of u1. In the next iteration, we solve for u2 using σ1 and again obtain the next iterate σ2 from the interior data (2) corresponding to this value of u2. We apply this method repeatedly, going back and forth between the two sets of interior data, till the difference in the L2 norm of two successive iterates of σ are small. The algorithm is summarized below.

Algorithm 4.2 (Picard-type algorithm).

  1. Input: Initial guess σ0, maximum iterations = K, TOL

    Initialize: err0 = 1, k = 0.

  2. While ‖errk2 > TOL and k < K do

  3. j = (k MOD 2) + 1 (To go back and forth between the two interior measurements.

  4. Solve for ujk from (4), i.e., solve
    L(ujk,σk,fDj)=0
  5. Update σk+1=Hj|ujk|

  6. errk+1 = ‖σk+1 − σk2

  7. k = k + 1

  8. end

The authors demonstrated in their paper that the Picard algorithm was fast and efficient. With 2 data sets and even in the presence of noisy interior data, the Picard method gave good conductivity reconstructions in CDII with a high resolution and less visible artifacts in comparison to other Newton based schemes. Thus, in order to test the performance of our CDII-SR scheme, we will compare the results of our scheme with the results of the Picard scheme in the next section.

5. Numerical experiments

In this section, we validate our CDII-SR framework using different experiments that support the choice of our proposed mathematical formulation and demonstrate its effectiveness in reconstructing a wide variety of objects. We choose the two boundary conditions as fD1=x,. fD2=y on Γ, which is the boundary of Ω = (−1, 1) × (−1, 1). For the experiments, we test our scheme with different values of the regularization parameters β, γ, δ in the functional (5), relative to the value of the data-fit parameters α1, α2. For this purpose, we fix the values of α1 = α2 = 1.0. Since the primary objective is to fit the measured interior data, the value of the regularization parameters are chosen to be less than 1. The parameters of our VIP scheme are chosen based on the convergence properties of the scheme given in [46]. We specified the tolerance level as TOL = 10−4 and maximum number of iterations as 20. But, due to the high non-linearity of the problem, our VIP scheme terminates because of the criteria of maximum number of iterations. For the Picard algorithm, described in Section 4.2, we choose the same stopping criterion. To solve the equation L(ujk,σk,fDj)=0 in Algorithm 4.2, we use the finite difference scheme described in (31).

In all the experiments, the domain Ω is uniformly discretized into N = 150 subintervals in both the x and y directions with h = 0.013. The generation of the synthetic interior electric field data Hδ is done as follows: we first solve for u in (4) with given value of σ on a finer mesh with N = 400 using the finite difference method outlined in Section 4. Then, we compute ∇u with one-sided finite differences to obtain Hδ on the finer mesh. In the final step, we restrict the obtained Hδ onto the coarser mesh with N = 150 and choose this as our given data to which we also add noise in some of the experiments.

In Test Case 1, we consider a phantom that is represented by a disk centered at (0.25, 0.25) with radius 0.25. The value of σ inside the disk is 1 with the background value chosen as 0. The plots of the actual σ and the reconstructed σ are shown in Figure 1.

Figure 1:

Figure 1:

Test Case 1- The actual and reconstructed disk with different choices of the values of the regularization weights.

Figure 1b shows the reconstruction of σLad without any regularization terms, i.e. β = γ = δ = 0 and no denoising, i.e., c = 0. Presence of strong artifacts can be observed in this case, which is inherent to the inverse problem and not the algorithm. A study of the pattern of such artifacts are very challenging and is out of the scope of the paper.

Figure 1c shows the result using the CDII-SR scheme without the denoising and the Perona-Malik regularization term, c = δ = 0, but with the L2L1 regularization. The L2L1 regularization parameter values were set at β = 0.03, γ = 0.3. We observe that the artifacts are reduced to some extent, but are still present. Figure 1d shows the reconstruction with the same L2L1 regularization parameter values and the H1 denoising parameter value c = 0.001 but without the PM filter. In this case, we observe that the artifacts diminish by a huge amount, but the edges are more blunt and the value of σ is lowered, leading to a loss of resolution and contrast. We correct this loss using the PM regularization term with the parameter value δ = 0.01. This can be observed in Figure 1e, where the edges are fairly well seen and the recovered parameter values are very close to the true ones. We also compare our results with the reconstruction in Figure 1f obtained with the Picard algorithm proposed in [20]. We observe a lot of artifacts and a significant loss of contrast in Figure 1f in comparison to the reconstruction shown in Figure 1e, which suggests that our CDII-SR scheme outperforms the Picard scheme.

We also implemented Test Case 1 with different values of the parameters θ, c1, c2 in the VIP algorithm along with values of β, γ, δ = 0, 0.001, 0.01, 0.1. The convergence results of the VIP algorithm are listed in Table 1. The variable v represents the value of each one of the parameters θ, c1, c2. In the table, each cell in the columns under the individual parameters represent the convergence property of the VIP algorithm for v in the prescribed range, when the other parameters are specified in their range of convergence.

Table 1:

Convergence properties of the VIP algorithm for various values v of θ, c1, c2

Values θ c1 c2
υ ≥ 2 No convergence No convergence Slow Convergence
υ < 2, υ ≈ 2 No convergence Good convergence Slow Convergence
1 ≤ υ < 2 No convergence Convergence Slow Convergence
υ < 1, υ ≈ 1 Convergence Convergence Slow Convergence
υ ≪ 1 Convergence Slow Convergence Good Convergence

We first remark that varying the parameters in the VIP algorithm results in change of convergence properties of the CDII-SR scheme. But choosing different values of θ, c1, c2 in the range of convergence, we did not observe any change in the qualitative or quantitative properties of the reconstructed conductivities. From Table 1, we can see that for values of θ ∈ [0, 1), c1 ≈ 2, c2 ≪ 1 we have good convergence of the VIP scheme. For a more detailed discussion of these range of values for convergence, we refer the reader to [46]. Thus, we choose θ = 0.5, c1 = 1.9, c2 = 0.001 in our subsequent experiments.

In Test Case 2, we consider the heart and lung phantom. It consists of two ellipses representing lungs with the value of σ = 1 and a circular region representing the heart with the value of σ = 0.5. The background value of σ is 0. The plots of the actual and reconstructed σ for various values of β, γ, δ, c are shown in Figure 2.

Figure 2:

Figure 2:

Test Case 2- The actual and reconstructed heart and lung phantom for various values of β, γ, δ

From Figure 2b, we observe that with a small value of the L1 regularization parameter γ = 0.01 in comparison to a higher L2 regularization parameter β = 0.3, there are artifacts present in the image. Increasing the value of γ to 0.1, keeping β fixed results in a decrease of the artifacts as seen in Figure 2c. A further increase in the value of γ to 0.3 results in a significant reduction of artifacts as observed in Figure 2d. In Figure 2f, we obtain a reconstruction with high resolution and contrast with a decrease in β to 0.03. This is expected because the phantom under consideration in this example is sparse. Thus, with a greater weight for L1 regularization term, coupled with the denoising operator controlled by the parameter c, described in Sec 4.1, helps in removing artifacts and, thus, promotes high contrast. We further note from Figures 2e and 2g that changing the value of the Perona-Malik regularization term δ to 0.1 does not visibly change the resolution of the edges. This is because the term δ is linked to the denoising parameter c that is fixed to be 0.001. For this choice of c, choosing δ = 0.01 is sufficient to guarantee sharp edges. We further compare our results to the Picard algorithm proposed in [20]. The corresponding reconstruction is shown in Figure 2h. It can be seen that the CDII-SR scheme provides a better contrast image, yet maintaining the same resolution as that of the Picard scheme. Also, there are far more artifacts through the Picard reconstruction method whereas the sparsity assumption and the H1 denoising in CDII-SR scheme results in an image with very less artifacts.

Further, to test the robustness of our method, we introduce 10% and 25% multiplicative Gaussian noises in the interior data Hδ, which is fed as input to our CDII-SR algorithm. The corresponding reconstructions are shown in Figure 3. We also plot the reconstructions obtained with the Picard algorithm.

Figure 3:

Figure 3:

Test Case 2- The reconstructed heart and lung phantom for various values of noise in the interior data

We note from Figure 3a that even with 10% noise, we still have a high contrast and high resolution reconstruction, which demonstrates the robustness of our method. When the noise percentage was increased to 25%, it led to appearance of some streak artifacts, as seen in Figure 3b. For this purpose, we used a higher value of L1 regularization parameter γ = 0.5 and denoising parameter c = 0.01 to remove artifacts. Due to a higher value of c, we increase the value of δ = 0.1. We also used the Picard algorithm with the noisy data and the reconstructions showed visible presence of streak artifacts in Figure 3c and 3d.

In Test Case 3, we consider a combination of phantoms, where one is supported on a square annulus Sa = {(x,y)2:0.8<x<0.7,0.2<x<0.1,0.8<y<0.7,0.2<y<0.1} with σ = 3.0; the other one consists of 2 disks centered at (0.7, 0.7) with radius 0.2 and σ = 1.0 and at (0.55, 0.55) with radius 0.15 and σ = 2.0. The value of σ inside the square annulus has a value −2.0. The plots of the actual and the reconstructed σ are shown in Figure 4.

Figure 4:

Figure 4:

Test Case 3- The actual and reconstructed heart and lung phantom for various values of β, γ, δ

We again observe a similar behavior as in Test Case 2. For the values of β = 0.3 and γ = 0.01, 0.1, we obtain reconstructions that have visible artifacts, especially at the corners of the square annulus. This can be seen in Figure 4b and 4c. When γ is increased to 0.3 and when β is subsequently lowered to 0.03, we have almost no artifacts except for some in front of the smaller disk as seen in Figure 4f and 4g. We also observe that the values inside the annular hole and the inclusion due to the intersection of the two disks are very well reconstructed with nice contrast and high resolution. With the Picard method, we obtain a large number of artifacts and the value inside the hole is also not reconstructed well, thus, leading to loss of contrast as seen in Figure 4h.

We also test the performance of our method in presence of 10% and 25% Gaussian noise in the interior data. The results are shown in Figure 5.

Figure 5:

Figure 5:

Test Case 3- The reconstructed heart and lung phantom for various values of noise in the interior data

We again observe that our method gives superior quality reconstructions as shown in Figures 5a and 5b. As with Test Case 2, we need greater values of β, γ, δ, c to obtain high contrast and high resolution images. From Figures 5c and 5d, we see that, in comparison to the reconstructions with the CDII-SR scheme, the Picard scheme does not provide good reconstructions of objects with holes and inclusions in presence of noise. In particular, the region inside the hole contains a large number of artifacts. This shows the robustness and efficiency of our method in the reocnstructing objects with holes and inclusions.

In this final Test Case 4, we consider another combination of phantoms that result in a lesser sparsity pattern. This experiment is conducted to demonstrate the effect of the L1 regularization term in improving contrast. The first phantom is supported on a square annulus Sa ={(x,y)2:0.8<x<0.7,0.2<x<0.1,0.8<y<0.7,0.2<y<0.1} with σ = 3.0. The value of σ inside the square annulus has a value −1.5. The second phantom consists of 4 disks: the first centered at (0.7, 0.7) with radius 0.2 and σ = 1.0, the second centered at (0.55, 0.55) with radius 0.15 and σ = 2.0, the third centered at (0, 0) with radius 0.25 and σ = 1.5 and the last centered at (0.05, 0.6) with radius 0.2 and σ = 2.5. The final phantom consists of two heart-shaped objects, represented by cardioids, with the value of σ = 4 in the bigger one and value of σ = 3.0 in the smaller one. The plots of the actual and the reconstructed σ are shown in Figure 6.

Figure 6:

Figure 6:

Test Case 4- The actual and reconstructed heart and lung phantom for various values of β, γ, δ

From Figure 6b, we observe the presence of a large number of artifacts. This is the case when the value of γ = 0.01 is small compared to the value of β. Increasing the value of γ to 0.3 and, subsequently, to 0.5, reduces the number of artifacts as can be seen in Figures 6c and 6d. This shows that the L1 regularization indeed leads to a nice contrast in the images, even when the sparsity level is low. Comparing the reconstruction obtained with the Picard algorithm, we see in Figure 6e the presence of a large number of artifacts and loss of contrast, especially in the hole and the inclusions.

6. Conclusions

In this paper, we propose a new framework to facilitate high contrast and high resolution reconstructions in CDII. Our framework is based on formulating the CDII inverse problem as a PDE-constrained optimization problem. In this setup, we minimize an objective functional comprising of least square interior data fitting terms corresponding to two boundary voltage measurements, a L2L1 penalization terms of the log-conductivity that helps promotes sparsity patterns, thus, improving contrast. Additionally we introduce a PM filtering term to sharpen the edges and obtain high resolution images.

We characterized the solution of the optimization problem through an optimality system that was solved using a proximal scheme, coupled with H1 denoising to remove artifacts in the reconstructions. We then demonstrated the effectiveness of our proposed scheme through several numerical experiments and compared our results with an existing Picard-type scheme. Our scheme facilitated reconstructions of a wide variety of conductivity patterns with superior contrast and resolution, even in the presence of noise in the data.

7. Acknowledgments

S. Roy was partly supported by the National Cancer Institute, National Institutes of Health, grant number: 1R21CA242933-01.

Biographies

1. Madhu Gupta: Madhu Gupta received her Masters degree from Indian Institute of Technology, Bombay, India. Currently, she is pursuing her Ph.D. under the supervision of Dr. Souvik Roy at the Department of Mathematics, University Texas at Arlington. Her current research interests include inverse problems related to medical imaging.

graphic file with name nihms-1544096-b0001.gif

2. Rohit Kumar Mishra: Rohit Kumar Mishra received his Ph.D. in 2017 under the guidance of Dr. Venkateswaran P. Krishnan at TIFR Centre for Applicable Mathematics, Bangalore, India. His primary research interests are in the field of inverse problems related to integral geometry, partial differential equations, microlocal analysis and medical imaging. Currently, he is a postdoctoral scholar at the Department of Mathematics, University of Texas at Arlington. Before joining UTA, he was a postdoctoral scholar at the University of California, Santa Cruz from August 2017 to July 2019.

graphic file with name nihms-1544096-b0002.gif

3. Souvik Roy: Souvik Roy recieved his Ph.D. from Tata Institute of Fundamental Research, Centre for Applicable Mathematics, Bangalore, India in 2015. Since 2018, he is an assistant professor at the Department of Mathematics, University of Texas at Arlington. His research interests include inverse problems in medical imaging and optimal control and mathematical frameworks for cancer treatment.

graphic file with name nihms-1544096-b0003.gif

Footnotes

Publisher's Disclaimer: This Author Accepted Manuscript is a PDF file of an unedited peer-reviewed manuscript that has been accepted for publication but has not been copyedited or corrected. The official version of record that is published in the journal is kept up to date and so may therefore differ from this version.

References

  • [1].Adesokan B, Knudsen K, Krishnan VP and Roy S. A fully non-linear optimization approach to acousto-electric tomography. Inverse Problems, 34:104004, 2018. [Google Scholar]
  • [2].Alessandrini G and Nesi V, Univalent σ-harmonic mappings, Archive for Rational Mechanics and Analysis, 158:155–171, 2001. [Google Scholar]
  • [3].Ambartsoumian G, Gouia-Zarrad R, Krishnan VP and Roy S. Image reconstruction from radially incomplete spherical Radon data. European Journal of Applied Mathematics, 29(3):470–493, 2018. [Google Scholar]
  • [4].Ammari H, Garnier J, Jing W and Nguyen LH. Quantitative thermo-acoustic imaging: An exact reconstruction formula. Journal of Differential Equations, 254(3):1375–1395, 2013. [Google Scholar]
  • [5].Ammari H, Grasland-Mongrain P, Millien P, Seppecher L, and Seo J-K. A mathematical and numerical framework for ultrasonically-induced Lorentz force electrical impedance tomography. Journal de Mathématiques Pures et Appliquées, 103(6):1390–1409, 2015. [Google Scholar]
  • [6].Ammari H, Boulmier S, and Millien P. A mathematical and numerical framework for magnetoacoustic tomography with magnetic induction. Journal of Differential Equations, 259(10):5379–5405, 2015. [Google Scholar]
  • [7].Ammari H, Qiu L, Santosa F, and Zhang W, Determining anisotropic conductivity using diffusion tensor imaging data in magneto-acoustic tomography with magnetic induction. Inverse Problems, 33:125006, 2017. [Google Scholar]
  • [8].Bal G, Guo C and Monard F. Imaging of anisotropic conductivities from current densities in two dimensions, SIAM Journal on Imaging Sciences, 7(4):2538–2557, 2014. [Google Scholar]
  • [9].Bal G, Guo C and Monard F. Inverse anisotropic conductivity from internal current densities, Inverse Problems, 30:025001, 2014. [Google Scholar]
  • [10].Bryan K and Leise T, Making do with Less: an introduction to compressed sensing, SIAM Review, 55:547–566, 2013. [Google Scholar]
  • [11].Candes EJ, Romberg JK, and Tao T, Stable signal recovery from incomplete and inaccurate measurements, Communications on Pure and Applied Mathematics, 59:1207–1223, 2006. [Google Scholar]
  • [12].Catté F, Lions P-L and Morel J-M and Coll T, Image selective smoothing and edge detection by nonlinear diffusion, SIAM Journal on Numerical Analysis, 29:182–193,1992. [Google Scholar]
  • [13].Chan TF, Esedoglu S and Nikolova M, Algorithms for finding global minimizers of image segmentation and denoising models, SIAM Journal on Applied Mathematics, 66(5):1632–1648, 2006. [Google Scholar]
  • [14.Cheney M, Issacson D and Newell JC, Electrical impedance tomography, SIAM Review, 41:85–101, 1999. [Google Scholar]
  • [15].Ekeland I and Témam R, Convex Analysis and Variational Problems, SIAM, Philadelphia, 1999. [Google Scholar]
  • [16].Gao H, Osher S and Zhao H. Quantitative photoacoustic tomography, Mathematical Modeling in Biomedical Imaging II, Lecture Notes in Mathematics 2035, Springer, Heidelberg, 131–158, 2012. [Google Scholar]
  • [17].Garmatter D and Harrach B. Magnetic resonance electrical impedance tomography (MREIT): Convergence and reduced basis approach, SIAM Journal of Imaging Sciences, 11(1):863–887, 2018. [Google Scholar]
  • [18].Gilbarg D and Trudinger NS. Elliptic partial differential equations of second order, Springer·Verlag; Berlin Heidelberg, 2001. [Google Scholar]
  • [19].Hasanov KF, Ma AW, Nachman A and Joy MLG. Current density impedance imaging, IEEE Transactions on Medical Imaging, 27(9):1301–1309, 2008. [DOI] [PubMed] [Google Scholar]
  • [20].Hoffman K and Knudsen K, Iterative reconstruction methods for hybrid inverse problems in impedance tomography, Sensing and Imaging, 15:96, 2014. [Google Scholar]
  • [21].Khang HS, Lee BI, Oh SH, Woo EJ, Lee SY, Cho MH, Kwon O, Yoon JR and Seo JK. J-substitution algorithm in magnetic resonance electrical impedance tomography (MREIT): phaontom experiments for static resistivity images, IEEE Transactions on Medical Imaging, 21:695–702, 2002. [DOI] [PubMed] [Google Scholar]
  • [22].Knudsen K, Lassas M, Mueller JL and Siltanen S. Regularized D-bar method for the inverse conductivity problem, Inverse Problems and Imaging, 3:599–624, 2009. [Google Scholar]
  • [23].Kuchment P and Steinhauer D. Stabilizing inverse problems by internal data, Inverse Problems, 28(8):084007, 2012. [Google Scholar]
  • [24].Kwon O, Woo EJ, Yoon JR and Seo JK. Magnetic resonance electrical impedance tomography (MREIT): simulation study of J-substitution algorithm, IEEE Transactions on Biomedical Engineering, 49:160–167, 2002. [DOI] [PubMed] [Google Scholar]
  • [25].Kim YJ, Kwon O, Seo JK and Woo EJ. Uniqueness and convergence of conductivity image reconstruction in magnetic resonance electrical impedance tomography, Inverse Problems, 19:1213–1225, 2003. [Google Scholar]
  • [26].Kim S, Kwon O, Seo JK and Yoon J-R. On a non-linear partial differential equation arising in magnetic resonance electrical impedance tomography, SIAM Journal Of Mathematical Analysis, 34(3):511–526, 2002. [Google Scholar]
  • [27].Kruger MVP, Tomography as a Metrology Technique for Semiconductor Manufacturing, PhD Thesis, University of California, Berkeley, 2003. [Google Scholar]
  • [28].Lee JY. A reconstruction formula and uniqueness of conductivity in MREIT using two internal current distributions, Inverse Problems, 20:847–858, 2004. [Google Scholar]
  • [29].Li M, Yang H and Kudo H. An accurate iterative reconstruction algorithm for sparse objects: application to 3D blood vessel reconstruction from a limited number of projections, Physics in Medicine and Biology, 47:2599–2609, 2002. [DOI] [PubMed] [Google Scholar]
  • [30].Liu JJ, Seo JK, Sini M and Woo EJ. On the convergence of the harmonic Bz algorithm in magentic resonance electrical impedance tomography, SIAM Journal of Applied Mathematics, 67:1259–1282, 2007. [Google Scholar]
  • [31].Liu JJ, Seo JK and Woo EJ. A posteriori error estimate and convergence analysis for conductivity image reconstruction in MREIT, SIAM Journal of Applied Mathematics, 70:2883–2903, 2010. [Google Scholar]
  • [32].Monard F and Bal G. Inverse diffusion problems with redundant information, Inverse Problems and Imaging, 6(2):289–313, 2012. [Google Scholar]
  • [33].Moradifam A, Nachman A and Timonov A. A convergent algorithm for the hybrid problem of reconstructing conductivity from minimal interior data, Inverse Problems, 28(8), 2012. [Google Scholar]
  • [34].Nachman A, Tamasan A and Timonov A. Conductivity imaging with a single measurement of boundary and interior data, Inverse Problems, 23:2551–2563, 2007. [Google Scholar]
  • [35].Nachman A, Tamasan A and Timonov A. Recovering the conductivity from a single measurement of interior data, Inverse Problems, 25, 2009. [Google Scholar]
  • [36].Nachman A, Tamasan A and Timonov A. Current density impedance imaging, Tomography and Inverse Transport Theory, 559:135–149, 2011. [Google Scholar]
  • [37].Perona P and Malik J, Scale-Space and edge detection using anisotropic diffusion, IEEE Transactions on Pattern Analysis and Machine Intelligence, 12:629–639, 1990. [Google Scholar]
  • [38].Qiu L and Santosa F. Analysis of the magnetoacoustic tomography with magnetic induction. SIAM Journal of Imaging Sciences, 8(3):2070–2086, 2015. [Google Scholar]
  • [39].Roy S, Annunziato M, and Borzì A, A Fokker–Planck feedback control-constrained approach for modelling crowd motion, Journal of Computational and Theoretical Transport, 45:442–458, 2016. [Google Scholar]
  • [40].Roy S, Annunziato M, Borzì A and Klingenberg C. A Fokker-Planck approach to control collective motion. Computational Optimization and Applications, 69(2):423–459, 2018. [Google Scholar]
  • [41].Roy S and Borzì A. A new optimisation approach to sparse reconstruction of log-conductivity in acousto-electric tomography, SIAM Journal of Imaging Sciences, 11(2):1759–1784, 2018. [Google Scholar]
  • [42].Roy S, Krishnan VP, Chandrasekhar P and Vasudeva Murthy AS. An efficient numerical algorithm for Radon transform inversion with applications in ultrasound imaging Journal of Mathematical Imaging and Vision, Springer, 53:78–91, 2015. [Google Scholar]
  • [43].Rudin LI, Osher S and Fatemi E, Nonlinear total variation based noise removal algorithms, Physica D: Nonlinear Phenomena, 60:259–268, 1992. [Google Scholar]
  • [44].Scherzer O and Weickert J, Relations between regularization and diffusion filtering, Journal of Mathematical Imaging and Vision, 12:43–63, 2000. [Google Scholar]
  • [45].Schindele A and Borzí A, Proximal Methods for Elliptic Optimal Control Problems with Sparsity Cost Functional, Applied Mathematics, 7:967–992, 2016. [Google Scholar]
  • [46].Schindele A and Borzí A, Proximal schemes for parabolic optimal control problems with sparsity promoting cost functionals, International Journal of Control, 10.1080/00207179.2016.1245870, 1–19, 2016 [DOI] [Google Scholar]
  • [47].Scott GC, Joy MLG, Armstrong RL and Henkelman RM. Measurement of nonuniform current density by magnetic resonance, IEEE Transactions on Medical Imaging, 10:262–374, 1991. [DOI] [PubMed] [Google Scholar]
  • [48].Seagar AD, Barber DC, and Brown BH, Theoretical limits to sensitivity and resolution in impedance imaging, Clinical Physics and Physiological Measurement 8, Suppl. A, 13–31, 1987. [DOI] [PubMed] [Google Scholar]
  • [49].Seo JK and Woo EJ. Magnetic resonance electrical impedance tomogrpahy, SIAM Review, 53:40–68, 2011. [Google Scholar]
  • [50].Stadler G. Elliptic Optimal Control Problems with L1 -Control Costs and Applications for the Placement of Control Devices. Computational Optimization and Applications, 44:159–181, 2009. [Google Scholar]
  • [51].Tamasan A and Veras J, Conductivity imaging by the method of characteristics in the 1-Laplacian, Inverse Problems 28, 084006 (13pp), 2012 [Google Scholar]
  • [52].Tröltzsch F, Optimal control of partial differential equations: theory, methods and applications, Americal Mathematical Society, 2010. [Google Scholar]
  • [53].Waterfall RC, He R and Beck CM, Visualizing combustion using electrical impedance tomography, Chemical Engineering Science, Vol. 52, No 13, pp. 2129–2138, 1997. [Google Scholar]
  • [54].Zain NM and Chelliah KK. Breast imaging using electrical impedance tomography: correlation of quantitative assessment with visual interpretation. Asian Pacific Journal of Cancer Prevention, 15(3), 1327–1331, 2014, [DOI] [PubMed] [Google Scholar]

RESOURCES