Skip to main content
Journal of Imaging logoLink to Journal of Imaging
. 2021 Jun 16;7(6):99. doi: 10.3390/jimaging7060099

Directional TGV-Based Image Restoration under Poisson Noise

Daniela di Serafino 1, Germana Landi 2,*, Marco Viola 3
Editor: Nicolas Papadakis
PMCID: PMC8321387  PMID: 39080887

Abstract

We are interested in the restoration of noisy and blurry images where the texture mainly follows a single direction (i.e., directional images). Problems of this type arise, for example, in microscopy or computed tomography for carbon or glass fibres. In order to deal with these problems, the Directional Total Generalized Variation (DTGV) was developed by Kongskov et al. in 2017 and 2019, in the case of impulse and Gaussian noise. In this article we focus on images corrupted by Poisson noise, extending the DTGV regularization to image restoration models where the data fitting term is the generalized Kullback–Leibler divergence. We also propose a technique for the identification of the main texture direction, which improves upon the techniques used in the aforementioned work about DTGV. We solve the problem by an ADMM algorithm with proven convergence and subproblems that can be solved exactly at a low computational cost. Numerical results on both phantom and real images demonstrate the effectiveness of our approach.

Keywords: directional image restoration, Poisson noise, DTGV regularization, ADMM method

1. Introduction

Poisson noise appears in processes where digital images are obtained by the count of particles (generally photons). This is the case of X-ray computed tomography, positron emission tomography, confocal and fluorescence microscopy and optical/infrared astronomical imaging, to name just a few applications (see, e.g., [1] and the references therein). In this case, the object to be restored can be represented as a vector uRn and the data can be assumed to be a vector bN0n, whose entries bj are sampled from n independent Poisson random variables Bj with probability

P(Bj=bj)=e(Au+γ)j(Au+γ)jbjbj!.

The matrix A=(aij)Rn×n models the observation mechanism of the imaging system and the following standard assumptions are made:

aij0 for all i,j,i=1naij=1 for all j. (1)

The vector γRn, with γ>0, models the background radiation detected by the sensors.

By applying a maximum-likelihood approach [1,2], we can estimate u by minimizing the Kullback–Leibler (KL) divergence of Au+γ from b:

DKL(Au+γ,b)=i=1nbilnbi[Au+γ]i+[Au+γ]ibi, (2)

where we set

bilnbi[Au+γ]i=0ifbi=0.

Regularization is usually introduced in (2) to deal with the ill-conditioning of this problem. The Total Variation (TV) regularization [3] has been widely used in this context, because it preserves edges and is able to smooth flat areas of the image. However, since it may produce staircase artifacts, other TV-based regularizers have been proposed. For example, the Total Generalized Variation (TGV) has been proposed and applied in [4,5,6,7] to overcome the staircasing effect while keeping the ability of identifying edges. On the other hand, to improve the quality of restoration for directional images, the Directional TV (DTV) regularization has been considered in [8], in the discrete setting. In [9,10], a regularizer combining DTV and TGV, named Directional TGV (DTGV), has been successfully applied to directional images affected by impulse and Gaussian noise.

Given an image uRn, the discrete second-order Directional TGV of u is defined as

DTGV2(u)=minwR2nα0˜uw2,1|R2n+α1E˜w2,1|R4n, (3)

where wR2n, ˜R2n×n and E˜R4n×2n are the discrete directional gradient operator and the directional symmetrized derivative, respectively, and α0,α1(0,+). For any vector vR2n we set

v2,1|R2n=j=1nvj2+vn+j2, (4)

and for any vector yR4n we set

y2,1|R4n=j=1nyj2+yn+j2+y2n+j2+y3n+j2. (5)

Given an angle θ[π,π] and a scaling parameter a>0, we have that the discrete directional gradient operator has the form

˜=DθDθ=cos(θ)DH+sin(θ)DVasin(θ)DH+cos(θ)DV,

where Dθ,DθRn×n represent the forward finite-difference operators along the directions determined by θ and θ=θ+π2, respectively, and DH,DVRn×n represent the forward finite-difference operators along the horizontal and the vertical direction, respectively. Moreover, the directional symmetrized derivative is defined in block-wise form as

E˜=Dθ012Dθ12Dθ12Dθ12Dθ0Dθ.

It is worth noting that, by fixing θ=0 and a=1, we have Dθ=DH and Dθ=DV, and the operators ˜ and E˜ define the TGV2 regularization [4].

We observe that the definition of both the matrix A and the finite difference operators DH and DV depend on the choice of boundary conditions. We make the following assumption.

Assumption 1.

We assume that periodic boundary conditions are considered for A, DH and DV. Therefore, those matrices are Block Circulant with Circulant Blocks (BCCB).

In this work we focus on directional images affected by Poisson noise, with the aim of assessing the behaviour of DTGV in this case. Besides extending the use of DTGV to Poisson noise, we introduce a novel technique for estimating the main direction of the image, which appears to be more efficient than the techniques applied in [9,10]. We solve the resulting optimization problem by using a customized version of the Alternating Direction Method of Multipliers (ADMM). We note that all the ADMM subproblems can be solved exactly at a low cost, thanks also to the use of FFTs, and that the method has proven convergence. Finally, we show the effectiveness of our approach on a set of test images, corrupted by out-of-focus and Gaussian blurs and noise with different signal-to-noise ratios. In particular, the KL-DTGV model of our problem is described in Section 2 and the technique for estimating the main direction is presented in Section 3. A detailed description of the ADMM version used for the minimization is given in Section 4 and the results of the numerical experiments are discussed in Section 5. Conclusions are given in Section 6.

Throughout this work we denote matrices with uppercase lightface letters, vectors with lowercase boldface letters and scalars with lowercase lightface letters. All the vectors are column vectors. Given a vector v, we use vi or (v)i to denote its i-th entry. We use R+ to indicate the set of real nonnegative numbers and · to indicate the two-norm. For brevity, given any vectors v and w we use the notation (v,w) instead of [vw]. Likewise, given any scalars v and w, we use (v,w) to indicate the vector [vw]. We also use the notation ([v]1,[v]2) to highlight the subvectors [v]1 and [v]2 forming the vector v. Finally, by writing v>0 we mean that all the entries of v are nonnegative and at least one of them is positive.

2. The KL-DTGV2 Model

We briefly describe the KL-DTGV2 model for the restoration of directional images corrupted by Poisson noise. Let bRn be the observed image. We want to recover the original image by minimizing a combination of the KL divergence (2) and the DTGV2 regularizer (3), i.e., by solving the optimization problem

minu,wλDKL(Au+γ,b)+α0˜uw2,1|R2n+α1E˜w2,1|R4ns.t.u0, (6)

where uRn, ARn×n, γ,bRn, wR2n, and ˜R2n×n and E˜R4n×2n are the linear operators defining the DTGV2 regularization. The parameters λ(0,+) and α0,α1(0,1) determine the balance between the KL data fidelity term and the two components of the regularization term.

We note that problem (6) is a nonsmooth convex optimization problem because of the properties of the KL divergence (see, e.g., [11]) and the DTGV operator (see, e.g., [10]).

3. Efficient Estimation of the Image Direction

An essential ingredient in the DTGV regularization is the estimation of the angle θ representing the image texture direction. In [10], an estimation algorithm based on the one in [12] is proposed, whose basic idea is to compute a pixelwise direction estimate and then θ as the average of that estimate. In [9], which focuses on impulse noise removal, a more efficient and robust algorithm for estimating the direction is presented, based on the Fourier transform. The main idea behind this algorithm is to exploit the fact that two-dimensional Fourier basis functions can be seen as images with one-directional patterns. However, despite being very efficient from a computational viewpoint, this technique does not appear to be fully reliable in our tests on Poissonian images (see Section 5.1). Therefore, we propose a different approach for estimating the direction, based on classical tools of image processing: the Sobel filter [13] and the Hough transform [14,15].

Our technique is based on the idea that if an image has a one-directional structure, i.e., its main pattern consists of stripes, then the edges of the image mainly consist of lines going in the direction of the stripes. The first stage of the proposed algorithm uses the Sobel filter to determine the edges of the noisy and blurry image. Then, the Hough transform is applied to the edge image in order to detect the lines. The Hough transform is based on the idea that each straight line can be identified by a pair (r,η) where r is the distance of the line from the origin, and η is the angle between the x axis and the segment connecting the origin with its orthogonal projection on the line. The output of the transform is a matrix in which each entry is associated with a pair (r,η), i.e., with a straight line in the image, and its value is the sum of the values in the pixels that are on the line. Hence, the elements with the highest value in the Hough transform indicate the lines that are most likely to be present in the input image. Because of its definition, the Hough transform tends to overestimate diagonal lines in rectangular images (diagonal lines through the central part of the image contain the largest number of pixels); therefore, before computing the transform we apply a mask to the edge image, considering only the pixels inside the largest circle centered in the center of the image. After the Hough transform has been applied, we compute the square of the two-norm of each column of the matrix resulting from the transform, to determine a score for each angle from 90 to 90. Intuitively, the score for each angle is related to the number of lines with that particular inclination which have been detected in the image. Finally, we set the direction estimate θ[π,π] as

θ=90ηmax180π,ηmax0,90ηmax180π,ηmax<0.

where ηmax is the value of η corresponding to the maximum score. A pseudocode for the estimation algorithm is provided in Algorithm 1 and an example of the algorithm workflow is given in Figure 1.

Algorithm 1 Direction estimation.
  • 1:

    Use the Sobel operator to obtain the image e of the edges of the noisy and blurry image b.

  • 2:

    Apply a disk mask to cut out some diagonal edges in e, obtaining a new edge image e˜ (Figure 1b).

  • 3:

    Compute the Hough transform h(e˜) (Figure 1c).

  • 4:

    Set ηmax as the value of η corresponding to the column of h(e˜) with maximum 2-norm. (Figure 1d)

  • 5:

    Set θ={90ηmax180π,ηmax0,90ηmax180π,ηmax<0. (yellow line in Figure 1a)

Figure 1.

Figure 1

Workflow of Algorithm 1 on a random directional image.

4. ADMM for Minimizing the KL-DTGV2 Model

Although problem (6) is a bound-constrained convex optimization problem, the nondifferentiability of the DTGV2 regularizer does not allow its solution by classical optimization methods for smooth problems, such as gradient methods (see [16,17,18] and the references therein). However, the problem can be solved by methods based on splitting techniques, such as [19,20,21,22,23]. Here we solve (6) by the Alternating Direction Method of Multipliers (ADMM) [20]. To this end, we first reformulate the problem as follows:

minu,w,z1,z2,z3,z4λDKL(z1+γ,b)+α0z22,1|R2n+α1z32,1|R4n+χR+n(z4)s.t.z1=Au,z2=˜uw,z3=E˜w,z4=u, (7)

where z1Rn, z2R2n, z3R4n, z4Rn, and χR+n(z4) is the characteristic function of the nonnegative orthant in Rn. A similar splitting has been used in [24] for TV-based deblurring of Poissonian images. By introducing the auxiliary variables x=(u,w) and z=(z1,z2,z3,z4) we can further reformulate the KL-DTGV2 problem as

minx,zF1(x)+F2(z)s.t.Hx+Gz=0, (8)

where we set

F1(x)=0,F2(z)=λDKL(z1+γ,b)+α0z22,1|R2n+α1z32,1|R4n+χR+n(z4), (9)

and we define the matrices HR8n×3n and GR8n×8n as

H=A0˜I2n0E˜In0,G=In0000I2n0000I4n0000In. (10)

We consider the Lagrangian function associated with problem (8),

L(x,z,ξ)=F1(x)+F2(z)+ξHx+Gz, (11)

where ξR8n is a vector of Lagrange multipliers, and then the augmented Lagrangian function

LA(x,z,ξ;ρ)=F1(x)+F2(z)+ξHx+Gz+ρ2Hx+Gz22, (12)

where ρ>0.

Now we are ready to introduce the ADMM method for the solution of problem (8). Let x0R3n, z0R8n, ξ0R8n. At each step k>0 the ADMM method computes the new iterate xk+1,zk+1,ξk+1 as follows:

xk+1=argminxR3nLA(x,zk,ξk;ρ),zk+1=argminzR8nLA(xk+1,z,ξk;ρ),ξk+1=ξk+ρHxk+1+Gzk+1. (13)

Note that the functions F1(x) and F2(z) in (8) are closed, proper and convex. Moreover, the matrices H and G defined in (10) are such that G=I8n and H has full rank. Hence, the convergence of the method defined by (13) can be proved by applying a classical convergence result from the seminal paper by Eckstein and Bertsekas [25] (Theorem 8), which we report in a form that can be immediately applied to our reformulation of the problem.

Theorem 1.

Let us consider a problem of the form (8) where F1(x) and F2(z) are closed, proper and convex functions and H has full rank. Let x0R3n, z0R8n, ξ0R8n, and ρ>0. Suppose {εk},{νk}R+ are summable sequences such that for all k

xk+1argminxR3nLA(x,zk,ξk;ρ)εk,zk+1argminzR8nLA(xk+1,z,ξk;ρ)νk,ξk+1=ξk+ρHxk+1+Gzk+1.

If there exists a saddle point (x*,z*,ξ*) of L(x,z,ξ), then xkx*, zkz* and ξkξ*. If such saddle point does not exist, then at least one of the sequences {zk} or {ξk} is unbounded.

Since we are dealing with linear constraints, we can recast (13) in a more convenient form, by observing that the linear term in (12) can be included in the quadratic one. By introducing the vector of scaled Lagrange multipliers μk=1ρξk, the ADMM method becomes

xk+1=argminxR3nρ2Hxzk+μk22, (14)
zk+1=argminzR8nF2(z)+ρ2Hxk+1z+μk22, (15)
μk+1=μk+Hxk+1+Gzk+1. (16)

In the next sections we show how the solutions to subproblems (14) and (15) can be computed exactly with a small computational effort.

4.1. Solving the Subproblem in x

Problem (14) is an overdetermined least squares problem, since H is a tall-and-skinny matrix with full rank. Hence, its solution can be computed by solving the normal equations system

HHx=Hvxk, (17)

where we set vxk=zkμk. Starting from the definition of H given in (10), we have

HH=In+AA+˜˜˜˜I2n+E˜E˜==In+AA+˜˜DθDθDθIn+DθDθ+12DθDθ12DθDθDθ12DθDθIn+12DθDθ+DθDθ.

System (17) may be quite large and expensive, also for relatively small images. However, as pointed out in Assumption 1, A, Dθ and Dθ have a BCCB structure, hence all the blocks of HH maintain that structure. By recalling that BCCB matrices can be diagonalized by means of two-dimensional Discrete Fourier Transforms (DFTs), we show how the solution to (17) can be computed expeditiously.

Let FCn×n be the matrix representing the two-dimensional DFT operator, and let F* denote its inverse, i.e., its adjoint. We can write HH as

HH=F*000F*000F*ΓΔθ*Δθ*ΔθΦ11Φ12ΔθΦ21Φ22F000F000F, (18)

where each block of the central matrix is the diagonal complex matrix associated with the corresponding block in HH, and Δθ*,Δθ* denote the (diagonal) adjoint matrices of Δθ,Δθ. By (18) and the definition of x, we can reformulate (17) as

ΓΔθ*Δθ*ΔθΦ11Φ12ΔθΦ21Φ22FuFw1Fw2=F[Hvxk]1F[Hvxk]2F[Hvxk]3, (19)

where we split w and vxk in two and three blocks of size n, respectively.

Now we recall a result about the inversion of block matrices. Suppose that a square matrix M is partitioned into four blocks, i.e.,

M=M11M12M21M22;

then, if M11 and M22 are invertible, we have

M1=M11M12M21M221=M11M12M221M21100M22M21M111M121IM12M221M21M111I. (20)

By applying (20) to the matrix consisting of the second and third block rows and columns of the matrix in (19), which we denote Φ, we get

Φ1=Φ11Φ12Φ21Φ221=Φ11Φ12Φ221Φ211Φ11Φ12Φ221Φ211Φ12Φ221Φ22Φ21Φ111Φ121Φ21Φ111Φ22Φ21Φ111Φ121. (21)

To simplify the notation we set

Ψ=Ψ11Ψ12Ψ21Ψ22=Φ1, (22)

and observe that the matrices ΨijCn×n are diagonal. Letting Δ*=Δθ*Δθ*, applying the inversion formula (20) to the whole matrix in (19), and using (21) and (22), we get

ΓΔ*ΔΦ1=Ξ100Ω1InΔ*ΨΔΓ1I2n, (23)

where

Ξ=ΓΔ*ΨΔ=ΓΔθ*Δθ*Ψ11Ψ12Ψ21Ψ22ΔθΔθ*,Ω=ΦΔΓ1Δ*=Φ11Φ12Φ21Φ22ΔθΔθΓ1Δθ*Δθ*.

We note that ΞCn×n is diagonal (and its inversion is straightforward), while ΩC2n×2n has a 2×2 block structure with blocks that are diagonal matrices belonging to Cn×n. Thus, we can compute Υ=Ω1 by applying (20):

Υ=Υ11Υ12Υ21Υ22=Ω11Ω12Ω21Ω221=Ω11Ω12Ω221Ω211Ω11Ω12Ω221Ω211Ω12Ω221Ω22Ω21Ω111Ω121Ω21Ω111Ω22Ω21Ω111Ω121 (24)

Summing up, by (19), (23) and (24), the solution to (17) can be obtained by computing

y1y2y3=Ξ100ΥInΔΨΔΓ1I2nF[Hvxk]1F[Hvxk]2F[Hvxk]3, (25)

and setting

uk+1=F*y1,w1k+1=F*y2,w2k+1=F*y3. (26)

Remark 1.

The only quantity in (25) that varies at each iteration is vxk. Hence, the matrices Δ, Γ, Ψ, Ξ1, and Υ can be computed only once before the ADMM method starts. This means that the overall cost of the exact solution of (14) at each iteration reduces to six two-dimensional DFTs and two matrix–vector products involving two 3×3 block matrices with diagonal blocks of dimension n.

4.2. Solving the Subproblem in z

By looking at the form of F2(z)–see (9)–and by defining the vector vzk=Hxk+1+μk, we see that problem (15) can be split into the four problems

z1k+1=argminz1RnλDKL(z1+γ,b)+ρ2z1[vzk]122, (27)
z2k+1=argminz2R2nα0z22,1|R2n+ρ2z2[vzk]222, (28)
z3k+1=argminz3R4nα1z32,1|R4n+ρ2z3[vzk]322, (29)
z4k+1=argminz4RnχR+n(z4)+ρ2z4[vzk]422, (30)

where vzk=([vzk]1,[vzk]2,[vzk]3,[vzk]4), with [vzk]1Rn, [vzk]2R2n,[vzk]3R4n, and [vzk]4Rn. Now we focus on the solution of the four subproblems.

4.2.1. Update of z1

By the form of the Kullback–Leibler divergence in (2), the minimization problem (27) is equivalent to

minz1Rnλi=1nbilnbi(z1)i+γi+(z1)i+γibi+ρ2i=1n(z1)idi2, (31)

where we set d=[vzk]1 to ease the notation. From (31) it is clear that the problem in z1 can be split into n problems of the form

minzRλbln(z+γ)+z+ρ2zd2. (32)

Since the objective function of this problem is strictly convex, its solution can be determined by setting the gradient equal to zero, i.e., by solving

λbz+γ+1+ρzd=0,

which leads to the quadratic equation

z2+λρ+γdzλρρλγdγ+b=0. (33)

Since, by looking at the domain of the objective function in (32), z+γ has to be strictly positive, we set each entry of z1k+1 as the largest solution of the corresponding quadratic Equation (33).

4.2.2. Update of z2 and z3

The minimization problems (28) and (29) correspond to the computation of the proximal operators of the functions f(z2)=α0ρz22,1|R2n and g(z3)=α1ρz32,1|R4n, respectively.

By the definitions given in (4) and (5), we see that the two (2,1)-norms correspond to the sum of two-norms of vectors in R2 and R4, respectively. This means that the computation of both the proximal operators can be split into the computation of n proximal operators of functions that are scaled two-norms in either R2 or R4.

The proximal operator of the function f(y)=cy, c>0, at a vector d is

proxc·(d)=argminycy+12yd2.

It can be shown (see, e.g., [26] [Chapter 6]) that

proxc·(d)=1cmaxd,cd=maxdcd,0d. (34)

Hence, for the update of z2 we proceed as follows. By setting d=[vzk]2 and c=α0ρ, for each i=1,,n we have

(z2k+1)i,(z2k+1)n+i=proxc·(di,dn+i).

To update z3, we set d=[vzk]3 and c=α1ρ and compute

(z3k+1)i,(z3k+1)n+i,(z3k+1)2n+i,(z3k+1)3n+i=proxc·(di,dn+i,d2n+i,d3n+i).

4.2.3. Update of z4

It is straightforward to verify that the update of z4 in (30) can be obtained as

z4k+1=ΠR+n[vzk]4,

where ΠR+n is the Euclidean projection onto the nonnegative orthant in Rn.

4.3. Summary of the ADMM Method

For the sake of clarity, in Algorithm 2 we sketch the ADMM version for solving problem (7).

In many image restoration applications, a reasonably good starting guess for u is often available. For example, if A represents a blur operator, a common choice is to set u0 equal to the the noisy and blurry image. We make this choice for u0. By numerical experiments we also verified that once x=(u,w) has been initialized, it is convenient to set u1=u0, w11=w10 and w21=w20 and to shift the order of the updates in the ADMM scheme (14)–(16), so that a “more effective” initialization of z and μ is performed. We see from line 9 of Algorithm 2 that the algorithm stops when the relative change in the restored image u goes below a certain threshold tol(0,1) or a maximum number of iterations kmax is reached. Finally, we note that for the case of the KL-TGV2 model, corresponding to θ=0 and a=1, we have that Dθ=DH and Dθ=DV; hence, we use the initialization w10=DHu0 and w20=DVu0.

Algorithm 2 ADMM for problem (7).
  •  1:

    Let u0Rn, w10=Dθu0, w20=Dθu0, μ0=0, z0=0, λ,ρ(0,+), α0,α1(0,1)

  •  2:

    Compute matrices Δ, Γ, Ψ, Ξ1, and Υ as specified in Section 4.1

  •  3:

    Let k=0, u1=u0, w1=w0, stop=false, tol(0,1), kmaxN

  •  4:

    while notstopandkkmax do

  •  5:

     Compute zk+1 by solving the four subproblems (27)–(30)

  •  6:

     Compute μk+1 as in (16)

  •  7: 

    k=k+1

  •  8:

     Compute uk+1, w1k+1 and w2k+1 by (25) and (26)

  •  9:

     Set stop=uk+1uk<toluk

  • 10:

    end while

5. Numerical Results

All the experiments were carried out using MATLAB R2018a on a 3.50 GHz Intel Xeon E3 with 16 GB of RAM and Windows operating system. In this section, we first illustrate the effectiveness of Algorithm 1 for the estimation of the image direction by comparing it with the one given in [9] and by analysing its sensitivity to the degradation in the image to be restored. Then, we present numerical experiments that demonstrate the improvement of the KL-DTGV2 model upon the KL-TGV2 model for the restoration of directional images corrupted by Poisson noise.

Four directional images named phantom (512×512), grass (375×600), leaves (203×300) and carbon (247×300) were used in the experiments. The first image is a piecewise affine fibre phantom image obtained with the fibre_phantom_pa MATLAB function available from http://www2.compute.dtu.dk/~pcha/HDtomo/ (accessed on 20 September 2020). The second and third images represent grass and veins of leaves, respectively, which naturally exhibit a directional structure. The last image is a Scanning Electron Microscope (SEM) image of carbon fibres. The images are shown in Figure 2, Figure 3, Figure 4 and Figure 5.

Figure 2.

Figure 2

Test problem phantom: original and corrupted images. The yellow dash-dotted line indicates the direction estimated by Algorithm 1 and the red dashed line the direction estimated by the method in [9].

Figure 3.

Figure 3

Test problem grass: original and corrupted images. The yellow dash-dotted line indicates the direction estimated by Algorithm 1 and the red dashed line the direction estimated by the method in [9].

Figure 4.

Figure 4

Test problem leaves: original and corrupted images. The yellow dash-dotted line indicates the direction estimated by Algorithm 1 and the red dashed line the direction estimated by the method in [9].

Figure 5.

Figure 5

Test problem carbon: original and corrupted images. The yellow dash-dotted line indicates the direction estimated by Algorithm 1 and the red dashed line the direction estimated by the method in [9].

To simulate experimental data, each reference image was convolved with two PSFs, one corresponding to a Gaussian blur with variance 2, generated by the psfGauss function from [27], and the other corresponding to an out-of-focus blur with radius 5, obtained with the function fspecial from the MATLAB Image Processing Toolbox. To take into account the existence of some background emission, a constant term γ equal to 1010 was added to all pixels of the blurry image. The resulting image was corrupted by Poisson noise, using the MATLAB function imnoise. The intensities of the original images were pre-scaled to get noisy and blurry images with Signal to Noise Ratio (SNR) equal to 43 and 37 dB. We recall that in the case of Poisson noise, which affects the photon counting process, the SNR is estimated as [28]

SNR=10log10NexactNexact+Nbackground,

where Nexact and Nbackground are the total number of photons in the image to be recovered and in the background term, respectively. Finally, the corrupted images were scaled to have their maximum intensity values equal to 1. For each test problem, the noisy and blurry images are shown in Figure 2, Figure 3, Figure 4 and Figure 5.

5.1. Direction Estimation

In Figure 2, Figure 3, Figure 4 and Figure 5 we compare Algorithm 1 with the algorithm proposed in [9], showing that Algorithm 1 always correctly estimates the main direction of the four test images. We also test the robustness of our algorithm with respect to noise and blur. In Figure 6 we show the estimated main direction of the phantom image corrupted by Poisson noise with SNR =35,37,39,41,43 dB and out-of-focus blurs with radius R =5,7,9. In only one case (SNR =35, R =7) Algorithm 1 fails, returning as estimate the orthogonal direction, i.e., the direction corresponding to the large black line and the background color gradient. Finally, we test Algorithm 1 on a phantom image with vertical, horizontal and diagonal main directions corresponding to θ=0,90,45. The results, in Figure 7, show that our algorithm is not sensitive to the specific directional structure of the image.

Figure 6.

Figure 6

Direction estimation for phantom with SNR =35,37,39,41,43 dB (from left to right) and out-of-focus blur with radius R =5,7,9 (from top to bottom).

Figure 7.

Figure 7

Direction estimation for phantom with SNR =37,43 (top, bottom) and out-of-focus blur with radius R =5.

5.2. Image Deblurring

We compare the quality of the restorations obtained by using the DTGV2 and TGV2 regularizers and ADMM for the solution of both models. In all the tests, the value of the penalty parameter was set as ρ=10 and the value of the stopping threshold as tol=104. A maximum number of kmax=500 iterations was allowed. By following [9,10], the weight parameters of DTGV were chosen as α0=β and α1=(1β) with β=2/3. For each test problem, the value of the regularization parameter λ was tuned by a trial-and-error strategy. This strategy consisted in running ADMM with initial guess u0=b several times on each test image, varying the value of λ at each execution. For all the runs the stopping criterion for ADMM and the values of α0, α1 and ρ were the same as described above. The value of λ yielding the smallest Root Mean Square Error (RMSE) at the last iteration was chosen as the “optimal” value.

The numerical results are summarized in Table 1, where the RMSE, the Improved Signal to Noise Ratio (ISNR) [29], and the structural similarity (SSIM) index [30] are used to give a quantitative evaluation of the quality of the restorations. As a measure of the computational cost, the number of iterations and the time in seconds are reported. Table 1 also shows, for each test problem, the values of the regularization parameter λ. The restored images are shown in Figure 8, Figure 9, Figure 10 and Figure 11. For the carbon test problem, Figure 12 shows the error images, i.e., the images obtained as the absolute difference between the original image and the restored one. The values of the pixels of the error images have been scaled in the range [m,M] where m and M are the minimum and maximum pixel value of the DTGV2 and TGV2 error images.

Table 1.

Numerical results for the test problems.

Blur SNR Model λ RMSE ISNR MSSIM Iters Time
phantom
Out-of-focus 43 DTGV 57.5 2.2558 ×102 9.5472 9.3007 ×101 86 10.95
TGV 275 2.8043 ×102 7.6568 8.9887 ×101 89 11.33
37 DTGV 3.25 3.7573 ×102 7.4431 8.5823 ×101 122 15.45
TGV 22.5 4.1719 ×102 6.5339 8.4061 ×101 52 6.64
Gaussian 43 DTGV 25 1.5530 ×102 9.1966 9.7829 ×101 56 7.17
TGV 100 1.8100 ×102 7.8667 9.7200 ×101 45 5.76
37 DTGV 3 2.5498 ×102 9.0841 9.2994 ×101 90 11.41
TGV 17.5 3.0674 ×102 7.4788 9.0199 ×101 53 6.76
grass
Out-of-focus 43 DTGV 60 3.6313 ×102 7.7364 8.7262 ×101 136 15.55
TGV 550 3.6575 ×102 7.6738 8.7188 ×101 179 20.39
37 DTGV 50 5.6164 ×102 4.7390 7.6165 ×101 160 18.56
TGV 55 5.7604 ×102 4.5191 7.4566 ×101 72 8.31
Gaussian 43 DTGV 65 2.9883 ×102 6.3343 9.2764 ×101 106 12.08
TGV 650 3.0814 ×102 6.0676 9.2523 ×101 136 15.48
37 DTGV 5.5 4.2274 ×102 4.7973 8.5615 ×101 98 11.13
TGV 35 4.3936 ×102 4.4624 8.4795 ×101 54 6.18
leaves
Out-of-focus 43 DTGV 125 6.2767 ×102 7.4978 8.2099 ×101 251 31.18
TGV 1100 8.2397 ×102 5.1342 7.1557 ×101 435 53.74
37 DTGV 12.5 9.5597 ×102 4.1497 6.3065 ×101 257 31.87
TGV 90 1.1874 ×101 2.2665 4.3294 ×101 113 14.03
Gaussian 43 DTGV 150 7.3332 ×102 4.8675 7.7456 ×101 236 29.13
TGV 1750 8.0857 ×102 4.0190 7.3001 ×101 380 46.77
37 DTGV 12.5 9.0999 ×102 3.3907 6.6469 ×101 148 18.36
TGV 100 1.0308 ×101 2.3081 5.6534 ×101 103 12.85
carbon
Out-of-focus 43 DTGV 150 1.8360 ×102 1.2830 ×101 9.4734 ×101 331 13.78
TGV 850 2.3825 ×102 1.0567 ×101 9.3671 ×101 233 9.73
37 DTGV 20 3.1682 ×102 8.2416 8.6294 ×101 171 7.07
TGV 150 3.8840 ×102 6.4723 8.2237 ×101 155 6.55
Gaussian 43 DTGV 250 2.0453 ×102 8.6178 9.5974 ×101 305 12.53
TGV 950 2.4839 ×102 6.9302 9.5698 ×101 171 7.12
37 DTGV 15 2.7995 ×102 6.2017 9.3007 ×101 128 5.36
TGV 150 3.3061 ×102 4.7572 8.9690 ×101 118 4.73

Figure 8.

Figure 8

Test problem phantom: images restored with DTGV2 (top) and TGV2 (bottom).

Figure 9.

Figure 9

Test problem grass: images restored with DTGV2 (top) and TGV2 (bottom).

Figure 10.

Figure 10

Test problem leaves: images restored with DTGV2 (top) and TGV2 (bottom).

Figure 11.

Figure 11

Test problem carbon: images restored with DTGV2 (top) and TGV2 (bottom).

Figure 12.

Figure 12

Test problem carbon: difference images with DTGV2 (top) and TGV2 (bottom).

From the results, it is evident that the DTGV2 model outperforms the TGV2 one in terms of quality of the restoration. A visual inspection of the figures shows that the DTGV2 regularization is very effective in removing the noise, while for high noise levels the TGV2 reconstructions still exhibit noise artifacts. Finally, by observing the “Iters” column of the table, we can conclude that, on average, the TGV2 regularization requires less ADMM iterations to achieve a relative change in the restoration that is below the fixed threshold. However, the computational time per iteration is very small and also ADMM for the KL-DGTV2 regularization is efficient.

Finally, to illustrate the behaviour of ADMM, in Figure 13 we plot the RMSE history for the carbon test problem. A similar RMSE behaviour has been observed in all the numerical experiments.

Figure 13.

Figure 13

Test problem carbon: RMSE history for the KL-DGTV2 (continuous line) and KL-TGV2 (dashed line) models.

6. Conclusions

We dealt with the use of the Directional TGV regularization in the case of directional images corrupted by Poisson noise. We presented the KL-DTGV2 model and introduced a two-block ADMM version for its minimization. Finally, we proposed an effective strategy for the estimation of the main direction of the image. Our numerical experiments show that for Poisson noise the DTGV2 regularization provides superior restoration performance compared with the standard TGV2 regularization, thus remarking the importance of taking into account the texture structure of the image. A crucial ingredient for the success of the model was the proposed direction estimation strategy, which proved to be more reliable than those proposed in the literature.

Possible future work includes the use of space-variant regularization terms and the analysis of automatic strategies for the selection of the regularization parameters.

Author Contributions

All authors have contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the Istituto Nazionale di Alta Matematica, Gruppo Nazionale per il Calcolo Scientifico (INdAM-GNCS). D. di Serafino and M. Viola were also supported by the V:ALERE Program of the University of Campania “L. Vanvitelli”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Bertero M., Boccacci P., Desiderà G., Vicidomini G. Image deblurring with Poisson data: From cells to galaxies. Inverse Probl. 2009;25:123006. doi: 10.1088/0266-5611/25/12/123006. [DOI] [Google Scholar]
  • 2.Shepp L.A., Vardi Y. Maximum Likelihood Reconstruction for Emission Tomography. IEEE Trans. Med. Imaging. 1982;1:113–122. doi: 10.1109/TMI.1982.4307558. [DOI] [PubMed] [Google Scholar]
  • 3.Rudin L.I., Osher S., Fatemi E. Nonlinear total variation based noise removal algorithms. Physica D. 1992;60:259–268. doi: 10.1016/0167-2789(92)90242-F. [DOI] [Google Scholar]
  • 4.Bredies K., Kunisch K., Pock T. Total generalized variation. SIAM J. Imaging Sci. 2010;3:492–526. doi: 10.1137/090769521. [DOI] [Google Scholar]
  • 5.Bredies K., Holler M. Regularization of linear inverse problems with total generalized variation. J. Inverse Ill-Posed Probl. 2014;22:871–913. doi: 10.1515/jip-2013-0068. [DOI] [Google Scholar]
  • 6.Gao Y., Liu F., Yang X. Total generalized variation restoration with non-quadratic fidelity. Multidimens. Syst. Signal Process. 2018;29:1459–1484. doi: 10.1007/s11045-017-0512-x. [DOI] [Google Scholar]
  • 7.Di Serafino D., Landi G., Viola M. TGV-based restoration of Poissonian images with automatic estimation of the regularization parameter; Proceedings of the 21th International Conference on Computational Science and Its Applications (ICCSA); Cagliari, Italy. 5–8 July 2021. [Google Scholar]
  • 8.Bayram I., Kamasak M.E. Directional Total Variation. IEEE Signal Process. Lett. 2012;19:781–784. doi: 10.1109/LSP.2012.2220349. [DOI] [Google Scholar]
  • 9.Kongskov R.D., Dong Y. Directional Total Generalized Variation Regularization for Impulse Noise Removal. In: Lauze F., Dong Y., Dahl A., editors. Scale Space and Variational Methods in Computer Vision. Volume 10302. Springer; Cham, Switzerland: 2017. pp. 221–231. SSVM2017; Lecture Notes in Computer Science. [DOI] [Google Scholar]
  • 10.Kongskov R.D., Dong Y., Knudsen K. Directional total generalized variation regularization. BIT. 2019;59:903–928. doi: 10.1007/s10543-019-00755-6. [DOI] [Google Scholar]
  • 11.Di Serafino D., Landi G., Viola M. ACQUIRE: An inexact iteratively reweighted norm approach for TV-based Poisson image restoration. Appl. Math. Comput. 2020;364:124678. doi: 10.1016/j.amc.2019.124678. [DOI] [Google Scholar]
  • 12.Setzer S., Steidl G., Teuber T. Restoration of images with rotated shapes. Numer. Algorithms. 2008;48:49–66. doi: 10.1007/s11075-008-9182-y. [DOI] [Google Scholar]
  • 13.Kanopoulos N., Vasanthavada N., Baker R. Design of an image edge detection filter using the Sobel operator. IEEE J. Solid-State Circuits. 1988;23:358–367. doi: 10.1109/4.996. [DOI] [Google Scholar]
  • 14.Hough P. Method and Means for Recognizing Complex Patterns. 3069654. U.S. Patent. 1962 Dec 18;
  • 15.Duda R.O., Hart P.E. Use of the Hough transform to detect lines and curves in pictures. Commun. ACM. 1972;15:11–15. doi: 10.1145/361237.361242. [DOI] [Google Scholar]
  • 16.Bonettini S., Zanella R., Zanni L. A scaled gradient projection method for constrained image deblurring. Inverse Probl. 2009;25:015002. doi: 10.1088/0266-5611/25/1/015002. [DOI] [Google Scholar]
  • 17.Birgin E.G., Martínez J.M., Raydan M. Spectral Projected Gradient Methods: Review and Perspectives. J. Stat. Softw. 2014;60:1–21. doi: 10.18637/jss.v060.i03. [DOI] [Google Scholar]
  • 18.Di Serafino D., Ruggiero V., Toraldo G., Zanni L. On the steplength selection in gradient methods for unconstrained optimization. Appl. Math. Comput. 2018;318:176–195. doi: 10.1016/j.amc.2017.07.037. [DOI] [Google Scholar]
  • 19.Goldstein T., Osher S. The Split Bregman Method for L1-Regularized Problems. SIAM J. Imaging Sci. 2009;2:323–343. doi: 10.1137/080725891. [DOI] [Google Scholar]
  • 20.Boyd S., Parikh N., Chu E., Peleato B., Eckstein J. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found. Trends Mach. Learn. 2011;3:1–122. doi: 10.1561/2200000016. [DOI] [Google Scholar]
  • 21.Parikh N., Boyd S. Proximal Algorithms. Found. Trends Optim. 2014;1:123–231. doi: 10.1561/2400000003. [DOI] [Google Scholar]
  • 22.Bonettini S., Loris I., Porta F., Prato M. Variable Metric Inexact Line-Search-Based Methods for Nonsmooth Optimization. SIAM J. Optim. 2016;26:891–921. doi: 10.1137/15M1019325. [DOI] [Google Scholar]
  • 23.De Simone V., di Serafino D., Viola M. A subspace-accelerated split Bregman method for sparse data recovery with joint ℓ1-type regularizers. Electron. Trans. Numer. Anal. 2020;53:406–425. doi: 10.1553/etna_vol53s406. [DOI] [Google Scholar]
  • 24.Setzer S., Steidl G., Teuber T. Deblurring Poissonian images by split Bregman techniques. J. Vis. Commun. Image Represent. 2010;21:193–199. doi: 10.1016/j.jvcir.2009.10.006. [DOI] [Google Scholar]
  • 25.Eckstein J., Bertsekas D.P. On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators. Math. Program. 1992;55:293–318. doi: 10.1007/BF01581204. [DOI] [Google Scholar]
  • 26.Beck A. First-Order Methods in Optimization. Society for Industrial and Applied Mathematics; Philadelphia, PA, USA: 2017. [DOI] [Google Scholar]
  • 27.Nagy J.G., Palmer K., Perrone L. Iterative methods for image deblurring: A Matlab object-oriented approach. Numer. Algorithms. 2004;36:73–93. doi: 10.1023/B:NUMA.0000027762.08431.64. [DOI] [Google Scholar]
  • 28.Bertero M., Boccacci P., Ruggiero V. Inverse Imaging with Poisson Data. IOP Publishing; Bristol, UK: 2018. pp. 2053–2563. [DOI] [Google Scholar]
  • 29.Banham M., Katsaggelos A. Digital image restoration. IEEE Signal Process. Mag. 1997;14:24–41. doi: 10.1109/79.581363. [DOI] [Google Scholar]
  • 30.Wang Z., Bovik A.C., Sheikh H.R., Simoncelli E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004;13:600–612. doi: 10.1109/TIP.2003.819861. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Journal of Imaging are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES