Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2018 Aug 16;12(1):e201800129. doi: 10.1002/jbio.201800129

Tensor regularized total variation for denoising of third harmonic generation images of brain tumors

Zhiqing Zhang 1,2,3,, Marie L Groot 1,3, Jan C de Munck 2
PMCID: PMC7065612  PMID: 29959831

Abstract

Third harmonic generation (THG) microscopy shows great potential for instant pathology of brain tissue during surgery. However, the rich morphologies contained and the noise associated makes image restoration, necessary for quantification of the THG images, challenging. Anisotropic diffusion filtering (ADF) has been recently applied to restore THG images of normal brain, but ADF is hard‐to‐code, time‐consuming and only reconstructs salient edges. This work overcomes these drawbacks by expressing ADF as a tensor regularized total variation model, which uses the Huber penalty and the L1 norm for tensor regularization and fidelity measurement, respectively. The diffusion tensor is constructed from the structure tensor of ADF yet the tensor decomposition is performed only in the non‐flat areas. The resulting model is solved by an efficient and easy‐to‐code primal‐dual algorithm. Tests on THG brain tumor images show that the proposed model has comparable denoising performance as ADF while it much better restores weak edges and it is up to 60% more time efficient.

graphic file with name JBIO-12-e201800129-g007.jpg

Keywords: anisotropic diffusion, convex optimization, tensor regularization, third harmonic generation, weak edges


Third harmonic generation (THG) microscopy is a label‐free imaging technique that shows great potential for in vivo instant pathology of human brain tissue. Here we have developed a robust and efficient image denoising method to facilitate the clinical applications of THG.

graphic file with name JBIO-12-e201800129-g008.jpg

1. INTRODUCTION

Third harmonic generation (THG) microscopy 1, 2, 3 is a non‐linear imaging technique for label‐free three‐dimensional (3D) imaging of live tissues without the need for exogenous contrast agents. THG microscopy has established itself as an important tool for studying intact tissues such as insect embryos, plant seeds and intact mammalian tissue 2, epithelial tissues 4, 5, 6, zebrafish embryos 3, 7 and zebrafish nervous system 8. This technique has been applied for in vivo mouse brain imaging, revealing rich morphological information 9. Brain cells appear as dark holes on a bright background of neuropil, and axons and dendrites appear as bright fibers. More important, THG microscopy has shown great potential for clinical applications. Excellent agreement with the standard histopathology of skin cancers has been demonstrated for THG 10, 11 and THG also shows great potential for breast tumor diagnosis 12, 13.

In particular, we have recently demonstrated that THG yields high‐quality images of fresh, unstained human brain tumor tissue 14. Increased cellularity, nuclear pleomorphism, and rarefaction of neuropil have been clearly recognized in the acquired THG images of human brain tissue. This finding significantly facilitates the in vivo pathology of brain tumors and helps to reveal the tumor margins during surgery, which will improve the surgical outcomes.

Reliable image processing tools will strengthen the potential of THG microscopy for in vivo brain tumor pathology. In the image analysis pipeline of THG images of brain tissue, image denoising is essential and challenging due to the rich cellular morphologies and the low signal‐to‐noise ratio (SNR) 15. Anisotropic diffusion filtering (ADF) lies in the core of image denoising techniques that are able to remove strong image noise while maintaining the edges of objects sharp 16, 17, 18. The structure tensor is responsible for capturing the distribution of local gradients, thus enabling ADF to reconstruct certain kinds of structures, such as one‐dimensional (1D) flow‐like 17, 19, 20 and two‐dimensional (2D) membrane‐like structures 21, 22, as well as 2D blob and ridges 23. In a previous study 24, we have applied the classical edge‐enhancing ADF model 16 to restore the “dark” brain cells observed in THG images of mouse brain tissue. We have further developed in 15 a salient edge‐enhancing ADF model to reconstruct the rich morphologies appearing in THG images of structurally normal human brain tissue. However, all the existing ADF models have the drawback that the restored edges are in fact smooth 25. So far, most ADF models 19, 20, 21 are implemented using an explicit or semi‐implicit scheme 17, 26 to solve the diffusion equation which converges slowly.

The combination of ADF and the total variation (TV) model 27, 28, 29 provides an approach to overcome the drawbacks of the ADF models. TV regularization is another standard denoising method that has been studied mathematically for over decades 30, 31, 32, 33, 34, 35. In 25, an ADF model is formulated as a tensor regularized total variation (TRTV) model to restore the truly sharp edges, but the presented algorithm is based on gradient descent and has a slow convergence rate. The adaptive TRTV (ATRTV) model 36 improves convergence by using the primal‐dual algorithm 35 to solve the accompanying convex optimization model. The structure tensor adapts to the local geometry at each point but the estimated tensor may not reflect the true local structures if the image is corrupt by strong noise. Other important regularization approaches include the structure tensor total variation (STV) 37, 38 that penalizes the eigenvalues of the structure tensor, but STV does not make use of the directional information 36. The higher‐order regularizations such as the total generalized variation 39 and the Hessian Schatten‐Norm regularization (HS) 40 have also been proposed and also ignore direction of derivatives. There are many important alternative approaches to the image denoising problem such as dictionary learning based methods 41, sparse representation based methods 42, non‐local based methods 43, prior learning based methods 44, 45, 46, low‐rank based methods 47 and deep learning based methods 48.

In this study, we present a robust and efficient TRTV model that inherits the advantages of both ADF and TV, that is, their abilities of suppressing strong noise, estimating and restoring complex structures, and efficient convergence, to reconstruct 2D and 3D THG images of human brain tissue. The contributions of this study are 3‐fold. First, the pointwise decomposition of a structure tensor, which is time‐consuming and necessary for both ADF and TRTV, is greatly accelerated by performing the tensor decomposition only in the non‐flat areas. We use the gradient magnitude of a Gaussian at each point to estimate the first eigenvalue of the structure tensor and to distinguish flat from non‐flat areas. In the flat areas, the identity matrix is used as the diffusion tensor and no tensor decomposition is needed, while in the non‐flat regions, the tensor decomposition is applied to construct the application‐driven diffusion tensor. Second, existing TRTV models adopt the L2 norm for the data fidelity term while we use the L1 norm to make the proposed model (TRTV‐L1) robust to outliers and image contrast invariant. In previous work, it has been shown that geometrical features are better preserved by the TV models with the L1 norm 49. Third, we solve the TRTV‐L1 model with an efficient and easy‐to‐code primal‐dual algorithm as in 35, 36. In a detailed comparison of methods we show the ability of the TRTV‐L1 model to reconstruct weak edges, which is not well possible with other TRTV models. Weak edges are commonly observed in THG images and are important for clinical applications.

This work is a considerably extended version of the robust TRTV model previously presented at a conference 50. The rest of this paper is organized as follows: we review the existing TRTV models in Section 2. The proposed TRTV‐L1 model is explained in detail in Section 3. Simulated and real THG images are tested to demonstrate the efficiency and robustness of the proposed TRTV‐L1 model in Section 4. Conclusions follow in Section 5.

2. RELATED WORK

2.1. Anisotropic diffusion filtering and regularization

Let u denote an m‐dimensional (m = 2 or 3) image, and f be the noisy image. An ADF model 16, 17, 18, 19, 20, 21, 22, 23, 24, 51 has originally been defined by the partial differential equation (PDE) as follows:

tu=divDu,uxt=0=f, (1)

together with an application‐driven diffusion tensor D, where the raw image f is used as the initial condition. D is computed from the gradient of a Gaussian smoothed version of the image ∇uσ in 3 consecutive steps. First, the structure tensor J is computed at each point to estimate the distribution of the local gradients:

Jρuσ=Kρ*uσuσ,ρ0. (2)

Here uσ is the Gaussian smoothed version of u, that is, u is convolved with a Gaussian kernel K of SD, σ,

Kσ*u=mKσxyuydy, (3)
Kσ=12πσ2m/2expx2σ22. (4)

The SD, σ, denotes the noise scale of the target image 17. To study the distribution of the local gradients, the outer product of ∇uσ is computed and each component of the resulting matrix is convolved with another Gaussian K of SD, ρ. ρ is the integration scale that reflects the characteristic size of the texture, and usually it is large in comparison to the noise scale σ 17.

Second, the structure tensor J is decomposed into the product of a diagonal matrix with eigenvalues μi and a matrix of eigenvectors qi that indicate the distribution of the local gradients 17:

J=QdiagμiQT. (5)

The diagonal matrix, diag(μ i), is the eigenvalue matrix of all the eigenvalues ordered in the descending order, and the matrix Q is formed by the corresponding eigenvectors qi.

Finally, the eigenvalue values in (5) are replaced by the application‐driven diffusion matrix diag(λ i):

D=QdiagλiQT, (6)

where λi represents the amount of diffusivity along the eigenvector qi.

By taking the input image f as the starting point and evolving Eq. 1 over some time, the image is smoothed in flat areas and along the object edges, whereas the prominent edges themselves are maintained. Both the explicit and semi‐implicit schemes 17 have been widely employed to implement Eq. 1. The explicit scheme is easy‐to‐code yet converging slowly. The semi‐implicit scheme is more efficient because a larger time step is allowed, but harder to code because the inverse of a large matrix is involved.

Mathematically, Eq. 1 closely relates to the regularization problem that is designed to achieve a balance between smoothness and closeness to the input image f:

minuRu+λuf. (7)

In this functional, the first term is the regularization term (regularizer) that depends on the diffusion tensor D. The second term is the data fidelity term that uses a mathematical norm ||.|| to measure the closeness of u to the input image f. The implementation of this functional therefore depends on the construction of the diffusion tensor D, the choice of the regularizer and the fidelity norm. If we use the L2 norm for the data fidelity and substitute:

Ru=ΩuTDu=ΩSu2,withD=STS, (8)

into Eq. 7, its E‐L equation has the form:

tu=divDuλuf, (9)

which has the same diffusion tensor as (1). Because ∇u appears quadratically, R behaves as a L2 regularizer which has been shown unable to recover truly sharp edges 25, and the relation between (9) and (1) explains why the output of ADF is intrinsically smooth.

2.2. Total variation

Another standard image denoising method is the TV model that was introduced into compute vision first by Rudin, Osher and Fatemi (ROF) 27 as follows:

minuΩu+λuf22. (10)

The TV regularization penalizes only the total height of a slope but not its steepness, which permits the presence of edges in the minimizer. Although the ROF model permits prominent edges, it tends to create the so‐called stair‐casing effect and the primal minimization method used converges slowly. To address these drawbacks, several modifications have been made to reduce the stair‐casing effect and accelerate the convergence rate: replacing the TV regularization by Huber regularization, replacing the L2 norm by the L1 or Huber norm 52, and solving the convex minimization problem by the Chambolle's dual method 31, the split Bregman method 33, or the hybrid primal‐dual method 30, 32, 34, 35, 53. These first‐order primal‐dual algorithms enable easy‐to‐code implementation of the TV model. However, all these methods cannot properly remove the noise on the edges and cannot restore certain structures like 1D line‐like structures, because only the modulus of a gradient is considered in the regularizer, not its directions. Total variation based methods have also been applied to other image processing fields such as compressive sensing, mixed noise removal and image deblurring of natural and brain images 54, 55, 56.

2.3. Anisotropic total variation filtering

In order to overcome the problems of ADF and TV and combine their benefits, it is helpful to notice the close relation between diffusion filtering and regularization, which was initially studied in 57 for isotropic diffusion. The relation between anisotropic diffusion and the TV regularization was studied in 25, via the TRTV model as follows:

minuΩSu+λ2uf22=minuΩuTDu+λ2uf22. (11)

The matrix S satisfies D = STS, with a given diffusion tensor D. The anisotropic regularizer used in (11) overcomes the drawbacks of ADF and reconstructs truly sharp edges. Because the directional information has been incorporated via the diffusion tensor D in this model, it is also able to remove the noise on the edges and restore the complex structure which is not possible with the TV model. Despite these improvements, the minimization used in 25 to solve this TRTV model was based on gradient descent which suffered from slow convergence.

The diffusion behavior of (11) can be analyzed in terms of the diffusion equation given by its E‐L equation:

tu=divDSuuλuf. (12)

The first term on the right corresponds to an ADF with the diffusion tensor D/ ∣ S∇u.

2.4. Adaptive regularization with the structure tensor

In 36, the convexity of the problem (11) was used to improve the computational performance of the TRTV model 25 by applying the primal‐dual algorithm 35, to solve the convex optimization of the proposed ATRTV model 36. Also, the Huber penalty gα was used to regularize the structure tensor and reduce the stair‐casing effect caused by the TV regularization:

minuΩgαSu+λ2uf22, (13)
gαu=u22αifu<α,uα2ifuα. (14)

Here the matrix S, with S=maxαμ1+μ24diagμi1/2QT, is the adaptive tensor used to rotate and scale the axes of the local ellipse to coincide with the coordinate axes of the image domain. This design of the adaptive regularizer has taken into account the local structure of each point to penalize image variations. However, we noticed that the asymmetry of S may create artifacts and reduce the applicability of the algorithm in practice. We also note that the diffusion strength along the ith direction is approximately proportional to 1/μi41, which is not enough to suppress the noise when the input is corrupted by strong noise.

3. MATERIALS AND METHODS

3.1. Image samples and acquisition

All procedures on human tissue were performed with the approval of the Medical Ethical Committee of the VU University Medical Center and in accordance with Dutch license procedures and the declaration of Helsinki. All patients gave a written informed consent for tissue biopsy collection and signed a declaration permitting the use of their biopsy specimens in scientific research. We imaged brain tissue samples from 6 patients diagnosed with low‐grade glioma and 2 patients diagnosed with high‐grade glioma, as well as 2 structurally normal references with THG microscopy 14. Structurally normal brain samples were cut from the temporal cortex and subcortical white matter that had to be removed for the surgical treatment of deeper brain structures affected by epilepsy. Tumor brain samples were cut from tumor margin areas and from the tumor core and peritumoral areas. For details of the imaging setup, the tissue preparation and the tissue histology, we refer to previous works 9, 14.

3.2. The proposed tensor regularized total variation

When applied to THG images of brain tissue, all the methods above have their specific problems. The ADF models are computationally expensive and they cannot restore weak edges. The TV model creates the stair‐casing effect and cannot restore thin 1D line‐like structures. The existing TRTV models are either too expensive in computation or lack of enough denoising capability. To deal with these drawbacks and to make the TRTV approach applicable to THG images corrupted by strong noise, we present an efficient estimation of the diffusion tensor and we replace the L2 norm used in the data fidelity term by the robust L1 norm. We solve the resulting model by an efficient primal‐dual method.

3.2.1. Efficient estimation of the diffusion tensor

One time‐consuming step of the ADF and TRTV models is that the diffusion matrix D or S needs to be estimated at each point to describe the distribution of local gradients. This is of no interest in flat areas because the gradients almost vanish. In 3D, this tensor decomposition procedure takes about half of the total computational time. If the tensor decomposition is only computed in non‐flat areas, the procedure will be substantially accelerated.

To do this, we exploit the fact that the flat regions consist of points whose first (largest) eigenvalue is small, and that this eigenvalue can be roughly estimated by |∇uσ|2 16. This fact motivates the idea of thresholding |∇uσ|2 to distinguish flat and non‐flat regions. Before thresholding, we use the following function g to normalize and scale exponentially |∇uσ|2 to the range [0,1]:

gs=expC4s/λ4,s>0. (15)

This function has been used in the edge‐enhancing ADF model 16 to define the diffusivity along the first direction. Following 16 we set C 4 = 3.31488. λ is the threshold to control the trend of the function 16. Then we regard the points with g(|∇uσ|2) < h (here h is always set to 0.9) as the flat regions and the other points as the non‐flat regions. In the flat regions, the diffusion along each direction is isotropic and the diffusion tensor D reduces to the identity matrix I. In the non‐flat regions, the diffusion tensor D is defined as a weighted sum of the identity matrix and the application‐driven diffusion tensor, with the weight g(|∇uσ|2):

D=1guσ2I+guσ2QdiagλiQT=Q1guσ2I+guσ2diagλiQTQΛQT. (16)

Therefore, the g(|∇uσ|2) has two roles here, one of which is acting as a threshold value and the other is acting as the weight for constructing the diffusion tensor D of the non‐flat areas. Note that most of the ADF and TRTV models could in principle be accelerated using the procedure described here with almost no loss of accuracy. When applied to 3D images, we use the following eigenvalue system to optimize the diffusivityλi:

λ1=1guσ2,λ2=λ1λ1λ3hτCplane,λ3=1. (17)

For 2D THG images, the second diffusivity λ2 is ignored. hτ(⋅) is a fuzzy threshold function between 0 and 1 that allows a better control of the transition between 2D plane structures and other regions 21, 58, as follows:

hτx=tanhγxτ+1tanhγ1.0τ+1,x01. (18)

where γ is a scaling factor that controls the transition and we set it to 100. Cplane is the plane‐confidence measure 21, 59 defined as follows:

Cplane=μ1μ2μ1+μ2. (19)

Smoothing behaviors of the diffusion matrix (17) are different for different regions: in background regions, λ1 is almost 1 and smoothing is encouraged from all the directions at an equal level (isotropic smoothing). In the vicinity of edges, λ1 ≈ 0, smoothing at the first direction is discouraged. In plane‐like regions, the fuzzy function hτ tends to 1, and λ2 = 1, and smoothing at the second and third directions is allowed. In 1D structure regions, λ2 tends to λ1 and both are close to 0. Smoothing at the third direction is allowed only.

3.2.2. Robust anisotropic regularization

Given a diffusion tensor D designed as (16), we consider the same regularizer as in Eq. 13 of the adaptive TRTV model 36:

Ru=ΩgαSu, (20)

but contrary to 36 we use a symmetric S, S = D. To analyze the behavior of this regularizer in terms of diffusion, we note that the E‐L equation that minimizes R(u) is:

tu=div1maxαSuSTSu. (21)

To analyze the diffusion behavior along each eigenvector direction, we only need to estimate the S∇u:

Su=QΛQTu=QTuTΛ2QTu=i=13Λi2qiTu2i=13Λi2μi. (22)

Hence, the regularization problem (20) is a scaled version of the diffusion problem with the diffusion tensor STS = QΛ2QT, whose behavior along each eigenvector is almost the same as the diffusion problem with diffusion tensor D. Note that in the flat regions, S becomes the identity matrix, and the regularization (20) reduced to the Huber regularization.

3.2.3. Tensor regularized total variation‐L1

Different from the existing TRTV models, we consider the robust minimization problem as follows:

minuΩgαSu+λuf1, (23)

where we have used the L1 norm in the data fidelity term. Compared to the L2 norm, the L1 norm is image contrast invariant, robust to noise and sensitive to fine details 49, 60.

3.2.4. Numerical minimization

To efficiently solve the minimization problem (23), we note that it is a convex problem which can be reformulated as a saddle‐point problem. Therefore, it can be solved efficiently by the primal‐dual approach 34, 35, 36. To describe the problem in matrix algebra language, we reorder the image matrix u row‐wisely into a vector with N points, that is, u ∈ ℝN. The minimization problem (23) is written as the following primal minimization problem:

minuJAu+λuf1, (24)

where Au(i) = S(i)∇u(i) at each point i, and J denotes the Huber norm, JAu=i=1NgαAui.

To convert problem (24) into a primal‐dual problem, we introduce a dual variable p ∈ ℝmN (m = 2 or 3, the dimension of an image), and the convex conjugate of J (we refer to 61 for a complete introduction to the classical theory of convex analysis) is:

J*p=supqmNpqJq. (25)

Since J** = J, we have

JAu=supppAuJ*p. (26)

Substituting (26) into (24), we obtain the equivalent saddle‐point problem of the minimization problem (24):

minumaxppAuJ*p+λuf1. (27)

According to the hybrid primal‐dual algorithm described in 34, 35, we need to solve the following dual, primal and approximate extra‐gradient steps iteratively,

pk+1=argmaxppAu¯kJ*p, (28)
uk+1=argminuA*pk+1u+λuf1, (29)
u¯k+1=uk+1+θuk+1uk,θ01. (30)

Similar to 35, 36, the maximization problem (28) has the closed‐form solution:

p^=pk+τ1Au¯k,pk+1i=p^imax1+ατ1p^i. (28a)

where τ1 is the dual step size and α is defined in the Huber regularization in (14) and (23). For an intuitive understanding of (28a), we note that J * can be interpreted as the indicator function for the unit ball in the dual norm,

J*p=0,p*1,,otherwise (31)

and then problem (28) is equivalent to solve the dual problem:

pk+1=argmaxpXpAu¯k, (32)

where X = {p, J*(p) ≤ 1}. Since the ascend direction of (32) is Au¯k, (28a) can be considered as updating p along the ascend direction and projecting p onto X.

We solve the primal problem (29) with the primal algorithm described in 35, where the L1 norm can be solved by the pointwise shrinkage operations:

u^k+1=ukτ2A*pk+1,uk+1i=u^iτ2λifu^ifi>τ2λ,u^i+τ2λifu^ifi<τ2λ,fiifu^ifiτ2λ. (29a)

Here τ2 is the primal step size and the conjugate of A is:

A*pk+1i=divSiTpk+1i. (33)

Problem (24) is convex and the efficiency of the proposed algorithm comes from the ability to find closed‐form solutions for each of the sub‐problems. We summarize the proposed algorithm, including the estimation of the diffusion tensor, in Algorithm 1. This algorithm is partially inspired by the work of Estellers et al. 36. Note that we use the forward differences to compute the discrete gradients and backward differences for the divergence to preserve their adjoint relationshipdiv =  − *.

Algorithm 1

The efficient algorithm for the convex minimization problem (24).

Initialization. Set u¯0=u0=f,p0=0,k=0. Choose the initial step size τ1, τ2 > 0 and θ ∈ [0, 1].

While (||u k + 1 − u k|| > ε).

  1. Compute the structure tensor J, using (2).

  2. Construct diffusion matrix S: in the flat areas, set S as the identity matrix; otherwise, compute S using (16).

  3. Update pk, uk and u¯k iteratively, using (28a), (29a) and (30).

  4. Set k = k + 1.

Output: u k + 1, when ||u k + 1 − u k|| ε is satisfied.

4. EXPERIMENTAL RESULTS

We validate the proposed TRTV‐L1 model on a 2D simulated image, and around 200 2D and 3D THG images of normal human brain and tumor tissue. The field of view of the 2D and 3D THG images is 273 × 273 μm2 (1125 × 1125 pixels) and 273 × 273 × 50 μm3 (1125 × 1125 × 50 voxels), respectively. The intensities of these images are scaled to [0, 255]. We have previously developed a salient edge‐enhancing ADF model (the SEED model) to process the THG images of normal brain tissue 15, while the images of tumor tissue have not been published for the purpose of image analysis before. We compare our 2D results with the TV model 34, the edge‐enhancing ADF model (the EED model) 16, the BM3D model 62, the HS model 40, the STV model 38, the ATRTV model 36 and our previous SEED model 15. We only compare our 3D results with the TV model and the SEED model because not all source codes are readily available for other models in 3D. A comparison between EED and SEED has already been made in 15 for 3D.

4.1. Implementation

The proposed TRTV‐L1 model and ADFs are implemented in Visual Studio C++ 2010 on a PC with 8 3.40‐GHz Intel(R) Core(TM) 64 processors and 8 GB memory. Multiple cores have been used to implement the 3D algorithms, and a single core has been used for the 2D implementation. The TV model is implemented using the primal‐dual algorithm described in 34. The ADF models are implemented in the semi‐implicit scheme 17. The Matlab source codes for the BM3D model 62, the HS model 40, the STV model 38 and the ATRTV model 36 are available online from the authors' websites. The parameters are manually optimized for each model. The key parameters used for the proposed TRTV‐L1 model involve λ = 0.15, τ1 = 0.02, τ2 = 8.5 and θ = 1.0 for 2D and λ = 0.15, τ1 = 0.05, τ2 = 1.5and θ = 1.0 for 3D. The convergence accuracy ε is set to 10−2.

4.2. Denoising effect

The performance of the proposed TRTV‐L1 model is first evaluated on a 2D simulated image (Figure 1). The simulated image consists of seven horizontal lines of the same width (255 pixels), but of different heights, 50, 30, 25, 10, 5, 3 and 1 pixels. The intensity of each line horizontally increases from 1 to 255, mimicking edges with varying gradients. Gaussian noise with SD of 60 is added to simulate strong noise. The TV model cannot remove the noise on the edges (blue square), creates stair‐casing effect, fails to restore the 1‐pixel line and restores the 3‐pixel line only partially. The ADF models, that is, the EED and SEED models, have the highest peak signal‐to‐noise ratio 36 and they provide the best denoising effect, but they also lose some weak edges of all the lines. The BM3D model has perfect performance on keeping fine details, for example, a large part of 1‐pixel line is kept, but it creates ripple‐like artifacts (yellow square) and its denoising effect is not comparable to the tensor methods. The HS model penalizes the second‐order derivatives and thus it is able to avoid the stair‐casing effect and capture blood‐vessel‐like structures, but it has limited denoising effect and creates dark‐dot‐like artifacts. The ATRTV model is able to get rid of most stair‐casing effect, but the noise on the edges (blue square) is not properly removed. This behavior remains for other parameter settings. A possible explanation could be that there is not enough diffusion strength along the edge direction, possibly caused by the design of diffusion tensor. Its ability of keeping fine details is also limited, for example, part of the 1‐pixel line is swiped out. STV suffers less stair‐casing effect than TV, but its performance on denoising and keeping fine details is also limited, because it does not consider the eigenvectors that are the key for restoring local structures. Our TRTV‐L1 model combines the benefits of the L1 norm and tensor regularization, and has a denoising performance that is comparable to the ADF models and higher than the other models. Moreover, TRTV‐L1 is also able to keep fine details as BM3D does. The weak edges of all the simulated lines are better restored by TRTV‐L1 than by other tensor methods and regularization methods.

Figure 1.

Figure 1

Comparison of denoising results on the simulated image

We then compare the performance of the proposed TRTV‐L1 model with the aforementioned models using around 200 THG images of normal human brain and tumor tissue. One 2D typical example of THG images of normal brain tissue from gray matter is depicted in Figure 2. Brain cells (mainly including neurons and glial cells) and neuropil (consisting of axons and dendrites) are the basic features in a human brain, which appear as dark holes with dimly seen nuclei inside and bright fibers, respectively. Brain cells and neuropil are sparsely distributed in gray matter. The strong noise and rich morphologies contained in these THG images make the image denoising challenging. The TV model is able to remove the noise but it causes the stair‐casing effect. It cannot restore the thin fiber‐like neuropil because it does not consider the distribution of the local gradients (blue square). The ADF models (the SEED result is similar to the EED result and thus it is omitted) already give very satisfying results. The noise has been properly removed, but a substantial amount of weak edges have been smoothed to some extent, for example, the weak edges of some fibers and dark brain cells (blue square), because they are equivalent to the anisotropic TV with the L2 regularizer. The BM3D keeps the most fine details (the thin neuropil in the blue square), but its denoising effect is limited and again it creates ripple‐like artifacts. Note that the parameter σ involved in BM3D reflects the noise level of an image and the result of σ = 100 shown in Figure 2 indicates that the noise level of THG images is comparable to Gaussian noise with SD of 100. The HS model has limited performance on suppressing the strong noise in THG images, the result seems a bit blurred and dark‐dot‐like artifacts are created as appeared in the simulated image (Figure 1, HS). The result of ATRTV is similar to TV (but with less stair‐casing effect), and it is not able to restore the thin neuropil with weak edges. The STV model causes little stair‐casing effect, and does not keep fine details due to the lack of directional information. Compared with BM3D and HS, our TRTV‐L1 model is able to keep reasonable amount of fine details yet has a significantly superior denoising performance. Compared to other tensor and regularization methods, TRTV‐L1 can keep all salient edges and many more weak edges and fine details. TRTV‐L1 also provides the best image contrast and suffers almost no stair‐casing effect, because of the L1 norm and the robust anisotropic regularizer used. Results presented in Supporting Information (Figures S1‐S7) indicate that the parameter settings in Figure 2 are optimal for BM3D, HS, ATRTV and STV. The comparison of the segmentations resulted from the denoised images (Figures S5 and S6) not only confirms our qualitative evaluation of the denoising performance but also suggests that the denoising effort of TRTV‐L1 can really benefit the following segmentation step.

Figure 2.

Figure 2

One 2D THG image of normal brain tissue from gray matter. Brain cells and neuropil appear as dark holes with dimly seen nuclei inside and bright fibers, respectively

3D THG images of normal brain tissue from white matter (Figure 3) are adequate testing materials to demonstrate the 3D performance of the proposed model, because of the presence of the complex morphologies, for example, nets of neuropil. The density of brain cells, for example, neurons with dimly seen nucleus or with lipofuscin granules inclusions (blue arrow), is low but the density of neuropil is higher than in gray matter. We see that the noise has been removed by all the models. Nevertheless, the TV model cannot enhance the fiber‐like structures (the left blue square) and suffers from the stair‐casing effect. The SEED model is able to enhance the fibers, but some weak edges have been in fact smoothed (the blue square). Only our TRTV‐L1 model succeeds to reconstruct almost all the sharp and weak edges.

Figure 3.

Figure 3

One 3D THG example of normal brain tissue from white matter, with the 33th slice shown. More neuropil is observed than in gray matter

One 2D example of THG images of the low‐grade tumor tissue obtained from an oligodendroglioma patient is shown in Figure 4. Compared to the THG images of normal brain tissue, more brain cells (including cell nuclei and the surrounding cytoplasm) are present that indicates the presence of a tumor. Again, the TV model suffers from the stair‐casing effect. The ADF models fail to restore the weak edges (blue arrow). BM3D and HS have weaker denoising effect than other methods. BM3D creates ripple‐like artifacts and HS blurs the image. In contrast to the conventional approach for tensor estimation, ATRTV attempts to capture the directionality and scale of local structures via another convex approximation, but our results on THG images do not suggest superior merits of this aspect of ATRTV over the conventional approach in restoring local structures. The result of STV is similar to that of ATRTV due to the lack of directional information. Compared to other models, TRTV‐L1 either has better denoising performance and/or restores more fine details and weak edges (blue and yellow arrows).

Figure 4.

Figure 4

One 2D THG example of low‐grade tumor tissue from an oligodendroglioma patient. High cell density and thick neuropil indicate the presence of a tumor

One 2D example of THG images of the high‐grade tumor tissue obtained from a glioblastoma patient is shown in Figure 5. All the fiber‐like neuropil are now completely absent and the whole area is filled with cell nuclei. The density of cell nuclei here is even higher than that of the low‐grade tumor tissue, indicating that those cells likely represent tumor cells. The TV model is able to reconstruct both the salient and weak edges but it again causes the stair‐casing effect around the edges. The ADF models provide quite similar results without any stair‐casing effect, but the weak edges have been blurred. BM3D and HS have limited denoising effect. ATRTV and STV suffer less stair‐casing effect than TV, but the contrast seems degenerated. The proposed TRTV‐L1 model has reconstructed the salient and weak edges, which will greatly facilitate applications like automatic cell counting.

Figure 5.

Figure 5

One 2D THG example of high‐grade tumor tissue from a glioblastoma patient. The whole area is occupied by tumor cells

4.3. Computational performance

We first evaluate the computational cost of TRTV‐L1 that has been saved by restricting the tensor decomposition to the non‐flat areas. Roughly, the flat regions estimated in each iteration increase from 50% up to 90% of the whole image domain, and on average, 80% pixels are considered as flat regions, indicating that tensor decomposition is needed only for 20% of the pixels (Figure 6A). The reconstruction with and without full estimation of the structure tensor everywhere has been compared using THG images, from the aspects of timing and restoration quality. No significant difference in the number of iterations needed for convergence is observed between the full and partial estimation of structure tensor. For 2D THG images the partial estimation approach saves ~10% of computation time, either in terms of convergence time or time per iteration. No significant degradation has been found in the restoration quality (Figure 6B,C and Figure S7) when h varies from 0.0 to 0.9, and thus we use h = 0.9 to obtain maximal gain in speed. We also find that the absolution difference per pixel between the two reconstructions is 3.8, indicating the small difference between the 2 solutions. As a reference, the absolution difference per pixel between the reconstruction using partial estimation of structure tensor and the input noisy image is 54.4. For 3D THG images, ~40% of computation time is saved by the partial estimation approach. A visual map of non‐flat regions that results from the last iterative step is shown in Figure 6D. This map actually consists of all the sharp edges of the image, which conversely suggests that the weak edges are restored from the regularization and L1 fidelity rather than from the diffusion. Similar tests on the EED model indicate that the same computational gains can be achieved for the ADF models, using the partial tensor decomposition.

Figure 6.

Figure 6

Computational performance of the proposed TRTV‐L1 model. (A) The percentage of pixels (y‐axis) that are considered as points in the flat regions, in each iteration (x‐axis) during the tensor decomposition of a 2D THG image. (B, C) Comparison of the TRTV‐L1 results with (B) and without (C) full tensor decomposition everywhere. (D) The visual map of non‐flat regions that results from the last iterative step. (E) The time per iteration (y‐axis) needed for TRTV‐L1 and EED, tested on the 30 2D THG images

To demonstrate the computational efficiency of the proposed TRTV‐L1 model, we compare the average computational time needed by TRTV‐L1 to the ADF models on 30 2D and 5 3D THG images. The semi‐implicit scheme used to implement ADFs allows larger time steps than the explicit scheme. We found that ADFs converge much slower and consume more time per iteration than TRTV‐L1. For example, TRTV‐L1 on average needs only ~1/3 number of iterations of EED to converge to the same accuracy 10−2 and TRTV‐L1 consumes ~2/3 time of EED per iteration (Figure 6E), which results in a ~75% higher speed than EED. In practice, a fixed number of iteration is also a strategy to stop the iterations, and we find that both ADFs and TRTV‐L1 have already produced quite satisfying results after 50 iterations. In this condition, TRTV‐L1 is on average ~30% more time efficient than ADFs on 2D THG images, and ~60% more time efficient on 3D THG images. Compared to other tensor regularization models, our TRTV‐L1 model is roughly as efficient as the STV model, faster than the ATRTV model that uses another convex optimization to estimate structure tensor.

5. DISCUSSION AND CONCLUSIONS

In this work, we have developed a robust and efficient TRTV‐L1 model to restore images corrupted by noise. THG images of structurally normal human brain and tumor tissue have been tested. The proposed model showed impressively better results on the reconstruction of weak edges and fine details and it was more efficient than existing ADF and TRTV models. Comparisons to other state‐of‐art denoising techniques that are able to keep fine details, indicate that TRTV‐L1 can restore a reasonable amount of fine details but it has significantly better denoising performance without creating artifacts. The artifacts created by other models may result in false positives in subsequent segmentation steps. Therefore, the proposed TRTV‐L1 model will greatly facilitate the following segmentation and cell counting of THG images of brain tumor, from which we conclude that the robust and efficient TRTV model will strengthen the clinical potential of the THG microscopy on brain tumor surgery. Moreover, based on the tests on the simulated image and the THG images with complex morphologies, we believe that the proposed method can be generalized to other application‐driven projects. The efficient estimation of the diffusion tensor we proposed here can be used to accelerate most of the existing tensor diffusion and regularization models, by performing tensor decomposition only in the non‐flat regions. Compared to existing TRTV models, for example, the ATRTV model, the approach we combined the diffusion tensor and TV can be easily used to derive other application‐driven TRTV models from existing ADF models. The L1 norm in the data fidelity term makes the proposed TRTV‐L1 model contrast invariant, robust to noise and sensitive to fine details. The primal‐dual algorithm used to optimize the proposed model is easy‐to‐code in comparison with the existing ADF models because no sparse matrix inversions are involved. Although there are many other important types of image denoising methods as aforementioned, in this study we emphasize the benefits of tensor‐based techniques that they are able to capture local structures. Compared to other alternative approaches, for example, the machine learning methods, a training step is usually included which needs a training set of clear images with high SNR, but such images are difficult to acquire for THG brain images.

AUTHOR BIOGRAPHIES

Please see Supporting Information online.

Supporting information

Author Biographies

Figure S1 Results of BM3D for σ = 50, 100 and 150. The denoising effect of BM3D increases with σ. The denoising effect starts to occur when σ = 50, and achieves its optimal effect for σ = 100. Larger σ does not contribute to further improvement. These images show that BM3D creates ripple‐like artifacts and has limited denoising performance.

Figure S2 Results of Hessian Schatten‐Norm regularization (HS) for λ = 0.1, 0.3 and 0.5. The denoising effect starts to occur when λ = 0.1, and achieves its optimal performance for λ = 0.3. The result becomes too blurred when λ = 0.5. HS creates dark‐dot artifacts, has limited denoising performance and blurs the image.

Figure S3 Results of ATRTV for λ = 18, 10, 5 and μ = 8.6, 5.0 and 3.0. ATRTV has better denoising effect when λ and μ are small. The denoising effect starts to occur when λ = 18, μ = 8.6, and achieves its optimal performance for λ = 10, μ = 5.0. The result becomes blurred when λ and μ get smaller. The result of ATRTV is similar to that of TV, with less stair‐casing effect created, but it is not able to restore fine details and weak edges corrupted by strong noise.

Figure S4 Results of STV for λ = 0.24, 0.32 and 0.4. The denoising effect starts to occur when λ = 0.24, and achieves its optimal performance for λ = 0.4. The result of STV is similar to that of TV, with less stair‐casing effect created, but it is not able to restore fine details and weak edges corrupted by strong noise.

Figure S5 Segmentations of the dark holes (brain cells) within the raw image and the denoised images in Figure 2, using manually optimized thresholds to detect most parts of the dark holes with least background included. The segmentation of the raw image indicates the strong noise present in the THG image. The segmentations of TV, EED, ATRTV, STV and TRTV‐L1 are similar but the small objects resident in segmentations of BM3D and HS illustrate their poor denoising performance.

Figure S6 Segmentations of the bright objects (neuropil) within the raw image and the denoised images in Figure 2, using manually optimized thresholds to detect most parts of the bright objects with least background included (eg, the fiber indicated by yellow arrow). The segmentation of the raw image indicates the strong noise present. The segmentation of TRTV‐L1 is comparable to those of BM3D and HS, where more fibers have been resolved than other models. Sometimes fibers (blue arrows) are even better segmented from the image denoised by TRTV‐L1, which suggests that BM3D and HS could visually keep more details than TRTV‐L1 but it is not necessarily beneficial for the segmentation followed.

Figure S7 Results of the proposed TRTV‐L1 model for h = 0.0 (full estimation), 0.2, 0.5, 0.8 and 0.9 (partial estimation). Almost no degradation has been found in the restoration quality when h varies from 0.0 to 0.9, and thus we use h = 0.9.

ACKNOWLEDGMENTS

This research is funded by China Scholarship Council (CSC).

Zhang Z, Groot ML, de Munck JC. Tensor regularized total variation for denoising of third harmonic generation images of brain tumors. J. Biophotonics. 2019;12:e201800129. 10.1002/jbio.201800129

Funding information China Scholarship Council

REFERENCES

  • 1. Barad Y., Eisenberg H., Horowitz M., Silberberg Y., Appl. Phys. Lett. 1997, 70, 922. [Google Scholar]
  • 2. Debarre D., Supatto W., Pena A. M., Fabre A., Tordjmann T., Combettes L., Schanne‐Klein M. C., Beaurepaire E., Nat. Methods 2006, 3, 47. [DOI] [PubMed] [Google Scholar]
  • 3. Olivier N., Luengo‐Oroz M. A., Duloquin L., Faure E., Savy T., Veilleux I., Solinas X., Debarre D., Bourgine P., Santos A., Peyrieras N., Beaurepaire E., Science 2010, 329, 967. [DOI] [PubMed] [Google Scholar]
  • 4. Weigelin B., Bakker G.‐J., Friedl P., Dermatol. Int. 2012, 1, 32. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Adur J., Pelegati V. B., de Thomaz A. A., Baratti M. O., Almeida D. B., Andrade L. A., Bottcher‐Luiz F., Carvalho H. F., Cesar C. L., PLoS One 2012, 7, e47007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Wu P. C., Hsieh T. Y., Tsai Z. U., Liu T. M., Sci. Rep. 2015, 5, 8879. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Luengo‐Oroz M. A., Rubio‐Guivernau J. L., Faure E., Savy T., Duloquin L., Olivier N., Pastor D., Ledesma‐Carbayo M., Debarre D., Bourgine P., Beaurepaire E., Peyrieras N., Santos A., IEEE Trans. Image Process. 2012, 21, 2335. [DOI] [PubMed] [Google Scholar]
  • 8. Chen S. Y., Hsieh C. S., Chu S. W., Lin C. Y., Ko C. Y., Chen Y. C., Tsai H. J., Hu C. H., Sun C. K., J. Biomed. Opt. 2006, 11, 054022. [DOI] [PubMed] [Google Scholar]
  • 9. Witte S., Negrean A., Lodder J. C., De Kock C. P., Silva G. T., Mansvelder H. D., Groot M. L., Proc. Natl. Acad. Sci. 2011, 108, 5970. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Chen S. Y., Chen S. U., Wu H. Y., Lee W. J., Liao Y. H., Sun C. K., IEEE J. Sel. Top. Quant. Electron 2010, 16, 478. [Google Scholar]
  • 11. Lee G. G., Lin H. H., Tsai M. R., Chou S. Y., Lee W. J., Liao Y. H., Sun C. K., Chen C. F., IEEE Trans. Biomed. Circ. Syst. 2013, 7, 158. [DOI] [PubMed] [Google Scholar]
  • 12. Lee W., Kabir M. M., Emmadi R., K. C. Toussaint Jr.. , J. Microsc. 2016, 264, 175. [DOI] [PubMed] [Google Scholar]
  • 13. Gavgiotaki E., Filippidis G., Markomanolaki H., Kenanakis G., Agelaki S., Georgoulias V., Athanassakis I., J. Biophotonics 2017, 10, 1152. [DOI] [PubMed] [Google Scholar]
  • 14. Kuzmin N. V., Wesseling P., Hamer P. C., Noske D. P., Galgano G. D., Mansvelder H. D., Baayen J. C., Groot M. L., Biomed. Opt. Express 2016, 7, 1889. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Zhang Z., Kuzmin N. V., Groot M. L., de Munck J. C., Bioinformatics 2017, 33, 1712. [DOI] [PubMed] [Google Scholar]
  • 16. Weickert J., Anisotropic Diffusion in Image Processing, B. G. Teubner Press, Stuttgart; 1998. [Google Scholar]
  • 17. Weickert J., Int. J. Comput. Vision 1999, 31, 111. [Google Scholar]
  • 18. Cottet G. H., Germain L., Math Comput. 1993, 61, 659. [Google Scholar]
  • 19. Maska M., Danek O., Garasa S., Rouzaut A., Munoz‐Barrutia A., Ortiz‐de‐Solorzano C., IEEE Trans. Med. Imaging 2013, 32, 995. [DOI] [PubMed] [Google Scholar]
  • 20. Manniesing R., Viergever M. A., Niessen W. J., Med. Image Anal. 2006, 10, 815. [DOI] [PubMed] [Google Scholar]
  • 21. Pop S., Dufour A. C., Le Garrec J. F., Ragni C. V., Cimper C., Meilhac S. M., Olivo‐Marin J. C., Bioinformatics 2013, 29, 772. [DOI] [PubMed] [Google Scholar]
  • 22. Mosaliganti K., Janoos F., Gelas A., Noche R., Obholzer N., Machiraju R., Megason S. G., Proceedings of 2010 I.E. International Symposium on Biomedical Imaging: From Nano to Macro, IEEE, Rotterdam: 2010, p. 588. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Qiu Z., Yang L., Lu W. P., Pattern Recogn. Lett. 2012, 33, 319. [Google Scholar]
  • 24. Zhang Z., Kuzmin N. V., Groot M. L., de Munck J. C., J. Biophotonics 2018, 11, e201600256. [DOI] [PubMed] [Google Scholar]
  • 25. Grasmair M., Lenzen F., Appl. Math. Opt. 2010, 62, 323. [Google Scholar]
  • 26. Weickert J., Romeny B. M. T., Viergever M. A., IEEE Trans. Image Process. 1998, 7, 398. [DOI] [PubMed] [Google Scholar]
  • 27. Rudin L. I., Osher S., Fatemi E., Physica D 1992, 60, 259. [Google Scholar]
  • 28. Chang C. W., Mycek M. A., J. Biophotonics 2012, 5, 449. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Xie D., Li Q., Gao Q., Song W., Zhang H. F., Yuan X., J. Biophotonics 2018, e201700360. DOI: 10.1002/jbio.201700360 [DOI] [PubMed] [Google Scholar]
  • 30. Chan T. F., Golub G. H., Mulet P., Siam J. Sci. Comput. 1999, 20, 1964. [Google Scholar]
  • 31. Chambolle A., J. Math. Imaging Vis. 2004, 20, 89. [Google Scholar]
  • 32. Zhu M., Chan T., UCLA CAM Report 08‐34, 2008.
  • 33. Goldstein T., Osher S., Siam J. Imaging Sci. 2009, 2, 323. [Google Scholar]
  • 34. Esser E., Zhang X. Q., Chan T. F., Siam J. Imaging Sci. 2010, 3, 1015. [Google Scholar]
  • 35. Chambolle A., Pock T., J. Math. Imaging Vis. 2011, 40, 120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Estellers V., Soatto S., Bresson X., IEEE Trans. Image Process. 2015, 24, 1777. [DOI] [PubMed] [Google Scholar]
  • 37. Lefkimmiatis S., Roussos A., Unser M., Maragos P., Proceedings of 2013 International Conference on Scale Space and Variational Methods in Computer Vision, Springer, Leibnitz: 2013, p. 48. [Google Scholar]
  • 38. Lefkimmiatis S., Roussos A., Maragos P., Unser M., Siam J. Imaging Sci. 2015, 8, 1090. [Google Scholar]
  • 39. Bredies K., Kunisch K., Pock T., Siam J. Imaging Sci. 2010, 3, 492. [Google Scholar]
  • 40. Lefkimmiatis S., Ward J. P., Unser M., IEEE Trans. Image Process. 2013, 22, 1873. [DOI] [PubMed] [Google Scholar]
  • 41. Aharon M., Elad M., Bruckstein A., IEEE Trans. Signal Process. 2006, 54, 4311. [Google Scholar]
  • 42. Dong W. S., Zhang L., Shi G. M., Li X., IEEE Trans. Image Process. 2013, 22, 1618. [DOI] [PubMed] [Google Scholar]
  • 43. Buades A., Coll B., Morel J. M., Multiscale Model. Sim. 2005, 4, 490. [Google Scholar]
  • 44. Xu J., Zhang L., Zuo W., Zhang D., Feng X., Proceedings of the 2015 I.E. International Conference on Computer Vision, IEEE, Santiago: 2015, p. 244. [Google Scholar]
  • 45. Zoran D., Weiss Y., Proceedings of the 2011 I.E. International Conference on Computer Vision, IEEE, Barcelona: 2011, p. 479. [Google Scholar]
  • 46. Xu J., Zhang L., Zhang D., IEEE Trans. Image Process. 2018, 27, 2996. [Google Scholar]
  • 47. Xu J., Zhang L., Zhang D., Feng X., Proceedings of the 2017 I.E. International Conference on Computer Vision, IEEE, Venice: 2017, p. 1096. [Google Scholar]
  • 48. Burger H. C., Schuler C. J., Harmeling S., Proceedings of the 25th IEEE Conference on Computer Vision and Pattern Recognition, IEEE, Providence: 2012, p. 2392. [Google Scholar]
  • 49. Chan T. F., Esedoglu S., Siam J. Appl. Math. 2005, 65, 1817. [Google Scholar]
  • 50. Zhang Z., Groot M. L., de Munck J. C., Joint Conference of the European Medical and Biological Engineering Conference and the Nordic‐Baltic Conference on Biomedical Engineering and Medical Physics, Springer, Tampere: 2017, p. 129. [Google Scholar]
  • 51. Roussos A., Maragos P., Proceedings of the 17th IEEE International Conference on Image Processing, IEEE, Hong Kong: 2010, p. 4141. [Google Scholar]
  • 52. Aujol J. F., Gilboa G., Chan T., Osher S., Int. J. Comput. Vision 2006, 67, 111. [Google Scholar]
  • 53. Esser E., Zhang X., Chan T., UCLA CAM Report 09‐67, 2009.
  • 54. Zhang M. L., Desrosiers C., Zhang C. M., Pattern Recogn. 2018, 76, 549. [Google Scholar]
  • 55. Zhang M., Desrosiers C., Zhang C., Proceedings of the 42nd IEEE International Conference on Speech and Signal Processing, IEEE, New Orleans: 2017, p. 1802. [Google Scholar]
  • 56. Ren D. W., Zhang H. Z., Zhang D., Zuo W. M., Neurocomputing 2015, 170, 201. [Google Scholar]
  • 57. Scherzer O., Weickert J., J. Math. Imaging Vis. 2000, 12, 43. [Google Scholar]
  • 58. Romulus T., Lavialle O., Borda M., Baylou P., Lect. Notes Comput. Sci. 2005, 3708, 316. [Google Scholar]
  • 59. Van Kempen G. M. P., van den Brink N., van Vliet L. J., Van Ginkel M., Verbeek P. W., Blonk H., Proceedings of the Scandinavian Conference on Image Analysis, Vol. 1, Pattern Recognition Society of Denmark Press, Copenhagen, 1999, p. 447. [Google Scholar]
  • 60. Rousseeuw P. J., Leroy A. M., Robust Regression and Outlier Detection, John Wiley & Sons, New York 2005. [Google Scholar]
  • 61. Hiriart‐Urruty J.‐B., Lemaréchal C., Convex Analysis and Minimization Algorithms I: Fundamentals, Springer Science & Business Media, Berlin 2013. [Google Scholar]
  • 62. Dabov K., Foi A., Katkovnik V., Egiazarian K., IEEE Trans. Image Process. 2007, 16, 2080. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Author Biographies

Figure S1 Results of BM3D for σ = 50, 100 and 150. The denoising effect of BM3D increases with σ. The denoising effect starts to occur when σ = 50, and achieves its optimal effect for σ = 100. Larger σ does not contribute to further improvement. These images show that BM3D creates ripple‐like artifacts and has limited denoising performance.

Figure S2 Results of Hessian Schatten‐Norm regularization (HS) for λ = 0.1, 0.3 and 0.5. The denoising effect starts to occur when λ = 0.1, and achieves its optimal performance for λ = 0.3. The result becomes too blurred when λ = 0.5. HS creates dark‐dot artifacts, has limited denoising performance and blurs the image.

Figure S3 Results of ATRTV for λ = 18, 10, 5 and μ = 8.6, 5.0 and 3.0. ATRTV has better denoising effect when λ and μ are small. The denoising effect starts to occur when λ = 18, μ = 8.6, and achieves its optimal performance for λ = 10, μ = 5.0. The result becomes blurred when λ and μ get smaller. The result of ATRTV is similar to that of TV, with less stair‐casing effect created, but it is not able to restore fine details and weak edges corrupted by strong noise.

Figure S4 Results of STV for λ = 0.24, 0.32 and 0.4. The denoising effect starts to occur when λ = 0.24, and achieves its optimal performance for λ = 0.4. The result of STV is similar to that of TV, with less stair‐casing effect created, but it is not able to restore fine details and weak edges corrupted by strong noise.

Figure S5 Segmentations of the dark holes (brain cells) within the raw image and the denoised images in Figure 2, using manually optimized thresholds to detect most parts of the dark holes with least background included. The segmentation of the raw image indicates the strong noise present in the THG image. The segmentations of TV, EED, ATRTV, STV and TRTV‐L1 are similar but the small objects resident in segmentations of BM3D and HS illustrate their poor denoising performance.

Figure S6 Segmentations of the bright objects (neuropil) within the raw image and the denoised images in Figure 2, using manually optimized thresholds to detect most parts of the bright objects with least background included (eg, the fiber indicated by yellow arrow). The segmentation of the raw image indicates the strong noise present. The segmentation of TRTV‐L1 is comparable to those of BM3D and HS, where more fibers have been resolved than other models. Sometimes fibers (blue arrows) are even better segmented from the image denoised by TRTV‐L1, which suggests that BM3D and HS could visually keep more details than TRTV‐L1 but it is not necessarily beneficial for the segmentation followed.

Figure S7 Results of the proposed TRTV‐L1 model for h = 0.0 (full estimation), 0.2, 0.5, 0.8 and 0.9 (partial estimation). Almost no degradation has been found in the restoration quality when h varies from 0.0 to 0.9, and thus we use h = 0.9.


Articles from Journal of Biophotonics are provided here courtesy of Wiley

RESOURCES