Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2023 Jul 4;23(13):6135. doi: 10.3390/s23136135

Multi-Focus Image Fusion via Distance-Weighted Regional Energy and Structure Tensor in NSCT Domain

Ming Lv 1, Liangliang Li 2, Qingxin Jin 3, Zhenhong Jia 1,*, Liangfu Chen 4, Hongbing Ma 5,*
Editor: Alessandro Bevilacqua
PMCID: PMC10346789  PMID: 37447984

Abstract

In this paper, a multi-focus image fusion algorithm via the distance-weighted regional energy and structure tensor in non-subsampled contourlet transform domain is introduced. The distance-weighted regional energy-based fusion rule was used to deal with low-frequency components, and the structure tensor-based fusion rule was used to process high-frequency components; fused sub-bands were integrated with the inverse non-subsampled contourlet transform, and a fused multi-focus image was generated. We conducted a series of simulations and experiments on the multi-focus image public dataset Lytro; the experimental results of 20 sets of data show that our algorithm has significant advantages compared to advanced algorithms and that it can produce clearer and more informative multi-focus fusion images.

Keywords: multi-focus image, image fusion, distance-weighted regional energy, structure tensor, non-subsampled contourlet transform

1. Introduction

In order to obtain richer and more useful information and to improve the completeness and accuracy of scene description, multi-focus images obtained from video sensors are usually fused [1,2]. Multi-focus image fusion uses effective information processing methods to fuse clear and focused information from different images, resulting in a high-definition fully focused image [3,4,5]. Figure 1 depicts an example of multi-focus image fusion.

Figure 1.

Figure 1

Multi-focus image fusion. (a) Near-focused image; (b) far-focused image; and (c) fused image.

After many years of extensive and in-depth research, the image fusion method has made significant progress. Image fusion approaches can be divided into three categories: transform domain-based, edge-preserving filtering-based, and deep learning-based algorithms [6,7,8,9].

The most widely used transform domain-based methods are the curvelet transform [10], contourlet transform [11,12], shearlet transform [13,14], etc. Kumar et al. [15] constructed an intelligent multi-modal image fusion technique utilizing the fast discrete curvelet transform and type-2 fuzzy entropy, and the fusion results demonstrate the efficiency of this model in terms of subjective and objective assessment. Kumar et al. [16] constructed an image fusion technique via an improved multi-objective meta-heuristic algorithm with fuzzy entropy in fast discrete curvelet transform domain; a comparison of the developed methodology over the state-of-the-art models observes enhanced performance with respect to visual quality assessment. Li et al. [17] integrated the curvelet and discrete wavelet transform for multi-focus image fusion, and it can obtain advanced fusion performance. Zhang et al. [18] constructed an image fusion model using the contourlet transform. The average-based fusion rule and region energy-based fusion rule are performed on the low- and high-frequency sub-bands, respectively; the fused image is analyzed using fast non-local clustering for multi-temporal synthetic aperture radar image change detection; and this technique generates state-of-the-art change detection performance on both small- and large-scale datasets. Li et al. [19] constructed an image fusion approach using sparse representation and local energy in shearlet transform domain; this method can obtain good fusion performance. Hao et al. [20] introduced a multi-scale decomposition optimization-based image fusion method via gradient-weighted local energy; the non-subsampled shearlet transform is utilized to decompose the source images into low- and high-frequency sub-images, and then the low-frequency components are divided into base layers and texture layers, the base layers are merged according to the intrinsic attribute-based energy fusion rule, and the structure tensor-based gradient-weighted local energy operator fusion rule is utilized to merge the texture layers and high-frequency sub-bands. This fusion method can achieve a superior fusion capability in both qualitative and quantitative assessments.

The methods based on edge-preserving filtering also hold an important position in the field of image fusion, such as guided image filtering [21], cross bilateral filtering [22], Gaussian curvature filtering [23], sub-window variance filtering [24], etc. Feng et al. [24] constructed the multi-channel dynamic threshold neural P systems-based image fusion method via a visual saliency map in the sub-window variance filter domain; this algorithm can obtain effective fusion performance both on visual quality and quantitative evaluations. Zhang et al. [25] introduced a double joint edge preservation filter-based image fusion technique via a non-globally saliency gradient operator; it can obtain excellent subjective and objective performance. Jiang et al. [26] introduced the image fusion approach utilizing entropy measure between intuitionistic fuzzy set joint Gaussian curvature filtering.

The image fusion approaches based on deep learning have generated good performance, and these deep learning-based fusion models can be divided into supervised- and unsupervised-based approaches [27,28]. In terms of supervised algorithms, some effective fusion methods have been generated. The deep convolutional neural network introduced by Liu et al. [29] is used for multi-focus image fusion, and it generates good fusion results. The multi-scale visual attention deep convolutional neural network introduced by Lai et al. [30] is utilized to fuse multi-focus images. Wang et al. [31] introduced weakly supervised image fusion with modal synthesis and enhancement. The self-supervised residual feature learning model constructed by Wang et al. [32] is used to fuse multi-focus images. The attention mechanism-based image fusion approach via supervised learning was constructed by Jiang et al. [33]. In terms of unsupervised algorithms, Jin et al. [34] introduced the Transformer and U-Net for image fusion. Zhang et al. [35] presented another unsupervised generative adversarial network with adaptive and gradient joint constraints for image fusion. Liu et al. [36] constructed a generative adversarial network-based unsupervised back project dense network for image fusion. These deep learning-based fusion methods also achieve good performance.

In order to improve the fusion performance of multi-focus images, a distance-weighted regional energy-based multi-focus image fusion algorithm via the structure tensor in non-subsampled contourlet transform domain was constructed. The distance-weighted regional energy-based fusion rule and the structure tensor-based fusion rule were used to merge low- and high-frequency sub-images, respectively. The experimental results of the Lytro dataset show that the proposed fusion method can produce fusion effects that are superior to other traditional and deep learning methods in terms of subjective visual and objective evaluation.

2. Backgrounds

2.1. NSCT

The non-subsampled contourlet transform (NSCT) is a transformation with multi-directional, multi-scale, and translation invariance [12]. Its basic framework structure is divided into two parts: the non-subsampled pyramid (NSP) decomposition mechanism and the non-subsampled directional filter bank (NSDFB) decomposition mechanism. The NSCT first utilizes NSP decomposition to perform multi-scale decomposition on the source image and then uses NSDFB decomposition to further decompose the high-frequency components in the direction, ultimately obtaining sub-band images of the source image at different scales and in different directions [12]. Figure 2 shows the NSCT basic framework structure diagram and the NSCT frequency domain partition diagram.

Figure 2.

Figure 2

(a) NSCT basic framework structure diagram; (b) NSCT frequency domain partition diagram.

2.2. Structure Tensor

For a multi-channel image, fx,y=f1x,y, f2x,y,, f3x,y, the grayscale images and color images are two special cases when m = 1 and m = 3, respectively. The square of variation of fx,y at position x,y in direction θ for any ε0+ can be given by the following [37]:

df2=fx+εcosθ,y+εsinθfx,y22 i=1mfixεcosθ+fiyεsinθ2 (1)

The rate of change Cθ of image fx,y at position x,y can be expressed as follows:

Cθ=i=1mfixcosθ+fiysinθ2=cosθ,sinθi=1mfix2i=1mfixfiyi=1mfiyfixi=1mfiy2cosθ,sinθT=cosθ,sinθi=1mfifiTcosθ,sinθT (2)

where fi=fix,fiyT, the following second-moment positive semi-definite matrix is

P=i=1mfifiT=EFFG (3)

where P is called the structure tensor and where E, F, and G are defined as

E=i=1mfix2 (4)
F=i=1mfixfiy (5)
G=i=1mfiy2 (6)

The structure tensor-based focus detection operator (STO) is given by

S=λ1+λ22+0.5λ1λ22 (7)

where λ1 and λ2 are the eigenvalues of the structure tensor and can be computed by the following:

λ1,2=12E+G±GE2+4F2 (8)

3. The Proposed Multi-Focus Image Fusion Method

The multi-focus image fusion method based on the distance-weighted regional energy and structure tensor in NSCT domain is proposed in this section, and the schematic of the proposed algorithm is depicted in Figure 3. This fusion method can be divided into four parts: NSCT decomposition, low-frequency components fusion, high-frequency components fusion, and the inverse NSCT. More details can be seen in the follows.

Figure 3.

Figure 3

The schematic of the proposed fusion algorithm.

3.1. NSCT Decomposition

The multi-focus images A and B were decomposed into low-frequency components and high-frequency components through the NSCT, and the corresponding coefficients are defined as LA,HAl,k and LB,HBl,k, respectively.

3.2. Low-Frequency Components Fusion

Low-frequency components contain more energy and information in the images. In this section, the distance-weighted regional energy (DWRE)-based rule was used to merge the low-pass sub-bands, and it is defined as follows [38]:

DWRELXi,j=p=11 q=11Wp+2,q+2LXi+p,j+q2 (9)

where W shows a 3×3 matrix that allocates weights to the neighboring coefficients and where W is defined as

W=11+21211+21211211+21211+2 (10)

Here, each entry of W is achieved by reciprocating ‘1 + distance of the respective position from its center, i.e., W2,2’, which is depicted in the following:

Wi,j=11+i22+j22 (11)

The fused low-frequency component LFi,j is constructed by the following:

LFi,j=LAi,j if DWRELAi,jDWRELBi,jLBi,j else (12)

where DWRELXi,j|XA,B is the DWRE of LXi,j estimated over a 3×3 neighborhood centered at the i,jth coefficient.

3.3. High-Frequency Components Fusion

High-frequency components contain more details and noise in the images. In this section, the structure tensor-based focus detection operator (STO) fusion rule was used to process the high-frequency components, and the fused coefficients HFl,ki,j can be computed by

HFl,ki,j=HXl,ki,j×M˜Xl,ki,jXA,B (13)

where

M˜Xl,ki,j=1 if a,bΩ1 MXl,ki+a,j+b>0.5×sizeΩ10 else (14)
MXl,ki,j=1 if argmaxSAl,ki,j,SBl,ki,j0 else (15)

where HXl,ki,j shows the high-frequency coefficient of image XA,B at the l-th scale and k-th direction at location i,j and where SXl,ki,j shows the structure salient image of HXl,ki,j computed by STO. Ω1 is a local region centered at i,j with the size of W1×V1. Equation (14) is the classical operator called the consistency verification, which can improve the robustness and rectify some wrong focusing detection of MXl,ki,j.

3.4. Inverse NSCT

The final fused image F was generated by the inverse NSCT, which was performed on fused low- and high-frequency components LF,HFl,k, and F is defined as follows:

Fi,j=LAi,j+HFl,ki,j (16)

The main steps of the proposed multi-focus image fusion approach are summarized in Algorithm 1.

Algorithm 1 Proposed multi-focus image fusion method
Input: the source images: A and B
Parameters: The number of NSCT decomposition levels: L, the number of directions at each decomposition level: K(l), l[1,L]
Main step:
Step 1: NSCT decomposition
For each source image X{A,B}
  Perform NSCT decomposition on X to generate {LX,HXl,k}, l[1,L], k[1,K(l)];
End
Step 2: Low-frequency components fusion
For each source image X{A,B}
  Calculate the DWRE for LX utilizing Equations (9)–(11);
End
Merge LA and LB utilizing Equations (12) to generate LF;
Step 3: High-frequency components fusion
For each Level l=1:L
  For each direction k=1:K(l)
   For each source image XA,B
    Calculate the structure salient image SXl,ki,j of HXl,ki,j computed by STO using Equation (7);
    Calculate the consistency verification using Equations (14) and (15);
   End
   Merge HAl,k and HBl,k using Equation (13)
  End
End
Step 4. Inverse NSCT
Perform inverse NSCT on LF,HFl,k to generate F;
Output: the fused image F.

4. Experimental Results and Discussions

In this section, the Lytro dataset with twenty pair images (Figure 4) are utilized to experiment, and eight state-of-the-art fusion algorithms are used to be compared with our method, namely, multi-focus image fusion using a bilateral-gradient-based sharpness criterion (BGSC) [39], the non-subsampled contourlet transform and fuzzy-adaptive reduced pulse-coupled neural network (NSCT) [40], multi-focus image fusion in gradient domain (GD) [41], the unified densely connected network for image fusion (FusionDN) [42], the fast unified image fusion network based on the proportional maintenance of gradient and intensity (PMGI) [43], image fusion based on target-enhanced multi-scale transform decomposition (TEMST) [44], the unified unsupervised image fusion network (U2Fusion) [45], and zero-shot multi-focus image fusion (ZMFF) [3]. Subjective and objective assessments are also used; in terms of objective assessment, eight metrics, including edge-based similarity measurement QAB/F [46], the feature mutual information metric QFMI [47], the gradient-based metric QG [48], the structural similarity-based metric QE [48], the phase congruency-based metric QP [48], the structural similarity-based metric QY introduced by Yang et al. [48], the average gradient metric QAG [22], and the average pixel intensity metric QAPI [22], are used. The larger the values of these indicators, the better the fusion effect. In the proposed method, the size of Ω1 is set as 5 × 5; the ‘vk’ and ‘pyrexc’ are used as the pyramid filter and directional filter, respectively; and the four decomposition levels with 2, 2, 2, and 2 directions from a coarser scale to a finer scale are used.

Figure 4.

Figure 4

Lytro dataset.

4.1. Qualitative Comparisons

In this section, we selected five sets of data from the Lytro dataset for result display; the qualitative comparisons of different methods are shown in Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9. We have enlarged some areas for easy observation and comparison. In Figure 5, we can see that the fusion result generated by the BGSC method has a weak far-focus fusion effect and that black spots appear around the building; the fusion results computed by the NSCT, GD, FusionDN, PMGI, and TEMST methods are somewhat blurry; the U2Fusion method generates a dark fusion image which limits the observation of some areas in the image; and the ZMFF method produces fully focused image results. However, compared to our method, our result produces moderate brightness and clear edges, especially the brightness of the red lock in the image, which is better than ZMFF.

Figure 5.

Figure 5

Figure 5

Lytro-06 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.

Figure 6.

Figure 6

Figure 6

Lytro-10 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.

Figure 7.

Figure 7

Figure 7

Lytro-13 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.

Figure 8.

Figure 8

Figure 8

Lytro-15 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.

Figure 9.

Figure 9

Figure 9

Lytro-18 fusion results. (a) BGSC; (b) NSCT; (c) GD; (d) FusionDN; (e) PMGI; (f) TEMST; (g) U2Fusion; (h) ZMFF; and (i) proposed.

In Figure 6, we can see that the fusion results generated by the BGSC and PMGI methods are blurry, making it difficult to observe the detailed information in the fused images; the fusion images computed by NSCT, FusionDN, TEMST, and ZMFF are similar, these images all yielding fully focused image results; the GD method generates a high-brightness fusion image, with the color information in the image being partially lost; the U2Fusion method generates a sharpened fused image, with some areas of the image having lost information, such as the branches of the tree on the right side and the branches of the small tree that is far away in the image, all of which are darker in the image, which is not conducive to obtaining information in a fully focused image; and the fusion image calculated by the proposed method is an all-focused image, with the details and edge information of the image being well preserved and with moderate brightness and ease of observation.

In Figure 7, we can denote that the fusion images calculated by the BGSC, GD, FusionDN, PMGI, and TEMST methods appear blurry, the images generated by BGSC and TEMST especially exhibiting severe distortion; the image computed by the NSCT method shows a basic full-focus result with some artifacts in the image, making it appear to have texture overlays; the U2Fusion method generates a dark image; and the fusion image calculated by the ZMFF method is very close to the result obtained by our algorithm. However, our fusion effect is better, with a clearer image and the almost seamless integration of all the information into the fully focused image.

In Figure 8, we can see that the fused images computed by the BGSC, FusionDN, and PMGI methods appear blurry, especially the forest information in the far distance of the images, making it difficult to observe the details; the images generated by the NSCT, TEMST, and ZMFF methods are similar; the fused image obtained by the GD method has high brightness and clarity; the fused image computed by the U2Fusion method has some dark regions, especially in the mouth area of the horse, with the loss of information being severe; and the fused image achieved by the proposed method has moderate brightness and retains more image information.

In Figure 9, we can see that the fusion images obtained by the BGSC, FusionDN, and PMGI approaches exhibit a certain degree of the blurring phenomenon, which makes it difficult to observe the information in the images, the image generated by the BGSC method especially not achieving full focus and the information on the left side of the image being severely lost; the NSCT, TEMST, and ZMFF methods all obtain a basic fully focused image from which information about the entire scene can be observed; the GD method gives a high-brightness fusion image but can also cause some details in the image to be lost; the fused image achieved by the U2Fusion method has many darker areas, causing severe information loss (for example, the information about the girl’s clothing and headscarf in the image cannot be captured correctly); the fusion image obtained by our algorithm fully preserves the information of the two source images. Additionally, it not only has moderate brightness, but the detailed information of the entire scene can also be fully observed and obtained.

4.2. Quantitative Comparisons

In this section, the quantitative comparisons of different methods are shown in Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6 and Figure 10. In Table 1, we can see that the metrics data QAB/F, QFMI, QG, QE, QP, QY, and QAG computed by the proposed method are the best; the metric data QAPI generated by the GD method is the best, but the corresponding indicator result of our algorithm still ranks third. In Table 2, we can see that the metrics data QAB/F, QFMI, QG, QE, QP, and QY generated by our method are the best. The metric data QAG generated by the U2Fusion method is the best, while our method ranks second; the metric data QAPI generated by the GD method is the best, and our method ranks fourth. In Table 3, we denoted that the metrics data QAB/F, QFMI, QG, QE, QP, QY and QAG computed by our method are the best; the metric data QAPI generated by the FusionDN method is the best, and our method ranks fourth. In Table 4, we can see that the metrics data QAB/F, QFMI, QG, QE, QP, and QY generated by our method are the best; the metrics data QAG and QAPI generated by the GD method are the best. In Table 5, we denoted that the metrics data QAB/F, QFMI, QG, QE, QP, QY, and QAG computed by our method are the best; the metric data QAPI computed by the FusionDN method is the best.

Table 1.

The quantitative assessment of different methods in Figure 5.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.5083 0.8910 0.5104 0.4608 0.4902 0.8714 8.2490 103.9694
NSCT 0.6931 0.8963 0.6934 0.8322 0.7818 0.9205 10.1363 104.1275
GD 0.6843 0.8914 0.6874 0.7312 0.7718 0.8186 9.6352 120.8782
FusionDN 0.5398 0.8872 0.5459 0.7274 0.5769 0.7789 7.8779 110.6457
PMGI 0.2819 0.8835 0.2829 0.3429 0.4029 0.6042 4.3240 105.6984
TEMST 0.6632 0.8924 0.6633 0.7713 0.7702 0.9323 9.8586 104.3450
U2Fusion 0.5919 0.8863 0.5957 0.7646 0.6907 0.6872 10.0790 97.9331
ZMFF 0.6702 0.8898 0.6736 0.8290 0.7763 0.9146 9.8506 101.9007
Proposed 0.7190 0.8968 0.7193 0.8419 0.8245 0.9461 10.3594 105.8613

Table 2.

The quantitative assessment of different methods in Figure 6.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.6042 0.8409 0.6059 0.6289 0.6181 0.9424 21.4182 100.9177
NSCT 0.7292 0.8431 0.7321 0.8345 0.7498 0.9861 24.1701 100.8183
GD 0.7172 0.8397 0.7211 0.8010 0.7403 0.9313 24.0096 118.4325
FusionDN 0.5762 0.8321 0.5800 0.7486 0.6380 0.9205 24.6566 109.2992
PMGI 0.4743 0.8352 0.4806 0.6013 0.5841 0.8205 13.6582 105.7407
TEMST 0.7159 0.8404 0.7193 0.8274 0.7421 0.9815 23.7422 101.4723
U2Fusion 0.5667 0.8320 0.5712 0.7219 0.6186 0.8666 25.5782 93.3662
ZMFF 0.7089 0.8429 0.7120 0.8316 0.7443 0.9772 24.0680 98.6039
Proposed 0.7381 0.8453 0.7412 0.8398 0.7649 0.9893 24.7951 102.7896

Table 3.

The quantitative assessment of different methods in Figure 7.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.3510 0.8138 0.3463 0.2734 0.2438 0.7150 14.5754 82.8519
NSCT 0.7609 0.8528 0.7562 0.8874 0.7451 0.9556 20.2824 84.1720
GD 0.7597 0.8477 0.7543 0.7906 0.7720 0.8668 19.6094 107.0923
FusionDN 0.6414 0.8398 0.6312 0.7718 0.6440 0.8356 17.8845 108.0876
PMGI 0.2369 0.8391 0.2309 0.3342 0.4773 0.6285 8.8447 101.0823
TEMST 0.6563 0.8416 0.6491 0.7598 0.6544 0.9165 18.4142 83.8760
U2Fusion 0.6239 0.8354 0.6157 0.7789 0.6156 0.7982 18.1150 70.1179
ZMFF 0.7585 0.8490 0.7513 0.8963 0.7652 0.9671 19.2529 79.9660
Proposed 0.7909 0.8535 0.7881 0.9041 0.7919 0.9720 20.7733 85.8043

Table 4.

The quantitative assessment of different methods in Figure 8.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.6808 0.9452 0.6766 0.8362 0.6360 0.8873 8.0699 112.4417
NSCT 0.7890 0.9460 0.7850 0.9335 0.8585 0.9412 9.9235 112.0279
GD 0.7360 0.9247 0.7328 0.8463 0.7966 0.8397 11.8692 124.0953
FusionDN 0.6570 0.9302 0.6529 0.8512 0.6466 0.8367 9.5025 110.2910
PMGI 0.4846 0.9326 0.4817 0.5948 0.4954 0.6824 4.6840 107.4757
TEMST 0.7850 0.9446 0.7830 0.9336 0.8464 0.9477 10.1213 112.5434
U2Fusion 0.6308 0.9205 0.6295 0.7828 0.6792 0.7869 11.6540 105.0806
ZMFF 0.7695 0.9390 0.7673 0.9270 0.8293 0.9289 9.7681 111.2188
Proposed 0.8106 0.9463 0.8102 0.9367 0.8903 0.9723 10.3106 112.7676

Table 5.

The quantitative assessment of different methods in Figure 9.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.6100 0.9308 0.6117 0.5300 0.6441 0.9339 4.6552 82.9920
NSCT 0.6804 0.9410 0.6808 0.8308 0.7105 0.9020 5.1003 82.5546
GD 0.6710 0.9301 0.6678 0.7118 0.7076 0.7964 5.3423 104.4747
FusionDN 0.5641 0.9328 0.5611 0.7192 0.4708 0.7252 4.7077 105.7243
PMGI 0.4156 0.9272 0.4147 0.4973 0.2699 0.6139 2.7373 97.1696
TEMST 0.6899 0.9375 0.6902 0.8064 0.7799 0.9433 5.1428 82.9629
U2Fusion 0.5494 0.9316 0.5469 0.7190 0.5362 0.6846 5.3458 72.9192
ZMFF 0.6831 0.9390 0.6807 0.8368 0.7691 0.9100 5.1441 81.7770
Proposed 0.7209 0.9444 0.7210 0.8440 0.8238 0.9587 5.4338 83.2692

Table 6.

The average quantitative assessment of different methods in Figure 10.

QAB/F QMFI QG QE QP QY QAG QAPI
BGSC 0.5242 0.8826 0.5210 0.5173 0.4742 0.8512 9.3365 111.7289
NSCT 0.7103 0.8972 0.7058 0.8644 0.7616 0.9208 11.3477 111.8662
GD 0.7034 0.8887 0.6987 0.7874 0.7466 0.8608 11.6844 126.7408
FusionDN 0.6018 0.8833 0.5952 0.7663 0.6221 0.8224 10.7865 110.9212
PMGI 0.3901 0.8815 0.3857 0.4736 0.4620 0.6738 5.8684 108.0326
TEMST 0.6722 0.8915 0.6664 0.8008 0.7263 0.9153 10.8599 112.2319
U2Fusion 0.6143 0.8844 0.6093 0.7835 0.6657 0.7912 12.0343 108.1656
ZMFF 0.7087 0.8925 0.7030 0.8687 0.7853 0.9313 11.2183 109.8769
Proposed 0.7388 0.8986 0.7366 0.8745 0.8100 0.9530 11.7371 113.3531

Figure 10.

Figure 10

Figure 10

The line chart of metrics data.

Figure 10 shows the objective experimental results of different algorithms with respect to 20 sets of image pairs, and the data results of the same method with respect to different image pairs are connected into a curve, with the average indicator value on the right side of the indicator graph. Table 6 shows the average quantitative assessment of the different methods in Figure 10. In Figure 10 and Table 6, we can see that the average metrics data QAB/F, QFMI, QG, QE, QP, and QY generated by our method are the best; the average metrics date QAG and QAPI generated by the U2Fusion and GD methods are the best, respectively, and the two corresponding average metrics data computed by our method rank second.

Through rigorous qualitative and quantitative evaluation and analysis, the results show that our algorithm stands out in the fusion effect of multi-focus images. Compared with state-of-the-art algorithms, we have achieved the best fusion effect and have the advantages of rich information and clear images.

5. Conclusions

In this paper, a novel multi-focus image fusion method based on the distance-weighted regional energy and structure tensor in NSCT domain was proposed. The structure tensor-based fusion rule was utilized to fuse low-frequency sub-bands, and the distance-weighted regional energy-based fusion rule was utilized to fused high-frequency sub-bands. The proposed method was experimented on the Lytro dataset with 20 paired images, and the fusion results demonstrate that our method can generate state-of-the-art fusion performance in terms of image information, definition, and brightness, making the seamless fusion of multi-focus images possible. In future work, we will improve and expand the application of this algorithm in the field of multi-modal image fusion so that it has better universality in image fusion.

Author Contributions

The experimental measurements and data collection were carried out by M.L., L.L., Z.J. and H.M. The manuscript was written by M.L. with the assistance of L.L., Q.J., Z.J., L.C. and H.M. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This work was supported by the National Science Foundation of China under Grant No. 62261053; the Shanghai Aerospace Science and Technology Innovation Fund under Grant No. SAST2019-048; and the Cross-Media Intelligent Technology Project of the Beijing National Research Center for Information Science and Technology (BNRist) under Grant No. BNR2019TD01022.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Zhang H., Xu H. Image fusion meets deep learning: A survey and perspective. Inf. Fusion. 2021;76:323–336. doi: 10.1016/j.inffus.2021.06.008. [DOI] [Google Scholar]
  • 2.Karim S., Tong G. Current advances and future perspectives of image fusion: A comprehensive review. Inf. Fusion. 2023;90:185–217. doi: 10.1016/j.inffus.2022.09.019. [DOI] [Google Scholar]
  • 3.Hu X., Jiang J., Liu X., Ma J. ZMFF: Zero-shot multi-focus image fusion. Inf. Fusion. 2023;92:127–138. doi: 10.1016/j.inffus.2022.11.014. [DOI] [Google Scholar]
  • 4.Zafar R., Farid M., Khan M. Multi-focus image fusion: Algorithms, evaluation, and a library. J. Imaging. 2020;6:60. doi: 10.3390/jimaging6070060. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Dong Y., Chen Z., Li Z., Gao F. A multi-branch multi-scale deep learning image fusion algorithm based on DenseNet. Appl. Sci. 2022;12:10989. doi: 10.3390/app122110989. [DOI] [Google Scholar]
  • 6.Singh S., Singh H. A review of image fusion: Methods, applications and performance metrics. Digit. Signal Process. 2023;137:104020. doi: 10.1016/j.dsp.2023.104020. [DOI] [Google Scholar]
  • 7.Li L., Ma H., Jia Z., Si Y. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021;80:12389–12409. doi: 10.1007/s11042-020-10462-y. [DOI] [Google Scholar]
  • 8.Li L., Si Y., Wang L., Jia Z., Ma H. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020;79:24303–24328. doi: 10.1007/s11042-020-09154-4. [DOI] [Google Scholar]
  • 9.Wu P., Hua Z., Li J. Multi-scale siamese networks for multi-focus image fusion. Multimed. Tools Appl. 2023;82:15651–15672. doi: 10.1007/s11042-022-13949-y. [DOI] [Google Scholar]
  • 10.Candes E., Demanet L. Fast discrete curvelet transforms. Multiscale Model. Simul. 2006;5:861–899. doi: 10.1137/05064182X. [DOI] [Google Scholar]
  • 11.Do M.N., Vetterli M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005;14:2091–2106. doi: 10.1109/TIP.2005.859376. [DOI] [PubMed] [Google Scholar]
  • 12.Da A., Zhou J., Do M. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006;15:3089–3101. doi: 10.1109/tip.2006.877507. [DOI] [PubMed] [Google Scholar]
  • 13.Guo K., Labate D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007;39:298–318. doi: 10.1137/060649781. [DOI] [Google Scholar]
  • 14.Easley G., Labate D., Lim W.Q. Sparse directional image representations using the discrete shearlet transform. Appl. Comput. Harmon. Anal. 2008;25:25–46. doi: 10.1016/j.acha.2007.09.003. [DOI] [Google Scholar]
  • 15.Kumar N., Prasad T., Prasad K. An intelligent multimodal medical image fusion model based on improved fast discrete curvelet transform and type-2 fuzzy entropy. Int. J. Fuzzy Syst. 2023;25:96–117. doi: 10.1007/s40815-022-01379-9. [DOI] [Google Scholar]
  • 16.Kumar N., Prasad T., Prasad K. Multimodal medical image fusion with improved multi-objective meta-heuristic algorithm with fuzzy entropy. J. Inf. Knowl. Manag. 2023;22:2250063. doi: 10.1142/S0219649222500630. [DOI] [Google Scholar]
  • 17.Li S., Yang B. Multifocus image fusion by combining curvelet and wavelet transform. Pattern Recognit. Lett. 2008;29:1295–1301. doi: 10.1016/j.patrec.2008.02.002. [DOI] [Google Scholar]
  • 18.Zhang W., Jiao L., Liu F. Adaptive contourlet fusion clustering for SAR image change detection. IEEE Trans. Image Process. 2022;31:2295–2308. doi: 10.1109/TIP.2022.3154922. [DOI] [PubMed] [Google Scholar]
  • 19.Li L., Lv M., Jia Z., Ma H. Sparse representation-based multi-focus image fusion method via local energy in shearlet domain. Sensors. 2023;23:2888. doi: 10.3390/s23062888. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Hao H., Zhang B., Wang K. MGFuse: An infrared and visible image fusion algorithm based on multiscale decomposition optimization and gradient-weighted local energy. IEEE Access. 2023;11:33248–33260. doi: 10.1109/ACCESS.2023.3263183. [DOI] [Google Scholar]
  • 21.Li S., Kang X., Hu J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013;22:2864–2875. doi: 10.1109/TIP.2013.2244222. [DOI] [PubMed] [Google Scholar]
  • 22.Shreyamsha Kumar B. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015;9:1193–1204. doi: 10.1007/s11760-013-0556-9. [DOI] [Google Scholar]
  • 23.Tan W., Zhou H. Fusion of multi-focus images via a Gaussian curvature filter and synthetic focusing degree criterion. Appl. Opt. 2018;57:10092–10101. doi: 10.1364/AO.57.010092. [DOI] [PubMed] [Google Scholar]
  • 24.Feng X., Fang C., Qiu G. Multimodal medical image fusion based on visual saliency map and multichannel dynamic threshold neural P systems in sub-window variance filter domain. Biomed. Signal Process. Control. 2023;84:104794. doi: 10.1016/j.bspc.2023.104794. [DOI] [Google Scholar]
  • 25.Zhang Y., Lee H. Multi-sensor infrared and visible image fusion via double joint edge preservation filter and non-globally saliency gradient operator. IEEE Sens. J. 2023;23:10252–10267. doi: 10.1109/JSEN.2023.3262775. [DOI] [Google Scholar]
  • 26.Jiang Q., Huang J. Medical image fusion using a new entropy measure between intuitionistic fuzzy sets joint Gaussian curvature filter. IEEE Trans. Radiat. Plasma Med. Sci. 2023;7:494–508. doi: 10.1109/TRPMS.2023.3239520. [DOI] [Google Scholar]
  • 27.Zhang C. Multifocus image fusion using a convolutional elastic network. Multimed. Tools Appl. 2022;81:1395–1418. doi: 10.1007/s11042-021-11362-5. [DOI] [Google Scholar]
  • 28.Ma W., Wang K., Li J. Infrared and visible image fusion technology and application: A review. Sensors. 2023;23:599. doi: 10.3390/s23020599. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Liu Y., Chen X., Peng H. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion. 2017;36:191–207. doi: 10.1016/j.inffus.2016.12.001. [DOI] [Google Scholar]
  • 30.Lai R., Li Y., Guan J. Multi-scale visual attention deep convolutional neural network for multi-focus image fusion. IEEE Access. 2019;7:114385–114399. doi: 10.1109/ACCESS.2019.2935006. [DOI] [Google Scholar]
  • 31.Wang L., Liu Y. MSE-Fusion: Weakly supervised medical image fusion with modal synthesis and enhancement. Eng. Appl. Artif. Intell. 2023;119:105744. doi: 10.1016/j.engappai.2022.105744. [DOI] [Google Scholar]
  • 32.Wang Z., Li X. A self-supervised residual feature learning model for multifocus image fusion. IEEE Trans. Image Process. 2022;31:4527–4542. doi: 10.1109/TIP.2022.3184250. [DOI] [PubMed] [Google Scholar]
  • 33.Jiang L., Fan H., Li J. A multi-focus image fusion method based on attention mechanism and supervised learning. Appl. Intell. 2022;52:339–357. doi: 10.1007/s10489-021-02358-7. [DOI] [Google Scholar]
  • 34.Jin X., Xi X., Zhou D. An unsupervised multi-focus image fusion method based on Transformer and U-Net. IET Image Process. 2023;17:733–746. doi: 10.1049/ipr2.12668. [DOI] [Google Scholar]
  • 35.Zhang H., Le Z., Shao Z. MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion. 2021;66:40–53. doi: 10.1016/j.inffus.2020.08.022. [DOI] [Google Scholar]
  • 36.Liu S., Yang L. BPDGAN: A GAN-based unsupervised back project dense network for multi-modal medical image fusion. Entropy. 2022;24:1823. doi: 10.3390/e24121823. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Li X., Zhou F., Tan H. Multi-focus image fusion based on nonsubsampled contourlet transform and residual removal. Signal Process. 2021;184:108062. doi: 10.1016/j.sigpro.2021.108062. [DOI] [Google Scholar]
  • 38.Panigrahy C., Seal A. Parameter adaptive unit-linking pulse coupled neural network based MRI-PET/SPECT image fusion. Biomed. Signal Process. Control. 2023;83:104659. doi: 10.1016/j.bspc.2023.104659. [DOI] [Google Scholar]
  • 39.Tian J., Chen L., Ma L. Multi-focus image fusion using a bilateral gradient-based sharpness criterion. Opt. Commun. 2011;284:80–87. doi: 10.1016/j.optcom.2010.08.085. [DOI] [Google Scholar]
  • 40.Das S., Kundu M.K. A neuro-fuzzy approach for medical image fusion. IEEE Trans. Biomed. Eng. 2013;60:3347–3353. doi: 10.1109/TBME.2013.2282461. [DOI] [PubMed] [Google Scholar]
  • 41.Paul S., Sevcenco I., Agathoklis P. Multi-exposure and multi-focus image fusion in gradient domain. J. Circuits Syst. Comput. 2016;25:1650123. doi: 10.1142/S0218126616501231. [DOI] [Google Scholar]
  • 42.Xu H., Ma J., Le Z. FusionDN: A unified densely connected network for image fusion; Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI); New York, NY, USA. 7–12 February 2020; pp. 12484–12491. [Google Scholar]
  • 43.Zhang H., Xu H., Xiao Y. Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity; Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI); New York, NY, USA. 7–12 February 2020; pp. 12797–12804. [Google Scholar]
  • 44.Chen J., Li X., Luo L. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 2020;508:64–78. doi: 10.1016/j.ins.2019.08.066. [DOI] [Google Scholar]
  • 45.Xu H., Ma J., Jiang J. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2022;44:502–518. doi: 10.1109/TPAMI.2020.3012548. [DOI] [PubMed] [Google Scholar]
  • 46.Qu X., Yan J., Xiao H. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Autom. Sin. 2008;34:1508–1514. doi: 10.3724/SP.J.1004.2008.01508. [DOI] [Google Scholar]
  • 47.Haghighat M., Razian M. Fast-FMI: Non-reference image fusion metric; Proceedings of the IEEE 8th International Conference on Application of Information and Communication Technologies; Astana, Kazakhstan. 15–17 October 2014; pp. 424–426. [Google Scholar]
  • 48.Liu Z., Blasch E., Xue Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012;34:94–109. doi: 10.1109/TPAMI.2011.109. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES