Skip to main content
Springer logoLink to Springer
. 2022 Nov 7;61(1):155–177. doi: 10.1007/s11517-022-02697-8

Multimodal medical image fusion algorithm based on pulse coupled neural networks and nonsubsampled contourlet transform

SaI Ibrahim 1,, M A Makhlouf 1, GhS El-Tawel 2
PMCID: PMC9816265  PMID: 36342598

Abstract

Abstract

Combining two medical images from different modalities is more helpful for using the resulting image in the healthcare field. Medical image fusion means combining two or more images coming from multiple sensors. This technology obtains an output image that presents more effective and useful information from two images. This paper proposes a multi-modal medical image fusion algorithm based on the nonsubsampled contourlet transform (NSCT) and pulse coupled neural networks (PCNN) methods. The input images are decomposed using the NSCT method into low- and high-frequency subbands. The PCNN is a fusion rule for integrating both low- and high-frequency subbands. The inverse of the NSCT method is to reconstruct the fused image. The results of medical image fusion help doctors with disease diagnosis and patient treatment. The proposed algorithm is tested on six groups of multi-modal medical images using 100 pairs of input images. The proposed algorithm is compared with eight fusion methods. We evaluate the performance of the proposed algorithm using the fusion metrics: peak signal to noise ratio (PSNR), mutual information (MI), entropy (EN), weighted edge information (QAB/F), nonlinear correlation information entropy (Qncie), standard deviation (SD), and average gradient (AG). Experimental results show that the proposed algorithm can perform better than other medical image fusion methods and achieve promising results.

Graphical abstract

graphic file with name 11517_2022_2697_Figd_HTML.jpg

Keywords: Medical image fusion, Pulse coupled neural networks, Nonsubsampled contourlet transform, Computed tomography, Magnetic resonance image

Introduction

Medical images play an essential role in healthcare applications like disease diagnosis and patient treatment [9]. These images are capturing from different modalities such as magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), and single-photon emission computed tomography (SPECT).

All of these images are spot on different organ information. The CT images are used to visualize bone structure, and the MR images are used to visualize the internal or soft structures of the organ where the CT image is more accurate than the MRI image. On the other hand, PET and SPECT images provide metabolic or functional information in low resolution for the organ and are more accurate in tumor detection [12, 14, 25]. Table 1 describes the advantages and disadvantages of multimodality medical image.

Table 1.

Multimodal medical image examples [9, 12, 25]

Modal Example Advantage Disadvantage
CT graphic file with name 11517_2022_2697_Fige_HTML.gif Scan in low time, less distortion, higher resolution, poor contrast for soft tissue
and more accurate than MRI
MRI graphic file with name 11517_2022_2697_Figf_HTML.gif high resolution in the spatial domain, more safe not accurate, difficult for dealing with
for pregnant women and preview anatomical details movement organs like mouth tumors,
and low sensitivity
PET graphic file with name 11517_2022_2697_Figg_HTML.gif high sensitivity and accuracy in tumor detection low resolution and high cost
SPECT graphic file with name 11517_2022_2697_Figh_HTML.gif accurate in tumor detection and high sensitivity low resolution, low image quality,
high cost, and blur effects

There are three categories of image fusion like pixel-level fusion methods, feature-level fusion methods, and decision-level fusion methods [22]. Pixel-level fusion seeks to obtain the fused image by integrating the pixel information of input images. Feature-level fusion extracted the meaningful features from the input images and merged them in a single vector [6]. Pixel-level fusion is performed in either the spatial or transform domain. It is widely used in medical image fusion.

The spatial domain image fusion techniques focused on the input image pixels. The main advantage of this domain is low computational time. On the other hand, it introduces spatial distortion and produces color distortion and low contrast images [20]. The common examples of spatial domain-based image fusion methods are the principal component analysis (PCA) method, average fusion method, weighted average fusion method, minimum fusion method, and maximum fusion method.

Transform domain image fusion techniques aim to get low- and high-frequency coefficients by transforming the input images into the frequency domain rather than a spatial domain. It is more accurate and efficient than spatial domain methods. The advantages of the transform domain method are avoiding distortion and dealing with multiple resolution images (Fig. 1).

Fig. 1.

Fig. 1

Image fusion process

The common medical image fusion methods in the transform domain are based on multiscale transform (MST) to obtain a good result. The MST fusion methods performed in three steps are decomposition, fusion, and reconstruction [14, 30]. The common MST methods are Laplacian pyramid (LP) [2, 4], discrete wavelet transform (DWT) [21], nonsubsampled shearlet transform (NSST) [23], convolutional neural networks (CNN) [14], and NSCT [31]. The basic image fusion process is described in these steps:

  • Image decomposition: convert the source images into an MST domain.

  • Fusion rule: apply the fusion rule to merge the transformed coefficients.

  • Image reconstruction: apply the inverse transform to reconstruct the fused image.

Motivations

Medical images are accurate images that require massive effort to clean and prepare for usage. These images face two challenges. To begin, collect medical images in high resolution. Second, create a good image fusion algorithm that preserves all the salient features in the source images.

The main motivations for this paper are choosing the most effective method for combining several source images with the following characteristics: high efficiency, high spatial resolution preservation, and low color distortion using the PCNN in the NSCT domain to aid doctors in accurately diagnosing diseases. It also creates a new accurate fused image with more detailed information than the input images.

Contribution

Our proposed medical image fusion method uses the NSCT features, including multi-scale, shift-invariance, and multi-directional properties, along with the PCNN to gain high fusion performance and capture the subtle differences and fine details present in the source medical images. The proposed method enhances the output fused image’s high contrast, clarity, and information content.

The main contribution of this paper is to create a high-performance fusion algorithm to detect whole brain regions from different multimodality medical images.

In summary, we propose a fused algorithm based on the PCNN method for multimodality medical images in the NSCT domain to improve the fused image quality to aid doctors in disease diagnosis. The rest of the paper is organized as follows. Section 2 focuses on some previous works. In Section 3 presents the proposed algorithm used in this paper. The experimental results and performance evaluation are discussed in Section 5. Finally, We conclude and summarize whole the paper in Section 6.

Related work

Researchers presented multiple medical image fusion methods. All of these methods are tested and achieved good results. In this section, we preview and analyze some of this research.

This paper designs an effective CT and MR image fusion method [6]. In this work, the NSCT decomposes the source images. A maximum entropy of the square of the coefficients within a local window merged the low-frequency sub-bands. Maximum-weighted sum-modified Laplacian merged the high-frequency sub-bands. Finally, the inverse NSCT creates the fused image. We evaluate the proposed method using the CT and MR images for different cases and then compare the results with the other conventional image fusion methods. Both visual analysis and quantitative evaluation of experimental results show the superiority of the proposed algorithm over other methods.

Nazrudeen et al. [19] proposed a medical image fusion method based on NSCT. In this paper, the fusion process can be stated as follows: apply input image decomposition using the NSCT domain into low and high-frequency subbands. Apply phase congruency and directive contrast methods as a fusion rule. To produce the fused image, use the inverse NSCT method. The proposed method tested on Alzheimer, stroke, and tumor data, using CT and MRI datasets as input images. Whole experiments are applied in the MATLAB toolbox. Results are evaluated using PSNR (peak signal to noise ratio) and RMSE (root mean square error) measures. The proposed method is compared with classical fusion methods and produces higher image performance than other compared methods.

Manker et al. [18] proposed the NSCT fusion method and pixel-level fusion to fuse multimodal medical images. In this paper, use CT and MRI as input images. The input images are decomposed by NSCT transformation. The Gabor filter bank is applied on low-frequency coefficients and used the gradient fusion method on high-frequency coefficients. The inverse of NSCT transformation is applied to the resulting image to obtain the fused image. The results were evaluated by using common metrics such as entropy, PSNR, correlation coefficient, and MSE (mean square error).

Gomathi et al. [7] presented the NSCT method to fuse medical images. In this paper, the input images are decomposed into low-frequency and high-frequency coefficients by using the NSCT method. The maximum local mean and the maximum local variance are two fusion rules used. The maximum local mean method is applied on low-frequency coefficients and the maximum local variance method for high-frequency coefficients. The inverse of the NSCT method is to reconstruct the fused image. The presented method is tested on CT, MRI, and PET images using MATLAB R2010a. The common quality metrics such as entropy, standard deviation, mean, and edge-based similarity measure (QAB/F) results declare that the applied method is better than compared methods.

Tain et al. [24] presented an improved PCNN (IPCNN) based on the NSCT domain. In this paper, apply the NSCT method to decompose input images into subbands. Next, apply the IPCNN method as a fusion rule into the merged low and high subbands. Finally, perform the inverse NSCT to get the fused image. The results were evaluated by using common metrics such as entropy, mutual information, and weighted edge information. The experiment results show that the proposed method is better than other compared methods to fused medical images.

Xia et al. [28] presented a combination of sparse representation, NSCT transform, and PCNN method to fuse medical images. This combination aims to solve the NSCT problem in a low subband coefficient that is not sparse. The proposed fusion strategy is performed in three steps. First, decompose the input image using NSCT transform. Second, use the sparse representation and PCNN algorithm as the fusion rules respectively on low subbands and high subbands. Finally, use the NSCT inverse to produce the fused image. The result was evaluated by seven metrics such as standard deviation (SD), information entropy (IE), average gradient (AG), spatial frequency (SF), mutual information (MI), and edge information delivery factor, and structural similarity model (SSIM). The result shows the fused image with higher performance and better contrast than other compared methods.

Zhu et al. [32] proposed a new multimodal medical image fusion strategy based on NSCT transform and also used phase congruency and local Laplacian energy algorithms. The procedure of the proposed method is performed in three main steps. First, apply the NSCT method to decompose the input images into both lowpass and highpass subbands. Then, apply the local Laplacian energy fusion rule on the lowpass subbands and use the phase congruency fusion rule on the highpass subbands. Finally, apply the inverse NSCT transformation on the merged result from both lowpass and highpass subbands to produce the final fused image. The experiment results show that the performed method obtained high-performance fusion result with low computational time. The main defect of this method is not good to fused PET-MRI images.

Material and methods

Non subsampled contourlet transform (NSCT)

The contourlet transform (CT) method is used in image processing especially in geometric transformations and produces good results in this field [7]. The main problem of the CT method is a shift variant caused by down- and upsampling [32]. The NSCT method is a shift-invariant, multi-directional transform, and multi-scale image representation that depends on the CT theory and is applied by a` trous algorithm.

This method is achieved by applying two basic stages: the nonsubsampled pyramid filter bank (NSP or NSPFB) and the nonsubsampled directional filter bank (NSDFB) [18, 19, 32]. The multiscale and multi-directional transform is ensured by both NSPFB and NSDFB filters. The image decomposition steps using the NSCT method are described as in Fig. 2.

Fig. 2.

Fig. 2

The NSCT image decomposition process [7]

The main steps of basic NSCT transform in medical image fusion are stated as in the following Algorithm 1

graphic file with name 11517_2022_2697_Figi_HTML.jpg

Algorithm 1 The basic NSCT for medical image fusion algorithm.

Nonsubsampled pyramid filter bank (NSPFB)

The NSPFB consists of a two-channel filter bank without downsamplers and upsamplers [6, 32]. This filter bank aims to achieve multiscale decomposition for input images into the low-pass and high-pass subbands. Each NSPFB decomposition level aims to obtain both low- and high-pass frequency images. Then, the low-frequency image is decomposed iteratively by NSPFB. The result is M+1 sub-images, where M represents high-frequency images, and 1 represents the low-frequency image [7, 32].

Nonsubsampled directional filter bank (NSDFB)

NSDFB is a nonsubsampled filter bank consisting of two channels that are obtained by merging the directional fan filter-banks [7]. This filter bank aims to decompose the high-frequency images resulted from NSP decomposition to result at the directional sub-images, where the size of the source image and directional sub-images are the same. The NSDFB ensures the NSCT produces accurate directional detail information and multi-directional feature [7, 32].

Pulse coupled neural networks (PCNN)

PCNN is the third generation of biological artificial neural network method that is used in many areas such as image processing, object detection, and image fusion. It aims to stimulate and utilize the synchronous pulse emission from the visual cortex for some mammals such as the cat’s brain established in 1990 by Eckhorn et al. [5, 28]. The main benefit of the PCNN method is applied image fusion without a training process [8]. The PCNN is represented as a one-layer network that involves multiple neurons connecting. The following Fig. 3 describes the main PCNN structure. This structure consists of three parts: a dendritic tree, linking modulation, and a pulse generator.

Fig. 3.

Fig. 3

Architecture of the PCNN model[27]

The inputs from the receptive fields are received from the dendritic tree. There are two types of receptive fields. The receptive field types consist of two branches named the linking and the feeding [26]. The role of linking is to receive an external stimulus; on the other hand, the feeding receives both local and external stimulus. The PCNN model can be described mathematically by the following equations [29]:

fij(n)=e-αffij(n-1)+vfklmijklykl(n-1)+sij 1
lij(n)=e-αlfij(n-1)+vlklwijklykl(n-1) 2
uij(n)=fij(n)(1+βlij(n)) 3
yij(n)=1,uij(n)>hij(n-1)0,otherwise 4
hij(n)=e-αhhij(n-1)+vhyij(n-1) 5

where the input channels in the PCNN model are represented by fij(feeding channel) and lij(linking channel) in the i and j position. The external stimulus is defined by sij, and mijkl and wijkl are considered as the local matrix. The neuron’s output is defined by yij. The αf, αl, and αh are represented as the time constants. The linking coefficient is β. The voltage is represented by vf, vl, and vh.

Proposed algorithm

In this paper, a multi-modality medical image fusion algorithm is proposed. The proposed algorithm is divided into three basic steps, namely image decomposition, fusion rule, and image reconstruction as shown in Fig. 4.

Fig. 4.

Fig. 4

The schematic diagram of the proposed fusion algorithm

Image decomposition

Image decomposition is considered the first step in the proposed algorithm. In this step, we use the NSCT method to decompose preprocessed images A and B into low- and high-frequency subbands LA, HA, LB, and HB. The LA is a low-frequency subband for image A, and the high-frequency subband is HA. The LBandHB have the same meaning as image A.

Fusion rule

Fusion of low- and high-frequency subbands applying the PCNN method as in Eqs. 1 to 4 and calculating the firing time as in Eq. 5, the fused low-and high-frequency coefficients LF and HF are calculated using the following equations:

LF(i,j)=LA(i,j),ifhA,ij[N]>hB,ij[N]LB(i,j),otherwise 6
HF(i,j)=HA(i,j),ifhA,ij[N]>hB,ij[N]HB(i,j),otherwise 7

where N represents the total number of iterations.

Image reconstruction

In the NSCT reconstruction step, we use the inverse of the NSCT transform to combine the fused low- and high-frequency coefficients LF and HF to produce the fused image F. Algorithm 2 discusses the steps of the proposed fusion method for multi-modality medical source images.

graphic file with name 11517_2022_2697_Figj_HTML.jpg

Algorithm 2 The proposed multi-modality medical image fusion algorithm.

Experiment results and discussion

In this section, we discuss the details about the results that are used in this paper. This section is divided into four subsections: Datasets, Quality measures, Performance evaluation, and Comparing with other techniques.

Datasets

In our experiments, these source images are collected from the Whole Brain Atlas database [11]. This database includes both CT and MRI images. We evaluate the proposed algorithm performance by using three pairs of multi-modal medical images. We use 100 pairs of multimodality medical images, 25 image pairs for CT-MRI fusion, 25 image pairs for T1-T2 weighted MRI fusion, and 25 image pairs for CT, MR-PD, and MR-Gad images for normal or abnormal brain diseases. We also use the 25 image pairs for MR-T2, SPECT, and PET images.

All of these images are accurately registered and have the same size of 256*256 pixels. We also use the Matlab R20l8a toolbox to obtain the results. Our experiments are tested on the device with Windows 10, one TB hard disk, 8 GB memory, and an Intel Core i7 processor. Samples of datasets are used in this experiment shown in Table 2.

Table 2.

Some parameters of PCNN

Parameter Link_arrange β αL VL αθ Vθ No. of iteration
Values 3 3 1 1 .8 20 100

The bold values represent the PCNN parameter used in our experiment

Quality measures

In this subsection, we present some evaluation metrics for medical image fusion. There are common evaluation metrics for image fusion. Our experiments use these fusion metrics to evaluate the performance of the proposed algorithm. They are entropy (EN), mutual information (MI), QAB/F, nonlinear correlation information entropy (Qncie), peak signal to noise ratio (PSNR), standard deviation (SD), and average gradient (AG). All of these metrics are discussed as follows.

  • Entropy (EN): It is useful for measuring the amount of information in the fused image. High EN value means the fused images with high quality and high performance. It is defined as follows:
    EN=-l=0L-1pllog2pl 8
    where pl is the ratio of pixels with the gray levels of l and L represents a total number of gray levels of an image [10, 24].
  • Mutual information (MI): this metric is used to evaluate the whole information in the fused image. It also measures the relevance or the dependence degree between two or more images [1, 10, 24]. MI is given by:
    MI=MIAF+MIBF 9
    where A and B represent the source images and the fused image is represented by F. The high MI value means the high-performance fused image. MIAF represents the mutual information between both the source image A and the fused image F. pA,F(m,n) represent the joint probability of the source and the fused image.
    MIAF=m,npA,F(m,n)log2pA,F(m,n)pA(m)pF(n) 10
  • Weighted edge information (QAB/F): total information transferred and edge intensity information from source images to the fused image, which is given as [1, 24]:
    QAB/F=m=1Mn=1NQAF(m,n)WA(m,n)+QBF(m,n)WB(m,n)m=1Mn=1N(WA(m,n)+WB(m,n)) 11
    where the preservation factors of the edge information are denoted by QAF and QBF, and the weighted items represented by both WA and WB. The QAB/F range is between 0 and 1.
  • Peak signal to noise ratio (PSNR): one of the main evaluation metrics to measure the quality of the fused image. The high PSNR values represent high-quality images [10] and is given by this equation:
    PSNR=10log102552/MSE 12
    where mean squared error (MSE) is the squared difference between the original image x(lk) and the output image x¯(l,k) and given by the following equation:
MSE=1MNl=1Mk=1Nx(l,k)-x¯(l,k)2 13
  • Standarad deviation (SD): It evaluates the contrast of the fused image by spreading the image data. The high SD value means the fused image with high visibility and good quality image [1, 10]. It is represent by the following equation:
    SD=m=1Mn=1NF(m,n)-μ2MN 14
    where MN represent the size of input image F(mn) and μ represent the average of pixel intensity value of the fused image. The μ is defined as follows:
μ=m=1Mn=1NF(m,n)MN 15
  • Average gradient (AG): the gradient Information of the combined image is evaluated by this metric. It also measures the texture detail such as sharpness and clarity of the fused image [1, 10]. High AG value means the fused image with high performance. The AG metric is given by this equation

AG=m=1Mn=1N(F(m,n)-F(m+1,n))2+(F(m,n)-F(m,n+1))2/2MN 16
  • The nonlinear correlation information entropy Qncie: measures the nonlinear information of the fused image. Qncie is denoted by the following formula [3]:
    Qncie(X,Y)=2+i=0b2niNlogbniN 17
    where N refers to the dataset size and ni refers to the number of samples.

Performance evaluation

In this subsection, we list some fusion methods used in multimodal image fusion in the medical area. The performance of the proposed algorithm is better if all of these metrics have higher values. We compared the proposed algorithm with eight fusion methods: the discrete wavelet transform (DWT) [13], the multi-channel model–pulse coupled neural networks (MPCNN) [26], the convolutional sparse representation (CSR) [15], the guided image filter and statistics (GFS) [1], the NSCT [13], the convolutional sparsity-based morphological component analysis (CSMCA) [16], the nonsubsampled contourlet transform–sparse representation (NSCT-SR) [17], and the nonsubsampled contourlet transform–phase congruency local Laplacian (NSCT-PCLP) [32].

The parameters in the proposed method are the following: In NSCT, the decomposition level is set 4; “pyrexc” and “vk” are selected. In PCNN, there are too parameters like β,αL,VL,αθ,Vθ Link_arrange, and number of iterations. The following table describes these parameters (Table 2).

Comparing with other techniques

In our experiments, we apply the proposed algorithm on gray images of four pairs of multi-modal medical images including the following: MR-T1 and MR-T2 images, CT and MR-Gad images, CT and MR-PD images, and CT and MR-T2 images. The following Figs. 5, 6, 7, 8, 9, and 10 show the experiments and results of the proposed algorithm.

Fig. 5.

Fig. 5

MR-T1/MR-T2 image fusion results

Fig. 6.

Fig. 6

Fusion result of (a) MR-T1 and (b) MR-T2 using (c) DWT method, (d) m-PCNN method, (e) CSR method, (f) NSCT method, (g) CSMCA method, (h) GFS method, (i) NSCT-SR method, (j) NSCT-PCLP method, and the proposed method (k) NSCT-PCNN

Fig. 7.

Fig. 7

Fusion result of (a) CT and (b) MR-Gad using (c) DWT method, (d) m-PCNN method, (e) CSR method, (f) NSCT method, (g) CSMCA method, (h) GFS method, (i) NSCT-SR method, (j) NSCT-PCLP method, and the proposed method (k) NSCT-PCNN

Fig. 8.

Fig. 8

Fusion result of (a) CT and (b) MR-PD using (c) DWT method, (d) m-PCNN method, (e) CSR method, (f) NSCT method, (g) CSMCA method, (h) GFS method, (i) NSCT-SR method, (j) NSCT-PCLP method, and the proposed method (k) NSCT-PCNN

Fig. 9.

Fig. 9

Fusion result of (a) CT and (b) MR-T2 using (c) DWT method, (d) m-PCNN method, (e) CSR method, (f) NSCT method, (g) CSMCA method, (h) GFS method, (i) NSCT-SR method, (j) NSCT-PCLP method, and the proposed method (k) NSCT-PCNN

Fig. 10.

Fig. 10

Fusion result of (a) MR-T2 and (b) SPECT using (c) DWT method, (d) NSCT method, (e) NSCT-SR method, (f) NSCT-PCLP method, and the proposed method (g) NSCT-PCNN

Figure 5a is an MR-T1 image and Fig. 5b is an MR-T2 image. In this figure, the fused images of DWT, MPCNN, CSR, NSCT, CSMCA, GFS, NSCT-SR, and NSCT-PCLP are displayed in Fig. 5c, d, e, f, g, h, i, j respectively. The image Fig. 5k represents the fused result of the proposed algorithm. The results show that the DWT and MPCNN methods lose some detailed information from the input image in MR-T2 modality and low contrast images as shown in Fig. 5c and d.

The fused images using the CSR method, the NSCT method, and the CMSCA method represented in Fig. 5e, f, and g are better than Fig. 5c, and d but some detailed information was not detected accurately. In Fig. 5h represented the fused image using the GFS method is good for detecting all image information in Fig. 5a but loses more information from the image in MR-T2 modality. Figure 5i represents the NSCT-SR fused image is detecting more edges and gradient information than Fig. 5j. Figure 5k is the proposed algorithm result with high contrast that preserves both MR-T1 and MR-T2 modality information and prevents visual artifacts.

In Fig. 6a, it is a CT image and Fig. 6b is an MR-Gad image. The results show that Fig. 6c, d, and e lose some detailed information from the input images and produce low contrast images. The results of using the CSMCA method and the GFS method as in Fig. 6g and h visually look good than using the DWT method, MPCNN method, and the CSR method as in Fig. 6c, d, e respectively but do not detect all edges in MR-Gad image. The result of the NSCT-SR method in Fig. 6i is better to fuse CT and MR-Gad images than using the NSCT method and the NSCT-PCLP method. Figure 6k is the fused image of the proposed algorithm with high performance and high contrast, and preserves both CT and MR-Gad modality information without preview visual artifacts.

In Fig. 7a, it is a CT image and Fig. 7b is an MR-PD image. The fused image of the proposed algorithm in Fig. 7k is a high-performance image that contains more mutual information from the input images than using the NSCT-SR method and the NSCT-PCLP method as shown in Fig. 7i and j. In Fig. 8a, it is a CT image and Fig. 8b is an MR-T2 image. The results show that the proposed algorithm in Fig. 8k accurately fused the CT and MR-T2 images and produced high contrast images without preview visual artifacts.

Figure 9 shows the fusion results for MR-T2 and SPECT images. Figure 9a is an MR-T2 image, and Fig. 9b is a SPECT image. The fusion results from the DWT, NSCT, NSCT-SR, and NSCT-PCLP methods perform well in extraction details from MR-T2 images but still have color distortion problems as well as the brain edges cannot detect successfully in Fig. 9c, d, e, and f. The proposed method can preserve color information and achieve higher quality than other methods; see Fig. 9g. Figure 9g shows that the proposed method performs better than NSCT-PCLP as in Fig. 9f on extraction details in some regions.

Figure 10 shows the fusion results for MR-T1 and PET images. Figure 10a is an MR-T1 image, and Fig. 10b is a PET image. The fusion results from the DWT, NSCT, and NSCT-SR can preserve the detailed MR-T1 information with the color fidelity problem in Fig. 10c, d, and e. Figure 10f is better than Fig. 10e in the color fidelity issue but loses some details from the MR-T1 image. The NSCT-PCLP can preserve functional information from the PET image, but some edge and structure information cannot be detected accurately; see Fig. 10f. In Fig. 10g, the proposed method can preserve color and structure information from the source images and achieve higher quality images than other methods.

Tables 3, 4, 5, 6, 7, and 8 report the performance evaluation results of the proposed algorithm and the compared methods. The performance evaluation metrics are calculated, and the highest values at each row shown in bold text are the best score values over all the different used methods. It shows that the proposed NSCT-PCNN algorithm effectively fused medical images and produced high-performance images as compared with other methods. The following figures displayed the values of fusion metrics applied to six pairs of multi-modal medical images, including the following: MR-T1 and MR-T2 images, CT and MR-Gad images, CT and MR-PD images, CT and MR-T2 images, MR-T2 and SPECT images, and MR-T1 and PET images.

Table 3.

Assessment of different fusion methods on MR-T1/MR-T2 images

Method and metrics DWT MPCNN CSR NSCT CSMCA GFS NSCT_SR NSCT_PCLP Proposed
EN 4.5754 4.3219 4.4240 4.7330 4.3413 4.6281 4.7036 4.8471 4.9101
MI 3.3859 3.0288 3.1392 3.1458 3.2119 3.4642 3.2056 3.1677 3.5121
QAB/F 0.5472 0.2319 0.5830 0.5817 0.5845 0.5644 0.5820 0.5433 0.5861
PSNR 64.4770 61.9325 65.4384 65.1782 64.5221 64.8578 64.6402 64.4505 66.1580
SD 80.4108 52.8049 82.4334 85.0782 81.9448 82.5296 85.4476 88.0566 88.2779
AG 8.80906 5.40107 12.0605 12.1523 12.1540 11.3904 12.3337 12.0567 12.5236
Time (sec) 3.0480 33.1315 59.9784 3.1754 1381.08 3.6158 33.6265 3.2978 90.4381

The bold values indicate the best result of the evaluation criterion

Table 4.

Assessment of different fusion methods on CT/MR-Gad images

Method and metrics DWT MPCNN CSR NSCT CSMCA GFS NSCT_SR NSCT_PCLP Proposed
EN 4.6619 4.0622 4.5380 4.8419 4.5102 4.6686 4.7651 4.8893 4.9821
MI 2.9486 2.6828 2.9653 3.1088 3.0709 3.0124 3.1098 3.0928 3.1953
QAB/F 0.3715 0.1279 0.4317 0.4152 0.4443 0.4497 0.4291 0.4080 0.4736
PSNR 64.4828 60.0427 64.8071 64.6118 64.8093 64.1848 64.7035 64.6530 64.9061
SD 69.8985 30.6469 64.5089 76.7335 70.8491 66.0041 78.0360 79.3871 84.4695
AG 8.0279 3.2498 6.6523 7.4735 6.6833 6.7030 7.4143 7.3386 9.6266
Time (sec) 0.4067 34.7309 55.5042 3.0921 1755.09 0.3146 32.9427 2.6856 101.439

The bold values indicate the best result of the evaluation criterion

Table 5.

Assessment of different fusion methods on CT/MR-PD images

Method and metrics DWT MPCNN CSR NSCT CSMCA GFS NSCT_SR NSCT_PCLP Proposed
EN 5.1882 4.5606 5.0707 5.2326 5.2264 5.0675 5.2382 5.3630 5.3986
MI 3.0102 2.8981 3.1521 3.2376 3.1540 3.1720 3.1314 2.9770 3.2543
QAB/F 0.3285 0.1949 0.4450 0.3787 0.4033 0.4591 0.3969 0.3244 0.4704
PSNR 62.8892 59.1924 62.9195 63.4793 63.2778 61.9589 63.4066 62.4729 63.9196
SD 72.5167 36.8359 66.2284 75.6873 70.2274 63.8462 76.9269 81.5905 85.0497
AG 8.5255 4.2299 7.3163 7.3904 6.6782 8.8754 7.3899 8.4993 9.4748
Time (sec) 0.3183 29.6324 51.4395 1.2437 327.855 0.2523 11.4953 2.5246 35.5117

The bold values indicate the best result of the evaluation criterion

Table 6.

Assessment of different fusion methods on CT/MR-T2 images

Method and Metrics DWT MPCNN CSR NSCT CSMCA GFS NSCT_SR NSCT_PCLP Proposed
EN 4.6183 4.1225 4.6325 4.8642 4.6938 4.6336 4.8684 4.9903 5.0649
MI 2.7317 2.4886 2.8179 2.7892 2.9445 2.7919 2.7294 2.9201 2.9673
QAB/F 0.3889 0.1880 0.4571 0.4780 0.4908 0.4986 0.4591 0.4280 0.5127
PSNR 63.0474 61.0370 61.8477 63.1047 62.8989 62.4608 63.1206 62.9206 63.2085
SD 69.6090 34.0936 75.9801 75.5034 71.2350 64.4672 80.3984 79.6444 81.7336
AG 7.8589 4.0245 7.8287 8.6292 6.9688 9.8363 7.6959 7.6707 9.8877
Time (sec) 0.3975 128.564 58.2724 3.0669 2023.22 0.2693 32.5230 4.2180 52.1087

The bold values indicate the best result of the evaluation criterion

Table 7.

Assessment of different fusion methods on MR-T2/SPECT images

Method and metrics DWT NSCT NSCT-SR NSCT-PCLP Proposed
MI 3.0506 2.9943 3.2353 3.4872 4.0813
QAB/F 0.6898 0.7186 0.7339 0.7064 0.7702
Qncie 0.8075 0.8073 0.8082 0.809 0.8118
SD 66.2309 65.9492 70.9162 78.4080 80.3603
AG 10.3300 9.9493 10.3019 10.0249 10.5105
Time (sec) 0.9985 4.5813 33.8711 6.941724 3.983949

The bold values indicate the best result of the evaluation criterion

Table 8.

Assessment of different fusion methods on MR-T1/PET images

Method and metrics DWT NSCT NSCT-SR NSCT-PCLP Proposed
MI 3.0242 2.8750 3.3302 3.5152 3.8364
QAB/F 0.6921 0.6159 0.7463 0.7617 0.7837
Qncie 0.8073 0.8069 0.8084 0.8093 0.8108
SD 80.0882 78.8085 88.1059 91.2964 94.3089
AG 11.2366 10.8065 11.4835 11.7440 11.7520
Time (sec) 2.1533 3.2737 36.3586 6.4638 4.8907

The bold values indicate the best result of the evaluation criterion

Table 7 shows the quantitative and objective assessments of the proposed algorithm and the compared methods on MR-T2/SPECT images. The proposed algorithm is better than other compared methods in MI, QAB/F, Qncie, SD, and AG values. The time in the DWT is better than the proposed algorithm time. Table 8 shows the quantitative and objective assessments of the proposed algorithm and the compared methods on MRT1/PET images. Our proposed algorithm has higher values than other compared methods in MI, QAB/F, Qncie, SD, and AG. The time in the DWT is better than the proposed algorithm time. The results show that the proposed algorithm performs better than other compared methods in both objective and visual quality, retaining more information from the source images.

In this paper, major objective metrics including EN, MI, QAB/F, PSNR, SD, and AG have evaluated the fusion performance for the DWT, MPCNN, CSR, NSCT, CSMCA, NSCT-SR, NSCT-PCLP, and the proposed algorithm using MR-T1/MRT2, CT/MR-GAD, CT/MR-PD, and CT/MR-T2 images. These metrics are represented in Figs. 11, 12, 13, 14, 15, 16, and 17. For MR-T2/SPECT and MR-T1/PET images, the fusion performance for the DWT, NSCT, NSCT-SR, NSCTPCLP, and the proposed algorithm is evaluated in Figs. 12, 13, 15, 16, 17, and 18.

Fig. 11.

Fig. 11

Fusion result of (a) MR-T1 and (b) PET using (c) DWT method, (d) NSCT method, (e) NSCT-SR method, (f) NSCT-PCLP method, and the proposed method (g) NSCT-PCNN

Fig. 12.

Fig. 12

Mutual information assessment of different fusion methods compared to proposed method

Fig. 13.

Fig. 13

QAB/F assessment of different fusion methods compared to proposed method

Fig. 14.

Fig. 14

PSNR assessment of different fusion methods compared to proposed method

Fig. 15.

Fig. 15

SD assessment of different fusion methods compared to proposed method

Fig. 16.

Fig. 16

AG assessment of different fusion methods compared to proposed method

Fig. 17.

Fig. 17

Time assessment of different fusion methods compared to proposed method

Fig. 18.

Fig. 18

Qncie assessment of different fusion methods compared to proposed method

Conclusion

In this paper, a new multimodal medical image fusion algorithm is proposed. The proposed algorithm is based on the NSCT and PCNN methods. This algorithm is divided into three main steps: decomposition, fusion rule, and reconstruction. First, the NSCT method is applied to decompose two input images from multi-sensors. In this step, the input images are decomposed by the NSCT method into low- and high-frequency subbands. Then, apply the PCNN method as a fusion rule that fuses both the high- and low-frequency subbands. Finally, apply the inverse of the NSCT method to both fused low- and high-frequency subbands and construct the final fused image. Our experiments are implemented on six sets of medical images: MR-T1 and MR-T2 images, CT and MR-Gad images, CT and MR-PD images, CT and MRT2 images, MR-T2 and SPECT images, and MR-T1 and PET images were obtained from the Whole Brain Atlas database. To evaluate the performance of the proposed algorithm, we use common fusion metrics, namely entropy, mutual information, QAB/F, PSNR, standard deviation, Qncie, and average gradient. The experimental results show that the proposed algorithm has high performance as compared with others.

Abbreviations

AG

Average gradient

CNN

Convolutional neural networks

CSMCA

Convolutional sparsity-based morphological component analysis

CSR

Convolutional sparse representation

CT

Computed tomography

CT

Contourlet transform

DWT

Discrete wavelet transform

EN

Entropy

GB

Giga byte

GFS

Guided image filter and statistics

IPCNN

Improved PCNN

LP

Laplacian pyramid

MI

Mutual information

MPCNN

Multi-channel model–pulse coupled neural networks

MRI

Magnetic resonance imaging

MSE

Mean square error

MST

Multiscale transform

NSCT

Nonsubsampled contourlet transform

NSCT-PCLP

Nonsubsampled contourlet transform–phase congruency local Laplacian

NSCT-SR

Nonsubsampled contourlet transform–sparse representation

NSPFB

Nonsubsampled pyramid filter bank

NSDFB

Nonsubsampled directional filter bank

NSST

Nonsubsampled shearlet transform

PCA

Principal component analysis

PCNN

Pulse coupled neural network

PET

Positron emission tomography

PSNR

Peak signal to noise ratio

RMSE

Root mean square error

SD

Standard deviation

SF

Spatial frequency

SPECT

Single-photon emission computed tomography

SSIM

Structural similarity model

TB

Terabyte

Biographies

Sa.I. Ibrahim

received B.Sc. and M.Sc. degrees from Suez Canal University in 2011 and 2017, respectively. She is a teacher assistant at Suez Canal University. Her interests are image fusion and image processing.graphic file with name 11517_2022_2697_Figa_HTML.jpg

M.A. Makhlouf

received the M.Sc. degree in expert systems from Cairo University and the Ph.D. degree in computer science from Zagazig University. He is currently an Associate Professor at Suez Canal University.graphic file with name 11517_2022_2697_Figb_HTML.jpg

Gh.S. El-Tawel

received the M.Sc. and Ph.D. degrees from Helwan and Cairo Universities in 2000 and 2005, respectively. Currently a professor and dean of the Faculty of Computer & Informatics at Suez Canal University.graphic file with name 11517_2022_2697_Figc_HTML.jpg

Author contribution

Sa.I. Ibrahim: participated in sample collection and performed data analyses. Drafted the manuscript, designed the experiment, wrote the MS, thoroughly revised and modified the MS. M.A. Makhlouf: participated in sample collection, read and approved the manuscript. Gh.S. El-Tawel: designed the research plan and organized the study, read and approved the manuscript.

Funding

Open access funding provided by The Science, Technology & Innovation Funding Authority (STDF) in cooperation with The Egyptian Knowledge Bank (EKB).

Declarations

Ethical standard

This article is original and contains unpublished material. The corresponding author confirms that all of the other authors have read and approved the manuscript and no ethical issues involved.

Conflict of interest

The authors declare no competing interests.

Contributor Information

Sa.I. Ibrahim, Email: sara_hassan@ci.suez.edu.eg

M. A. Makhlouf, Email: m.abdallah@ci.suez.edu.eg

Gh.S. El-Tawel, Email: ghada@ci.suez.edu.eg

References

  • 1.Bavirisetti DP, Kollu V, Gang X, Dhuli R. Fusion of MRI and CT images using guided image filter and image statistics. Int J Imaging Syst Technol. 2017;27(3):227–237. doi: 10.1002/ima.22228. [DOI] [Google Scholar]
  • 2.Burt PJ, Adelson EH (1987) The Laplacian pyramid as a compact image code. In: Readings in computer vision. Elsevier, pp 671–679
  • 3.Ding Z, Zhou D, Nie R, Hou R, Liu Y (2020) Brain medical image fusion based on dual-branch CNNs in NSST domain. BioMed Res Int 2020. 10.1155/2020/6265708 [DOI] [PMC free article] [PubMed]
  • 4.Du J, Li W, Xiao B, Nawaz Q. Union Laplacian pyramid with multiple features for medical image fusion. Neurocomputing. 2016;194:326–339. doi: 10.1016/j.neucom.2016.02.047. [DOI] [Google Scholar]
  • 5.Eckhorn R, Reitboeck HJ, Arndt M, Dicke P. Feature linking via synchronization among distributed assemblies: simulations of results from cat visual cortex. Neural Comput. 1990;2(3):293–307. doi: 10.1162/neco.1990.2.3.293. [DOI] [Google Scholar]
  • 6.Ganasala P, Kumar V. CT and MR image fusion scheme in nonsubsampled contourlet transform domain. J Digit Imaging. 2014;27(3):407–418. doi: 10.1007/s10278-013-9664-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Gomathi PS, Kalaavathi B, et al. Multimodal medical image fusion in non-subsampled contourlet transform domain. Circuits and Syst. 2016;7(08):1598. doi: 10.4236/cs.2016.78139. [DOI] [Google Scholar]
  • 8.Gong J, Wang B, Qiao L, Xu J, Zhang Z (2016) Image fusion method based on improved NSCT transform and PCNN model. In: 2016 9th international symposium on computational intelligence and design (ISCID). IEEE, vol 1, pp 28–31
  • 9.James AP, Dasarathy BV. Medical image fusion: a survey of the state of the art. Inf fusion. 2014;19:4–19. doi: 10.1016/j.inffus.2013.12.002. [DOI] [Google Scholar]
  • 10.Kaur H, Koundal D, Kadyan V (2021) Image fusion techniques: a survey. Archives of Computational Methods in Engineering pp 1–23 [DOI] [PMC free article] [PubMed]
  • 11.Keith A, Johnson JAB Whole brain atlas. http://www.med.harvard.edu/aanlib/. Last accessed on 10 April 2021
  • 12.Li B, Peng H, Wang J. A novel fusion method based on dynamic threshold neural p systems and nonsubsampled contourlet transform for multi-modality medical images. Signal Process. 2021;178:107793. doi: 10.1016/j.sigpro.2020.107793. [DOI] [Google Scholar]
  • 13.Li S, Yang B, Hu J. Performance comparison of different multi-resolution transforms for image fusion. Information Fusion. 2011;12(2):74–84. doi: 10.1016/j.inffus.2010.03.002. [DOI] [Google Scholar]
  • 14.Liu Y, Chen X, Cheng J, Peng H (2017) A medical image fusion method based on convolutional neural networks. In: 2017 20th international conference on information fusion (Fusion). IEEE, pp 1–7
  • 15.Liu Y, Chen X, Ward RK, Wang ZJ. Image fusion with convolutional sparse representation. IEEE Signal Process Lett. 2016;23(12):1882–1886. doi: 10.1109/LSP.2016.2618776. [DOI] [Google Scholar]
  • 16.Liu Y, Chen X, Ward RK, Wang ZJ. Medical image fusion via convolutional sparsity based morphological component analysis. IEEE Signal Process Lett. 2019;26(3):485–489. doi: 10.1109/LSP.2019.2895749. [DOI] [Google Scholar]
  • 17.Liu Y, Liu S, Wang Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf fusion. 2015;24:147–164. doi: 10.1016/j.inffus.2014.09.004. [DOI] [Google Scholar]
  • 18.Mankar R, Daimiwal N (2015) Multimodal medical image fusion under nonsubsampled contourlet transform domain. In: 2015 International Conference on Communications and Signal Processing (ICCSP). IEEE, pp 0592–0596
  • 19.Nazrudeen M, Rajalakshmi MM, Sureshkumar MS. Medical image fusion using non-subsampled contourlet transform. Int J Eng Res (IJERT) 2014;3(3):1248–1252. [Google Scholar]
  • 20.Polinati S, Dhuli R (2019) A review on multi-model medical image fusion. In: 2019 international conference on communication and signal processing (ICCSP). IEEE, pp 0554–0558
  • 21.Polinati S, Dhuli R. Multimodal medical image fusion using empirical wavelet decomposition and local energy maxima. Optik. 2020;205:163947. doi: 10.1016/j.ijleo.2019.163947. [DOI] [Google Scholar]
  • 22.Tan W, Thitøn W, Xiang P, Zhou H. Multi-modal brain image fusion based on multi-level edge-preserving filtering. Biomed Signal Process Control. 2021;64:102280. doi: 10.1016/j.bspc.2020.102280. [DOI] [Google Scholar]
  • 23.Tan W, Zhang J, Xiang P, Zhou H, Thitøn W (2020) Infrared and visible image fusion via NSST and PCNN in multiscale morphological gradient domain. In: Optics, photonics and digital technologies for imaging applications VI, vol. 11353, p. 113531E. International Society for Optics and Photonics
  • 24.Tian Y, Li Y, Ye F (2016) Multimodal medical image fusion based on nonsubsampled contourlet transform using improved PCNN. In: 2016 IEEE 13th international conference on signal processing (ICSP). IEEE, pp 799–804
  • 25.Tirupal T, Mohan BC, Kumar SS. Multimodal medical image fusion techniques-a review. Curr Signal Transduct Ther. 2020;15(1):1–22. [Google Scholar]
  • 26.Wang Z, Ma Y. Medical image fusion using m-PCNN. Inf fusion. 2008;9(2):176–185. doi: 10.1016/j.inffus.2007.04.003. [DOI] [Google Scholar]
  • 27.Wang Z, Wang S, Zhu Y. Multi-focus image fusion based on the improved PCNN and guided filter. Neural Process Lett. 2017;45(1):75–94. doi: 10.1007/s11063-016-9513-2. [DOI] [Google Scholar]
  • 28.Xia J, Chen Y, Chen A, Chen Y (2018) Medical image fusion based on sparse representation and PCNN in NSCT domain. Computational and Mathematical Methods in Medicine 2018 [DOI] [PMC free article] [PubMed]
  • 29.Xu X, Shan D, Wang G, Jiang X. Multimodal medical image fusion using PCNN optimized by the QPSO algorithm. Appl Soft Comput. 2016;46:588–595. doi: 10.1016/j.asoc.2016.03.028. [DOI] [Google Scholar]
  • 30.Yin M, Liu X, Liu Y, Chen X. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas. 2018;68(1):49–64. doi: 10.1109/TIM.2018.2838778. [DOI] [Google Scholar]
  • 31.Zhang Q, Guo BI. Multifocus image fusion using the nonsubsampled contourlet transform. Signal Process. 2009;89(7):1334–1346. doi: 10.1016/j.sigpro.2009.01.012. [DOI] [Google Scholar]
  • 32.Zhu Z, Zheng M, Qi G, Wang D, Xiang Y. A phase congruency and local Laplacian energy based multi-modality medical image fusion method in NSCT domain. IEEE Access. 2019;7:20811–20824. doi: 10.1109/ACCESS.2019.2898111. [DOI] [Google Scholar]

Articles from Medical & Biological Engineering & Computing are provided here courtesy of Springer

RESOURCES