Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2024 Dec 4;14:30246. doi: 10.1038/s41598-024-81594-6

Optimized dual-tree complex wavelet transform aided multimodal image fusion with adaptive weighted average fusion strategy

Jampani Ravi 1,, R Narmadha 1
PMCID: PMC11618366  PMID: 39632891

Abstract

Image fusion is generally utilized for retrieving significant data from a set of input images to provide useful informative data. Image fusion enhances the applicability and quality of data. Hence, the analysis of multimodal image fusion is a new to the research topic, which is designed by combining the images of multimodal into single image in order to preserveexact details. On the other hand, the existing approaches face challenges in the precise interpretation of source images, and also it have only captured local information without considering the wide range of information. To consider these weaknesses, a multimodal image fusion model is planned to develop according to the multi-resolution transform along with the optimization strategy. At first, the images are effectively analyzed from standard public datasets and further, the images given into the Optimized Dual-Tree Complex Wavelet Transform (ODTCWT) to acquire low frequency and high frequency coefficients. Here, certain parameters in DTCWT get tuned with the hybridized heuristic strategy with the Probability of Fitness-based Honey Badger Squirrel Search Optimization (PF-HBSSO) to enhance the decomposition quality. Then, the fusion of high-frequency coefficients is performed using adaptive weighted average fusion technique, whereas the weights are optimized using PF-HBSSOto achieve the optimal fused results. Similarly, the low-frequency coefficients are combined by average fusion. Finally, the fused images undergo image reconstruction using the inverse ODTCWT. The experimental evaluation of the designed multimodal image fusion illustratessuperioritythat distinguishes this work from others.

Subject terms: Health care, Medical research

Introduction

Image fusion is a recently emerging concept owing to the amplifying requirements of diverse image processing applications especially in medical-aided diagnosis, video surveillance, remote sensing, and so on1. Image fusion is rapidly growing with various imaging sensors and the accessibility of a huge range of imaging techniqueslikeComputed Tomography (CT), and Magnetic Resonance Images (MRI), etc. It has explored the healthcare community for efficient decision-making and treatment to patients2. In addition, the major goal of image fusion also considers various factors like the fused image should be reliable and robust, inconsistencies or artifacts have to be eliminated, and salient information in any of the inputs must not be eradicated2. On the other hand, the major issues of image fusion research are similarity across modalities as data formation can be statistically uncorrelated and completely different, the efficient feature illustration of every modality and image noise, etc.3. In addition to these requirements in real-time applications, image fusion-guided disease prognosis and diagnosis have been formulated to assist medical professionals in decision-making as there is a restriction on human interpretation of clinical images owing to their subjectivity4. The major reason of multimodal clinical image fusion is to get the information with superior quality by combining the complementary information from various source images5.

Image fusion can be performed in various ways including multi-focus, multi-temporal, multi-modal, and multi-view fusion techniques6. Among these approaches, multi-modal image fusion is more essential, and it is carried out on images gathered by different sensors. Moreover, it is more helpful in getting precise results, especially in the medical field7. Generally, multi-modal image fusion combines both complementary and supplementary information of source images8.The multimodal image fusion results in final fused images that is free from redundant and random information. It minimizes the storage space, and stores one final fused image instead of storing two individual images9. The amalgamation of two different modalities results in precise localization or detection of abnormalities.The fundamental features of multi-modal image fusion are listed here. It reduces the uncertainty as the joint information from various sensors minimizes the vagueness related to the decision or sensing process10. The temporal and spatial coverage is extended for better performance. Fusion requires condensed representations to give complete information on images11. Multi-modal image fusion increases the system efficiency by reducing the redundancy in different measurements. It enhances reliability and reduces noise12.

Generally, image fusion is conducted via different approaches like transform domain as well as spatial domain methods. The consideration of high pass filtering, Intensity and Hue Saturation (IHS), Brovey method etc., are performed in the spatial domain methods. Thus, there is a need of adopting transform domain approaches come into the picture. Some popular transforms used for image fusion process include Contourlet (CT)13, Curvelet (CVT)14, Stationary Wavelet (SWT)12, Discrete Wavelet (DWT)11, DTCWT15, and Non-Subsampled Contourlet Transform (NSCT)16. While estimating with spatial domain methods, transform domain techniques get higher efficiency in terms of image fusion.On the other hand, a reliable, accurate, and appropriate image fusion approach is needed for several classes of images in diverse domains that must be simply interpretable for getting superior image fusion performance17. Some challenges are inexpensive computation time, uncontrollable acquisition conditions, and errors found in fusing the images. Thus, there is a need of suggesting an innovative multi-modal image fusion approach with the integration of two medical images by adopting transform domain techniques.

The innovationssuggested in this paper are listed here.

  • To recommend a new Multimodal Image Fusion with intelligent approaches like ODTCWT and adaptive weighted average fusion strategy along with the suggestion of hybrid nature-inspired approaches for performing medical image fusion for better localization of abnormalities and disease diagnosis through the gathered images.

  • To propose a novel ODTCWT and adaptive weighted average fusion strategy for efficient image fusion using a new hybrid nature-inspired algorithm termed PF-HBSSO for obtaining the salient information from the input images and formulates the objective as the maximization of fusion mutual information.

  • To implement the PF-HBSSO algorithm by the combination of Honey Badger Algorithm (HBA) as well as Squirrel Search Algorithm (SSA) for recommending ODTCWT by optimizing the filter coefficients and weights for adaptive weighted average fusion model for increasing the convergence rate that aids in maximizing the fused image quality.

Followed by introduction, the forthcoming section is shown.Part II specifiesthe discussion on existing works. Part III recommends an innovativemodel for multimodal image fusion. Part IV specifies the generation of low and high frequency coefficientsusingODTCWT. Part V derives the fusion of high frequency and low frequency by heuristic algorithm. Part VI estimates the results and Part VII completes this paper.

Study on existing works

Literature review

Research work based on deep learning models

In 2021, Zuoet al.18 has presented a new automated multi-modal medical image fusion strategy using classifier-based feature synthesis with a deep multi-fusion scheme. They have used a pre-trained autoencoder to analyze the fusion strategy by multi-cascade fusion decoder and a feature classifier. The public datasets were used for analyzing the image fusion results. TheParameter-Adaptive Pulse Coupled Neural Network (PAPCNN)was used in thelow and high-frequency coefficients. This image fusion was especially used for classifying brain diseases via the final fused images.

In 2022, Sun et al.19 have suggested a new deep MFNet using LiDAR data and multimodal VHR aerial images for performing medical image fusion. The multimodal learning and attention strategy was utilized for adaptively fusing the intramodal and intermodal features. A multilevel feature fusion module, pyramid dilation blocks, and a multimodal fusion strategy were implemented. This proposed network has adopted the adaptive fusion of multimodal features, enhanced the effects of global-to-local contextual fusion, and improved the receptive field. Moreover, this network was optimized using a multiscale supervision training strategy. The ablation studies and simulation outcomes have supreme performance of the recommended MFNet.

In 2021, Fu et al.20 have proposed a novel multimodal biomedical image fusion approach through deep Convolutional Neural Network (CNN) and rolling guidance filter”. The VGG model was applied for enhancing the image details and edges. Here, the rolling guidance filter was intended for extracting the detail and base images. Here, the fusion was done on both perceptual, detail, and base images with three diverse fusion methods. They have then chosen the image decomposition constraints by simulation for getting the suitable structure and texture images. In addition, the normalization operation was used to the perceptual images to eradicate noise and feature variations. Finally, it has shown the superior fusion outcomes and achieved better performance in terms of various objective measures.

In 2022, Goyal et al.21 have collected the images from standard sources, where NSCT was used for extracting the features. Next, a Siamese Convolutional Neural Network (sCNN) was applied for getting the significant features by weighted fusion. In order to eradicate noise, a new method has beenenhanced rate. Finally, the combination of NSCT + sCNN + FOTGV strategies has helped in enhancing the image fusion and also exhibited higher performance on both quantitative and visual analysis.

In 2022, Venkatesan and Ragupathyan22 have suggested a medical image fusion approach for fusing both MRI and CT images for recommending a healthcare model. To get both spectral and spatial domain features, the hybrid technique by integrating Deep Neural Network and DWT was suggested for getting more accuracy rate while estimating with traditional approaches. The performance enhancement was noticed in terms of standard deviation and average entropy for the designed DWT-CNN fusion approach compared with other wavelet transform methods. The superior efficiency on image fusion was noticed and has achieved considerable fusion performance rate.

In 2018, Bernal et al.23 have suggested a supervised deep multimodal fusion model for automated human activity and egocentric action recognition to monitor and assist patients. This model has collected video data using body-mounted or egocentric camera and motion data gathered with wearable sensors. The performance was estimated on multimodal public dataset and has analyzed the efficiency. They have used CNN-LSTM architecture for performing the multimodal fusion to get the results regarding automated human activity and egocentric action recognition.

Research work based on machine learning algorithms

In 2021, Duan et al.24 have recommended a new regional medical multimodal image fusion by adopting Genetic Algorithm (GA)-derived optimized approach. A weighted averaging technique was recommended for averaging the images of the source clinical images. Next, a fast Linear Spectral Clustering (LSC) superpixel technique was used for getting the homogenous regions and preserved the detailed information of images, which has segmented the average images and obtained the superpixel labels. The most significant regions were chosen and produced a decision map. The efficiency of the designed fusion approach was estimated via various experimental evaluations. Finally, the performance estimation on GA-based image fusion has shown the superiority on final fused images over others.

Research work based on image processing techniques

In 2022, Kong et al.25 have implemented a new medical image fusion approach via Side Window Filtering (SWF) and Gradient Domain-Guided Filter Random Walk (GDGFRW) in the Framelet Transform (FT) domain. Initially, FT was used on standard multimodal images for getting the residual and approximate illustrations. Then, a new GDGFRW was used for integrating the superiority of GD and GFRW that has built for interpreting the sub-bands, and the fusion was done by SWF. Next, inverse FT was performed for getting the residual models have fused images. The performance were addressed the fusion issues and outperformed the recent representative ones regarding objective estimation and subjective visual efficiency.

Comparative analysis of the existing techniques and proposed model

In 2023, Zhang et al.26 have implemented the novel fusion approach using Infrared-To-Visible Object Mapping (IVOMFuse) for extracting the target region from the infrared image. Further, the Expectation–Maximization (EM) has evolved to tune the probabilities in the target region. The fused image was attained by combining PCA and average fusion strategy. Hence, the final validation was attained by considering the TNO, CVC14, and RoadScene to get the final outcomes. In 2022, Zhou et al.27 have suggested the differential image registration model termed as robust image fusion to assist thermal anomaly detection (Re2FAD). The fusion strategy has been effectively done to enhance the accuracy. In 2023, Gu et al.28 have implemented the improved end-to-end image fusion approach (FSGAN) model to enhance the image fusion approach. Here, the auxiliary network has been extracted to enhance performance with diverse experiments.

Due to the heterogeneous nature, the multimodal image fusion is challenging in accordance of misalignment and non-linear relationships between the input data26. Also, the decomposition based methods are not highly preferred in the fusion model27. However, there is a complexity of discovering better multimodal images with fusion quality estimation in the suggested image fusion approaches28.To eradicate the drawbacks in the existing techniques, an effective deep learning method is implemented in the multimodal image fusion model. By considering the decomposition model, the multimodal image fusion gets enhanced by analyzing the texture details and smoothen the layers. It has the ability to maximize image quality to detect the performance effectively. The diverse implementation outcome is done whereas the recommended framework ensures to get better reliable outcomes.

Problem specification

In multimodal image fusion, it is very challenging to perform the multi-scale analysis that intends to analyze the feature maps extracted using shearlet domain. The analysis of the strengths and weakness of the existing models is given in Table 1. LSC and GA24 are very efficient in both the objective evaluation and visual effects in the segmentation of medical images. However, when increasing the region count, the fusion efficiency may get reduced and increase the running time in the image fusion. PAPCNN18 provides accurate and detailed information present in the fusion results. On the other hand, it does not completely utilize the fusion layer and decoding layer, which is observed through the quantification evaluations. Deep MFNet19 attains better performance regarding the visualization and quantification when considering the quantitative and qualitative evaluations. Yet, it does not consider the multi-scale decomposition for encoding and decoding to get performance enhancement. VGG network20 ensures the final fused images through the combination of three informative images like fused base image, detail image as well as perceptual image. Still, it has provided the results with increased color distortion and fusion noise without considering the fusion quality. sCNN21 are trained with the concatenated features for considering huge significance.But, it is time-consuming and cannot perform region-based medical image fusion. DWT-CNN22 is efficient on capturing the high-level association among the modalities and obtains the feature descriptors from the spatiotemporal regions. Yet, it may fail on preserving the shift-invariance. GDGFRW25 has an ability to understand the temporal patterns of behavior over data modalities, which have been hidden through the overriding of individual modality. Still, it is unable to perform multi-focus image fusion. RNN23 has efficiently utilized the location and hand presence as their significant cues for automatically classifying the images. Yet, it does not support the practical, feedback and in-device inference in the fusion. Hence, it is important to develop an enhanced multimodal image fusion model with superior optimization strategy.

Table 1.

Strengths and weakness of existing multimodal image fusion models.

Author [citation] Techniques Strengths Weakness
Duan et al.24 LSC and GA It is very efficient in providing clear visual effects When increasing the regions count, the fusion efficiency may get reduced and increases the running time in the image fusion
Zuoet al.18 PAPCNN It provides accurate and detailed information present in the fusion results It does not completely utilize the fusion layer and decoding layer, which is observed through the quantification evaluations
Sun et al.19 Deep MFNet It attains better performance regarding the visualization and quantification when considering the quantitative and qualitative evaluations It does not consider the multi-scale decomposition for encoding and decoding to get performance enhancement
Fu et al.20 VGG network It ensures the final fused images with the help of three informative images like fused base image, detail image as well as perceptual image It has provided the results with increased color distortion and fusion noise without considering the fusion quality
Goyal et al.21 sCNN It is trained with concatenated features for considering huge significance It is time-consuming and it cannotperformregion-based medical image fusion
Venkatesan and Ragupathyan22 DWT-CNN It is efficient on getting the high-level association among the different modal images and obtains the feature descriptors from the spatiotemporal regions It may fail on preserving the shift-invariance
Kong et al.25 GDGFRW It has an ability to understand the temporal patterns of behaviour over data modalities, which have been hidden through the overriding of individual modality It is unable to perform multi-focus image fusion
Bernal et al.23 RNN It has efficiently utilized the location and hand presence as their significant cues for automatically classifying the images It does not support the practical, feedback and in-device inference in the fusion

An intelligent model for multimodal image fusion

Collection of dataset

This multimodal image fusion approach gathers medical images for performing the fusion, which helps in healthcare systems regarding better treatment planning and decision making.It is available from https://www.med.harvard.edu/aanlib/home.htm. The proposed model considers the images of MRI and Single Photon/Positron Emission Computed Tomography (SPECT/PET). The resolution of theimageis taken as 256 × 256. In total, the 66 images are considered for the evaluation. The gathered images are known as Ca, where a=1,2,3,,A, and the final number of obtained images is denoted as A.The sample images of MRI and SPECT/PET are visualized in Fig. 1.

Figure 1.

Figure 1

Sample images in terms of MRI and SPECT/PET.

Proposed multimodal image fusion model

Imaging technology in the healthcare applications need a huge amount of information that helps to provide the additional requirement for medical image fusion. It is further split into single-modal and multimodal image fusion. Various researchers are focused on designing multi-modal image fusion with several complications in the information offered uisng single-modal fusion. The multimodal image fusion comprises both physiological and anatomical data, which makes disease detection easier. Various modalities in the medical field are SPECT, PET, MRI, CT, etc. It has offered medical information regarding the human body’s structural properties such as soft tissue, etc. Different imaging approaches preserve diverse characteristics regarding the similar part. Hence, the reason for image fusion is to get a superior perceived experience, fusion quality, and contrast. The better image fusion results must require to follow the below constraints like avoiding bad states like noise and misregistration from the images. In classical approaches, the issues presented in the fusion effects are enhanced but not addressed effectively including feature information extraction and color distortion. Thus, there is a need to utilize innovative and intelligent approaches for medical image fusion that remains a major complication in this research area. Finally, it is concluded that there is a huge requirement for medical image fusion with multimodal medical images for getting better functional and structural information about the same part and thus, the fused images will be high-quality information preserved images.Consequently,the multimodal image fusion model is designed with intelligent approaches in this paper, which promotes medical image fusion. The visual representation ofdevelopedmodel is depicted in Fig. 2.

Figure 2.

Figure 2

Architectural representation of multimodal image fusion with developed framework.

A new multimodal image fusion approach is recommended here, especially for the medical fieldwith the help of multi-resolution transform with the optimization strategy. Firstly, the source medical images are collected from benchmark sources. The next process is to decompose both images using the ODTCWTto acquire the frequency coefficient of low and high. Here, the filter coefficients of DTCWT are tuned using recommended PF-HBSSO algorithm.The decomposition helps in distinguishing the frequencies from the images to get the texture as well as smooth layer. The individual processing of both frequency parts helps in better preservation of images. Next, the fusion of high-frequency coefficients is performed bythe adaptive weighted average fusion technique, where the weights are optimized using the same PF-HBSSO algorithmto achieve the optimal fused results. Consequently, the low-frequency coefficients are fused usingstandard average fusion technique. At last, the fused image is retained using inverse ODTCWT to maximize thefused mutual information and ensure the quality of fused images.

Generating low and high “frequency coefficients” by optimized dual-tree complex wavelet transform

Optimization concept by PF-HBSSO

In this proposed multimodal image fusion approach, a new heuristic algorithm is recommended with the adoption of two recently familiar algorithms like SSA29 and HBA30. Here, the suggested model uses this new PF-HBSSO algorithm for maximizing the performance of image fusion regarding maximization of fused mutual information. PF-HBSSO algorithm optimizes the frequency coefficientsin DT-CWT and also the weights used for weighted average fusion for fusing the high frequency coefficients. This innovation increases the efficiency of the designed model while estimating with other approaches, which is detailed in result section. Here, HBA is chosen for performance enhancement owing to their vast range of features like skill towards sustaining the swapping among exploitation and exploration phases, good diversity, convergence speed, and provides statistical significanceto handle complex optimization problems, and utilization of the dynamic search schemes. Conversely, it faces complications in handling the local optima solutions. Henceforth, a SSA is adopted into this mechanism for giving better performance owing to their higher efficiency, capability towards producing the better solutions at faster manner even for critical and high dimensional optimization problems.

A new PF-HBSSO algorithm is recommended in this paper by modifying the random parameter b used in HBA technique, where this random parameter b is assigned in the range of [0, 1] in conventional HBA. On the other hand, this sameparameter b is implemented in this PF-HBSSO by taking the probability of fitness-based solutions as shown in Eq. (1).

b=Pα 1

Here, the P determines the number of population and α is specified as the capability of the individuals to reach food (α ≥ 1) as given in traditional HBA. Moreover, in this recommended PF-HBSSO,α is found by determininghow many number of fitness is less than mean fitness, which results in getting the better reach of optimal solutions at higher convergence rate.

Consequently, this new parameter b is used for updating the solutions from either HBA or SSA with the following conditions. If b<0.5 is verified, then the solution updating is carried out via digging phase of HBA or else the solution updating is performed using SSA. Here, the solution updating in SSA is carried out via formulating the case 1 with the condition of total flying squirrels. Hence, it produces higher convergence rate and free from local optimal with superior outcomes, which increases the efficiency of image fusion.

Initially, the population of search individual is created based on HBA, as derived in Eq. (2).

zj=lj+b1uj-lj 2

The jth position of honey badger is specified in Eq. (4). In Eq. (2), a random number with limit of [0, 1] is specified as b1, the upper and lower bounds of search range istermed as uj and lj and the jth position of individual referring candidate solution is derived as zj.

z=z11z12z21z22z1Sz2Szp1zp2zpS 3
zj=zj1,zj2,,zjS 4

The size of the population is shown as P. In the next stage,the parameter b is implemented by Eq. (1) and verifies b<0.5. Then, the solution updating is started with HBA while satisfying the b<0.5.

HBA has two phases for updating their solutions, which aredigging phase and honey phase. But, PF-HBSSO algorithm only formulates the digging as well as honey phase is replaced by SSA.

The digging phase is executed while honey badger performs in Cardioid shape as shown in Eq. (5).The position is updated by evaluating digging as well as the honey phase.Here, the position is updated by the several movement patterns that help to sort better solution during the optimization problem. The updated position is generally suited for the global optimization in which it maximizes or minimizes the multivariate function to find out the optimal solution. It helps to set the desired and reliable outcomes. The position update mechanism helps to avoid from the premature convergence to maximize the algorithms efficiency and effectiveness.

znew=zprey+D×α×It×zprey+D×b3×δ×hj×cos2πb4×1-cos2πb5 5
D=1ifb60.5-1else 6
It=b2×CS4π·hj2 7
CS=yj-yj+12 8
hj=zprey-zj 9
δ=G×exp-iimax 10

The variables b2,b2,b4,b5,b6 and b7 is denoted as random numbers with a specific interval of [0, 1].Also,the global prey position is denoted as zprey, the flag utilized for modifying the search direction is mentioned as D, and also the term It specifies the smell intensity, distance among the individual and prey is known as hj, the concentration strength is specified as CS, density factor is termed as δ and the constant term is derived as G, where the maximum number of iterations denotes the imax and iteration number is specified as i.

Then, the proposed PF-HBSSO verifies the b<0.5 factor, if it is not satisfied then; the solutions are executed by formulating the case 1 of SSA with the condition of total flying squirrels.

The location of the search individuals is updated while they move from the acorn treesto hickory nut tree as derived in Eq. (11).

zactnew=zactold+rG×KC×zhiti-zactiifv1ρdgrandompositionor 11
rG=yGtanϕ×o 12
tanϕ=QX 13
Q=12χe2srCrdd 14
X=12χe2srCrll 15

In the afore mentioned equations, the constant terms are correspondingly derived as χ,o,yG,e,sr,Crll and Crdd, the and the lift force is represented by X, Q specifies the drag force, and the gliding angle is known as tanϕ, random gliding distance is indicated by rG, the random function v1 is computed at the interval of 0,1 and the gliding constant is derived by KC and the position of search individual that reached hickory nut tree is shown as zhiti, the position of search individual in acorn tree is derived as zacti. Moreover, the predator presence probability ρdg is the essential role in updating position of individuals.

Finally, the search individuals are updated and the optimal solutions are attained for enhancing the efficiency of image fusion. Here, the new parameter updating for integrating two familiar algorithms gives the better efficiency and it is exhibited in results.

The pseudo-code of PF-HBSSOisshown in Algorithm 1.

graphic file with name 41598_2024_81594_Figa_HTML.jpg

Algorithm 1: PF-HBSSO

The hybrid heuristic algorithms exhibit better significance in recent days and thus, this paper also recommend one hybrid algorithm to increase efficiency of image fusion.The applications of the suggested PF-HBSSO model are solving unimodal, multimodal, and multi-dimensional optimization problems, system control, machine design and engineering planning. The flowchart of the designed PF-HBSSO is visualized in Fig. 3.

Figure 3.

Figure 3

Flowchart of the designed PF-HBSSO algorithm.

Optimized DT-CWT-based image decomposition

In the recommended model, the decomposition of both MRI and SPECT/PET is done using ODT-CWT, where a newly recommended PF-HBSSO optimizes the filter coefficients of DTCWT for getting the better fusion effects that promotes the medical diagnosis approach.The gathered images Ca are given to ODT-CWT, where the decomposition of images is done via ODT-CWT for getting low frequency and high frequency coefficients.

DT-CWT31 is the most eminent approach used in image fusion approaches, where the masks are used for extracting the information from the decomposed structure. It is the extended version of DWT, which is processed by executing two parallel trees. It is useful in eradicating the aliasing effects and achieved shift invariance.It helps reveal the visual sensitivity, which comprises of real and complex coefficients. The gathered images Ca are given to ODT-CWT for getting frequencies of low L and high H in Eq. (16).

L,H=ODT-CWTCa 16
LFU=ϕLL1,L2 17
HFU=ϕHH1,H2 18

Here, the fusion rules for high and low frequency coefficients are correspondingly known as ϕH, and ϕL, which are optimized by PF-HBSSO algorithm.

Finally,ODT-CWT offers ideal reconstruction over the traditional wavelet transform for getting better multimodal image fusion approach for medical images, which will be explained in upcoming sections.The gathered images Ca are given to ODT-CWT, where decomposition of images is done via ODT-CWT for getting low and high frequency.

The framework of ODT-CWT for image decomposition using recommended PF-HBSSO algorithm is expressed in Fig. 4.

Figure 4.

Figure 4

Illustration of ODT-CWT for image decomposition using recommended PF-HBSSO algorithm

High frequency and low frequency image fusion by proposed heuristic algorithm

Developed objective model

The implemented multimodal image fusion approach aims to improve the performance rate with the help of PF-HBSSO algorithm. Here, the PF-HBSSO algorithm is used for optimizing the frequency coefficients in DT-CWT and also the weights used for the weighted average fusion method for fusing the high frequency coefficients. This model considers the major goal as the maximization of fused mutual information as equated in Eq. (19).

OF=argmaxϕH,ϕL,W1,W2FMI 19

Here, the high and low frequency coefficientsare optimized using PF-HBSSO algorithm and the weights used for the weighted average fusion method is represented as W1,W2 that is also optimized by PF-HBSSO algorithm. The range of ϕH,ϕL are assigned among [-20, 20] and W1,W2 are assigned among the range of [0, 1]. The optimal tuning of frequency coefficients results in better image decomposition whereas the weight optimization in weighted average fusion increases performance of high frequency fusion method. Term FMI represents fused mutual information. The fused mutual information is determined among the fused image and the source image as derived here.

FMI=FMIIF+FMIIS 20
FMIIF=ig=0Ysjg=0QsgtFiAiigjglog2gtFiAiigjggtFiigjggtAiigjg 21
FMIIS=ig=0Ysjg=0QsgtGiAiigjglog2gtGiAiigjggtGiigjggtAiigjg 22

In the aforementioned equations, the joint histogram among source and fused images are correspondingly specified as gtGAigjg, and gtFAigjg, the column and row size of the image are correspondingly known as Qs and Ys, and the normalized histogram of the source image 1 Fi, source image 2 Ai, and fused image Bi are specified accordingly.The higher mutual information value represents the superior quality of fused images.

High frequency optimization by adaptive weighted average fusion

The recommended model gets the high frequency coefficients of two different imaging modalities using ODT-CWT.It is fused using adaptive weighted average fusion strategy. This scheme is modeled in Eq. (23).

HBi=W1×FiH+W2×BiH 23

Here, the fused images are known as HBi, the weights utilized for fusing the high frequency coefficients of images are correspondingly specified as W1 and W2, where the range of weights is assigned byPF-HBSSO algorithm in the range of [0, 1], the high frequency coefficients of two source images are given as FiH and BiH, where the weights W1 and W2 are optimized using PF-HBSSO algorithm.The high frequency coefficients are fused via this recommended adaptive weighted average fusion scheme, which is carried out via optimizing the weights utilized for the fusion process.High frequency coefficients are fused for specifying the edge information to increase the image fusion quality. There is a need of maintaining the superiority of frequency information in the images for increasing the better contrast in the final fused images.

Finally, the final high frequency coefficients of two source images is attained as HBi, which is further given to the reconstruction process to get final fused images. The sample representation of adaptive weighted average fusion for fusing the high frequency images is given in Fig. 5.

Figure 5.

Figure 5

Fusion process of high frequency by adaptive weighted fusionmodel.

Low frequency optimization by average fusion

In this suggested image fusion model, the average fusion is performed on low frequency coefficients of two source images, which is used for storing the local information in terms of increasing the image fusion. This process is derived in Eq. (24).

LBi=FiL+BiL2 24

The average fusion method is performed via taking the averaging among two different modal source images. Averaging is the simplest technique in implementation, where the average of entire pixels from the input low-frequency coefficients of medical images is considered as the intensity of the output pixel. The averaging operation is useful in reducing the bad information and enhancing the good information from the images by taking a mean image. Although this approach is not eminent in image fusion, it is helpful in fusing the low-frequency coefficients.The final fused low frequency of two source images is attained as LBi. In Fig. 6, it represents the averaging fusion-based low frequency source image model.

Figure 6.

Figure 6

Fusion process of low frequency by adaptive weighted fusion model.

Image reconstruction by inverse ODT-CWT

In this designed multimodal image fusion approach, the final fused images are attained using inverse ODT-CWT from both MRI and SPECT/PET images, which is derived in Eq. (25).

Fua=IODT-CWTHBi,LBi 25

Here, the final fused images Fua are attained using inverse ODT-CWT via fusing both low frequency fused images LBi and high frequency fused images HBi. At last, the final fused images are attained via the image reconstruction stage and ensured its higher quality regarding the maximization of fused mutual information.

Experimental analysis

Validation setting

There commended multimodal image fusion framework was implemented in MATLAB 2020a, with different quantitative measures. Some techniques like Dragon Algorithm (DA)32, Grey Wolf Optimizer (GWO)33, HBA30, SSA29, and traditional transform approaches such as PCA34, DWT35, IHS36, DCT37, CWT38, NSCT39 and DT-CWT28. The experimentation was done via measures like number of population as 10, chromosome length as 82, and the maximum Iteration was considered as 10.The recent methods like Spatial-Frequency Information Integration Network (SFINet)40, Channel Attention dual adversarial Balancing network (CABnet)41, DUSMIF42 and Dense-ResNet43 are compared using the developed model.

Validation metrics

  1. SSIM: The SSIM measure helps to evaluate the local patterns of different pixel intensities. It is formulated in Eq. (26).
    SSIMRCa,Fua=2μRCaμFua+Vn12σRCaFua+Vn2μRCa2+μFua2+Vn1σRCa2+σFua2+Vn2 26

    Here, the constants are represented as Vn1 and Vn2, it is measured among two images RCa,Fua, and the average of Fua is termed as μFua,and the average of RCa is termed as μRCa. The covariance of RCa and Fua is termed as σRCaFua and the variance of RCa and Fua are termed as σRCa and σFua, respectively.

  2. BRISQUE: The non-referenced image Fua is calculated using score = Brisque (Fua).

  3. Entropy: It is used to measure the information content of a fused image. The high entropy value indicates the fused image has rich information content and given in Eq. (27).
    Ent=-ig=0NghhFuaiglog2hhFuaig 27

    Here, the term hhFua is specified as the probability of the gray level of fused image.

  4. PSNR: The PSNR is expressed in Eq. (28).
    PSNR=20log10Ng2RMSE2 28

    Here, the number of gray levels is denoted as Ng.

  5. RMSE: It is formulated in Eq. (29).
    RMSE=1Ys×Qsig=1Ysjg=1QsRCaig,jg-Fuaig,jg2 29

    Here, RCa is the reference image,Fua is the fused image and the intensity values of the reference and fused imageare accordingly represented as RCaig,jg and Fuaig,jg.

  6. Standard Deviation: It is used for measuring the fusion performance, where the larger standard deviation results show better fusion results. The STD is described in Eq. (30).
    STD=1Ys×Qsig=1Ysjg=1QsFuaig,jg-μ^212 30

    In Eq. (30), the mean value of the image is given as μ^.

Experimental images

Some of the resultant images attained using different techniques and proposed model are shown in Fig. 7.The recent image fusion techniques like Unified and Unsupervised end-to-end image fusion network (U2Fusion)37, Information Gate Network for multimodal medical image fusion (IGNFusion)38 and Fast, Lightweight Image Fusion Network (FLFuse-Net) are compared in the image fusion model. This resultant analysis helps the developed model show effective outcomes.

Figure 7.

Figure 7

Resultantimages attained using the existing method and proposed model.

Estimation over heuristic approaches

The effectiveness of the achieved fused images isanalyzed over various techniques as given in Fig. 8. From the evaluation, it is clearly shown that the designed model has exhibited its higher performance over traditional approaches. For example, while comparing with the recent heuristic algorithms, the recommended ODT-CWT-based PF-HBSSO algorithm specifies superior effectiveness over traditional methods.

Figure 8.

Figure 8

Figure 8

Efficiency analysis on designed model over heuristic approaches for (a) Fused mutual information, (b) Standard deviation, (c) Entropy, (d) Brisque, (e) SSIM, (f) PSNR, (g) RMSE and (h) SNR.

Estimation over transform approaches

The effectiveness of the designed image fusion model is estimated over traditional transform domain approaches as listed in Fig. 9. The final fused images are compared over conventional approaches using standard statistical measures to illustrate the efficiency of the multi-modal image fusion approach.

Figure 9.

Figure 9

Figure 9

Efficiency analysis on designed model over transform domain approaches for (a) Fused mutual information, (b) Standard deviation, (c) Entropy, (d) Brisque, (e) SSIM, (f) PSNR, (g) RMSE and (h) SNR.

Comparative estimationof image fusion over heuristic algorithms

The comparative estimation of the designed image fusion over various optimization algorithms is given in terms of various performance metrics from Tables 2, 3, 4, 5, 67, 8 and 9. The investigation clearly exhibited superior performance over both positive measures and error measures. The positive measures show the superior performance regarding image fusion quality whereas the negative measures specify the performance enhancement by getting lower error rates in image fusion. Hence, the designed model exhibit better performance and thus, it is more applicable for medical applications.

Table 2.

Efficiency estimation on multi-modal image fusionover heuristic algorithms for fused mutual information.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT
1 0.8525 0.85151 0.85586 0.83259 0.87444
2 0.8682 0.86711 0.87459 0.83907 0.87969
3 0.91896 0.91857 0.90787 0.91284 0. 93,112
4 0.86645 0.86428 0.87401 0.84555 0.88374
5 0.88159 0.88059 0.88691 0.85839 0.89245
6 0.87339 0.87252 0.87054 0.85232 0.89079
7 0.86841 0.86764 0.8725 0.83856 0.88029
8 0.86672 0.86576 0.86961 0.84187 0.87899

Table 3.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for standard deviation.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 0.18441 0.19791 0.17944 0.15973 0.075319
2 Image 2 0.20864 0.22428 0.20117 0.18487 0.085404
3 Image 3 0.14315 0.15275 0.14858 0.11729 0.058243
4 Image 4 0.17726 0.19063 0.17099 0.16133 0.072862
5 Image 5 0.1729 0.18498 0.17227 0.14846 0.070638
6 Image 6 0.15999 0.17152 0.15698 0.14393 0.065629
7 Image 7 0.21188 0.22769 0.20464 0.18641 0.086649
8 Image 8 0.19778 0.21236 0.19183 0.17146 0.080775

Table 4.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for entropy.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 4.9364 5.1441 4.8991 4.8477 4.1346
2 Image 2 4.5038 4.6521 4.5709 4.3923 3.694
3 Image 3 4.4418 4.6527 4.5535 5.0215 3.7606
4 Image 4 5.1221 5.3214 4.971 5.0402 4.2493
5 Image 5 3.2539 3.3675 2.8838 3.4634 2.5914
6 Image 6 4.1076 4.2943 4.1479 4.4421 3.4474
7 Image 7 4.5866 4.7862 4.5091 4.6408 3.8333
8 Image 8 4.6094 4.8272 4.57 4.5284 3.8149

Table 5.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for Brisque.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 34.72 33.499 33.975 34.113 50.491
2 Image 2 42.318 41.635 40.762 42.538 50.694
3 Image 3 35.131 34.103 33.798 29.979 46.193
4 Image 4 32.027 31.129 31.873 31.719 50.234
5 Image 5 51.256 51.128 51.811 50.567 49.221
6 Image 6 43.749 43.235 43.971 42.765 47.404
7 Image 7 40.128 39.314 39.872 39.115 52.629
8 Image 8 40.612 40.059 39.504 40.413 51.674

Table 6.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for SSIM.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 0.54999 0.64369 0.58262 0.61801 0.71643
2 Image 2 0.64224 0.69363 0.65925 0.59674 0.79625
3 Image 3 0.50736 0.48648 0.37731 0.45969 0.55752
4 Image 4 0.56224 0.69228 0.61818 0.63233 0.75819
5 Image 5 0.76257 0.77901 0.70539 0.77529 0.86377
6 Image 6 0.6102 0.65061 0.55874 0.55778 0.71829
7 Image 7 0.61764 0.6774 0.61425 0.65039 0.78316
8 Image 8 0.63523 0.70821 0.64291 0.69575 0.81592

Table 7.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for PSNR.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 20.684 20.646 20.168 19.739 15.254
2 Image 2 19.031 18.993 18.557 18.208 14.26
3 Image 3 15.789 15.792 15.627 15.511 11.32
4 Image 4 20.909 20.891 20.286 19.888 15.536
5 Image 5 20.754 20.727 20.073 20.1 16.6
6 Image 6 17.599 17.592 17.317 17.195 14.507
7 Image 7 20.585 20.544 19.911 19.41 14.535
8 Image 8 22.005 21.949 21.287 20.575 15.22

Table 8.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for RMSE.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 0.092423 0.092835 0.098082 0.10305 0.1727
2 Image 2 0.1118 0.11229 0.11808 0.12292 0.19364
3 Image 3 0.16239 0.16234 0.16545 0.16767 0.19566
4 Image 4 0.09006 0.090251 0.096756 0.1013 0.16718
5 Image 5 0.091685 0.091975 0.099161 0.098854 0.14791
6 Image 6 0.13183 0.13195 0.13619 0.13811 0.18822
7 Image 7 0.093485 0.093934 0.10103 0.10703 0.1876
8 Image 8 0.079389 0.079897 0.086228 0.093593 0.17338

Table 9.

Efficiency estimation on multi-modal image fusion over heuristic algorithms for SNR.

Description/images DA-DT-CWT32 GWO-DT-CWT33 HBA-DT-CWT29 SSA- DT-CWT30 PF-HBSSO- DT-CWT Description/images
1 Image 1 27.276 26.959 26.083 31.602 34.994
2 Image 2 30.031 29.38 29.161 34.029 38.593
3 Image 3 10.784 10.188 10.746 12.821 18.597
4 Image 4 27.033 26.664 25.178 31.938 35.092
5 Image 5 25.217 24.837 23.238 30.157 33.764
6 Image 6 22.915 22.257 22.519 25.582 31.039
7 Image 7 30.158 29.725 28.518 34.634 38.259
8 Image 8 31.562 30.903 29.694 37.241 40.532

Comparative estimation on image fusion over transform algorithms

The comparative analysis on the designed image fusion approach over various transform approaches for various performance metrics are given from Tables 10, 11, 12, 13, 14, 15, 16 and 17. By analyzing the values, the designed fusion model exhibits the better performance and clearly shows the superiority over the traditional methods.

Table 10.

Efficiency estimation on multi-modal image fusion over transform algorithms for fused mutual information.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 0.86364 0.85586 0.85153 0.85153 0.84941 0.84766 0.85613 0.87444
2 Image 2 0.87701 0.87451 0.86703 0.86703 0.86762 0.86405 0.85621 0.87969
3 Image 3 0.9253 0.90785 0.91939 0.91939 0.90232 0.90073 0.90937 0.93112
4 Image 4 0.8598 0.87402 0.86413 0.86413 0.86384 0.8418 0.87987 0.88374
5 Image 5 0.8844 0.88694 0.87999 0.87999 0.88161 0.86363 0.86911 0.89245
6 Image 6 0.88664 0.87043 0.87261 0.87261 0.86174 0.8568 0.8538 0.89079
7 Image 7 0.87038 0.87259 0.86816 0.86816 0.86618 0.85799 0.86487 0.88029
8 Image 8 0.87029 0.86968 0.86552 0.86552 0.86486 0.85432 0.86476 0.87899

Table 11.

Efficiency estimation on multi-modal image fusion over transform algorithms for standard deviation.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 0.1999 0.17952 0.23414 0.1883 0.22214 0.21621 0.16046 0.075319
2 Image 2 0.22644 0.20125 0.27255 0.21351 0.25925 0.21735 0.18211 0.085404
3 Image 3 0.25139 0.1487 0.26477 0.14561 0.20006 0.25318 0.11279 0.058243
4 Image 4 0.20299 0.17106 0.19827 0.18215 0.23181 0.20938 0.14284 0.072862
5 Image 5 0.18549 0.17237 0.21269 0.17659 0.22224 0.18368 0.14942 0.070638
6 Image 6 0.21435 0.15706 0.24439 0.16407 0.21501 0.23724 0.13394 0.065629
7 Image 7 0.22666 0.20472 0.24793 0.21662 0.25713 0.24881 0.18497 0.086649
8 Image 8 0.20907 0.19191 0.23295 0.20194 0.23704 0.20858 0.17424 0.080775

Table 12.

Efficiency estimation on multi-modal image fusion over transform algorithms for entropy.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 5.0084 4.9011 4.7997 4.9731 5.2523 5.9934 6.0335 4.1346
2 Image 2 4.4475 4.5728 4.3917 4.3967 4.9338 5.7999 5.8847 3.694
3 Image 3 4.8464 4.5564 4.8666 4.5999 4.9917 6.5128 5.7598 3.7606
4 Image 4 5.1871 4.9719 4.6253 5.147 5.4067 5.7485 5.8177 4.2493
5 Image 5 3.0693 2.8837 2.9753 3.0683 3.8068 5.1957 4.8821 2.5914
6 Image 6 4.3546 4.1496 4.1066 4.2122 4.713 5.8315 5.285 3.4474
7 Image 7 4.5861 4.5105 4.5238 4.5818 4.9843 5.7495 5.9722 3.8333
8 Image 8 4.5989 4.5713 4.5101 4.5678 4.8022 5.9588 5.9966 3.8149

Table 13.

Efficiency estimation on multi-modal image fusion over transform algorithms for Brique.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 34.558 33.97 36.564 33.974 31.267 47.121 39.126 50.491
2 Image 2 42.448 40.744 42.271 42.608 40.115 41.949 44.597 50.694
3 Image 3 33.791 33.806 33.519 33.301 32.158 34.555 39.778 46.193
4 Image 4 31.38 31.875 36.7 31.255 33.521 38.943 37.771 50.234
5 Image 5 51.26 51.808 51.589 51.25 49.286 51.231 51.001 49.221
6 Image 6 42.898 43.972 43.309 42.255 40.613 46.189 44.823 47.404
7 Image 7 39.965 39.841 39.953 39.584 34.686 46.161 42.549 52.629
8 Image 8 40.524 39.5 39.53 40.864 37.968 51.274 43.344 51.674

Table 14.

Efficiency estimation on multi-modal image fusion over transform algorithms forSSIM.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 0.64535 0.58229 0.63141 0.68435 0.45334 0.21178 0.27059 0.71643
2 Image 2 0.78573 0.65898 0.74121 0.64224 0.46706 0.23733 0.23503 0.79625
3 Image 3 0.53106 0.37661 0.52761 0.48648 0.22034 0.12426 0.20906 0.55752
4 Image 4 0.75748 0.6178 0.66031 0.56224 0.46006 0.25329 0.26437 0.75819
5 Image 5 0.86136 0.70487 0.8353 0.76257 0.54155 0.17554 0.32254 0.86377
6 Image 6 0.68648 0.55807 0.61084 0.6102 0.34708 0.14925 0.29471 0.71829
7 Image 7 0.78005 0.61389 0.73722 0.61764 0.49649 0.25615 0.24889 0.78316
8 Image 8 0.80643 0.64257 0.75402 0.63523 0.53907 0.22809 0.3035 0.81592

Table 15.

Efficiency estimation on multi-modal image fusion over transform algorithms for PSNR.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 20.082 20.166 17.755 20.729 18.929 17.713 18.092 15.254
2 Image 2 18.63 18.554 15.841 19.068 17.167 15.705 16.467 14.26
3 Image 3 13.264 15.626 12.787 15.8 14.834 12.767 14.842 11.32
4 Image 4 20.245 20.283 17.405 20.958 18.446 17.044 17.661 15.536
5 Image 5 20.475 20.069 17.59 20.797 18.096 17.444 18.3 16.6
6 Image 6 15.963 17.315 14.508 17.615 16.15 14.507 16.149 14.507
7 Image 7 20.367 19.908 17.363 20.645 18.301 17.063 17.281 14.535
8 Image 8 21.749 21.284 18.582 22.068 19.652 18.47 18.379 15.22

Table 16.

Efficiency estimation on multi-modal image fusion over transform algorithms for RMSE.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 0.099055 0.098105 0.1295 0.091946 0.11312 0.13013 0.12457 0.1727
2 Image 2 0.11709 0.11811 0.16143 0.11133 0.13856 0.16396 0.1502 0.19364
3 Image 3 0.21717 0.16547 0.22943 0.16219 0.18126 0.19566 0.18109 0.22996
4 Image 4 0.09722 0.096792 0.13482 0.089558 0.11959 0.14055 0.13091 0.16718
5 Image 5 0.094678 0.099213 0.13197 0.091237 0.1245 0.13421 0.12162 0.14791
6 Image 6 0.15916 0.13622 0.18818 0.1316 0.15577 0.17455 0.15579 0.18822
7 Image 7 0.095864 0.10107 0.13547 0.092846 0.1216 0.14023 0.13676 0.1876
8 Image 8 0.081761 0.086256 0.11774 0.078812 0.10408 0.11926 0.12052 0.17338

Table 17.

Efficiency estimation on multi-modal image fusion over transform algorithms for SNR.

Description/images Description PCA34 DWT35 IHS36 DCT37 CWT38 NSCT36 DT-DWT31 PF-HBSSO-DT-CWT
1 Image 1 24.406 26.071 20.891 27.026 23.795 20.081 20.907 34.994
2 Image 2 27.532 29.148 22.478 29.831 26.862 20.259 18.763 38.593
3 Image 3 5.0268 10.74 4.5191 10.619 8.9618 4.7744 10.753 18.597
4 Image 4 27.943 25.164 20.919 27.042 22.847 21.512 20.537 35.092
5 Image 5 23.971 23.222 18.46 25.326 20.223 15.572 16.862 33.764
6 Image 6 17.343 22.509 15.456 22.806 20.613 14.629 20.12 31.039
7 Image 7 29.738 28.505 24.158 30.279 25.941 20.235 19.091 38.259
8 Image 8 29.974 29.675 23.437 31.718 26.734 18.896 19.753 40.532

Comparative analysis of the developed model with recent methods

The comparative analysis of the developed model is done using the recent methods in the multimodal image fusion approach shown in Table 18. This table analysis is performed by analyzing the entropy measure that helps to show the potential outcome in the designed framework. The recent methods like SFINet, CABnet, DUSMIF, and Dense-ResNet model is compared with different images. In this experimental evaluation, the developed PF-HBSSO-DT-CWT model shows 29.2, 29.13, 26.25, and 30.9% decreased than SFINet, CABnet, DUSMIF, and Dense-ResNet in terms of validating the image 6. The analysis of performance evaluation shows the developed model shows lower entropy value. Thus, it ensures to reduce the dimensionality problem that occur between the data points. Commonly, the developed model used to evaluate better data quality outcomes to make the possible predictions in the multimodal image fusion approach.

Table 18.

Performance analysis of multimodal image fusion model with recent methods.

Description/images SFINet40 CABnet41 DUSMIF42 Dense-ResNet43 PF-HBSSO-DT-CWT
1 4.8956 5.7375 4.7904 4.974 4.1346
2 4.8576 4.8947 4.1643 4.8947 3.694
3 4.8432 4.5736 4.7465 5.8624 3.7606
4 5.8946 5.6436 4.7895 5.0782 4.2493
5 3.8965 3.7541 2.6854 3.7398 2.5914
6 4.8756 4.8645 4.6748 4.9898 3.4474
7 4.6574 4.7728 4.8935 4.8936 3.8333
8 4.8365 4.8554 4.6732 4.6283 3.8149

Conclusion

A multimodal image fusion model wasrecommendedin this paper via heuristic derived transform approaches. Initially, the medical source images wereacquired from the standard public datasets, and further, decomposed was done using the ODTCWT to acquire low frequency and high frequency coefficients. Here, frequency coefficients in DTCWT were tuned with the PF-HBSSO to enhance the better fusion quality. Then, the fusion of high-frequency coefficients was performed with the adaptive weighted average fusion technique, where the weights were optimized using the same PF-HBSSO algorithm to achieve the optimal fused results. Similarly, the low-frequency coefficients were fused by average fusion. Finally, the fused images were given to inverse ODTCWT for image reconstruction. The experiments have demonstrated that the recommended multimodal image fusion method has illustrated superior efficiency than the conventional image fusion approaches. For example, the SNR of the designed PF-HBSSO-DTCWT has achieved higher rate while estimating with traditional approaches for image 8, which was 35.2, 36.5, 72.9, 27.7, 51.6, 53, and 51% maximized than PCA, DWT, IHS, DCT, CWT, NSCT, and DT-CWT. However, the proposed model has to be improved in future in terms of statistical analysis to eradicate the performance overlap with the traditional heuristic algorithms.

Acknowledgements

I would like to express my very great appreciation to the co-authors of this manuscript for their valuable and constructive suggestions during the planning and development of this research work.

Author contributions

J.R. and R.N. designed the model, computational framework and carried out the implementation. Jampani Ravi performed the calculations and wrote the manuscript with all the inputs. J.R. and R.N. discussed the results and contributed to the final manuscript.

Funding

This research did not receive any specific funding.

Data availability

The data underlying this article are available in https://www.med.harvard.edu/aanlib/home.htm.

Declarations

Competing interests

The authors declare no competing interests.

Ethical approval

Not Applicable.

Informed consent

Not Applicable.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Gao, X. W. & Hui, R. A deep learning based approach to classification of CT brain images. SAI Computing Conference (SAI). (2016)
  • 2.Yang, H., Sun, J. Li, H., Wang, L. & Xu, Z. Deep fusion net for multi-atlas segmentation: Application to cardiac mr images.MICCAI 2016: Medical Image Computing and Computer-Assisted Intervention. pp 521–528. (2016)
  • 3.Nie, D., Zhang, H., Adeli, E., Liu, L. & Shen, D. 3D deep learning for multi-modal imaging-guided survival time prediction of brain tumor patients. MICCAI 2016: Medical Image Computing and Computer-Assisted Intervention. pp 521–528. (2016) [DOI] [PMC free article] [PubMed]
  • 4.James, A. P. & Dasarathy, B. V. Medical image fusion: A survey of the state of the art. Inform. Fusion.19, 4–19 (2014). [Google Scholar]
  • 5.Li, S., Kang, X., Fang, L., Hu, J. & Yin, H. Pixel-level image fusion: A survey of the state of the art. Inform. Fusion.33, 100–112 (2017). [Google Scholar]
  • 6.James, A. P. & Dasarathy, B. A review of feature and data fusion with medical images. Computer Vision and Pattern Recognition. (2015)
  • 7.Mangai, U. G., Samanta, S., Das, S. & Chowdhury, P. R. A survey of decision fusion and feature fusion strategies for pattern classification. IETE Tech. Rev.27(4), 293–307 (2010). [Google Scholar]
  • 8.Hermessi, H., Mourali, O. & Zagrouba, E. Convolutional neural network-based multimodal image fusion via similarity learning in the shearlet domain. Neural Comput. Appl.30, 2029–2045 (2018). [Google Scholar]
  • 9.Yang, Y. et al. Multimodal medical image fusion based on fuzzy discrimination with structural patch decomposition. IEEE J. Biomed. Health Inform.23(4), 1647–1660 (2019). [DOI] [PubMed] [Google Scholar]
  • 10.Gómez-Chova, L., Tuia, D., Moser, G. & Camps-Valls, G. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE103(9), 1560–1584 (2015). [Google Scholar]
  • 11.Ioannidou, S. & Karathanassi, V. Investigation of the dual-tree complex and shift-invariant discrete wavelet transforms on quickbird image fusion. IEEE Geosci. Remote Sens. Lett.4(1), 166–170 (2007). [Google Scholar]
  • 12.Jiang, Q., Jin, X., Lee, S. & Yao, S. A novel multi-focus image fusion method based on stationary wavelet transform and local features of fuzzy sets. IEEE Access5, 20286–20302 (2017). [Google Scholar]
  • 13.Bhateja, V., Patel, H., Krishn, A., Sahu, A. & Lay-Ekuakille, A. Multimodal medical image sensor fusion framework using cascade of wavelet and contourlet transform domains. IEEE Sens. J.15(12), 6783–6790 (2015). [Google Scholar]
  • 14.Madheswari, K., Venkateswaran, N. & Sowmiya, V. Visible and thermal image fusion using curvelet transform and brain storm optimization. 2016 IEEE Region 10 Conference (TENCON). pp 2826–2829. (2016)
  • 15.Tao, J. Li, S. & Yang, B. Multimodal image fusion algorithm using dual-tree complex wavelet transform and particle swarm optimization. Communications in Computer and Information Science. vol 93, pp 296–303. (2010)
  • 16.Kumari, D. & Agwekar, A. Survey paper on image fusion using hybrid non-subsampled contourlet transform and neural network. International Conference on Intelligent Computing and Control Systems (ICICCS). pp 1564–1568. (2021)
  • 17.Zheng, S., Shi, W.-Z., Liu, J., Zhu, G.-X. & Tian, J.-W. Multisource image fusion method using support value transform. IEEE Trans. Image Process.16(7), 1831–1839 (2007). [DOI] [PubMed] [Google Scholar]
  • 18.Zuo, Q., Zhang, J. & Yang, Y. DMC-Fusion: Deep multi-cascade fusion with classifier-based feature synthesis for medical multi-modal images. IEEE J. Biomed. Health Inform.25(9), 3438–3449 (2021). [DOI] [PubMed] [Google Scholar]
  • 19.Sun, Y., Fu, Z., Sun, C., Hu, Y. & Zhang, S. Deep multimodal fusion network for semantic segmentation using remote sensing image and LiDAR data. IEEE Trans. Geosci. Remote Sens.60, 1–18 (2022). [Google Scholar]
  • 20.Fu, J., Li, W., Ouyang, A. & He, B. Multimodal biomedical image fusion method via rolling guidance filter and deep convolutional neural networks. Optik.237, 166726 (2021). [Google Scholar]
  • 21.Goyal, S., Singh, V., Rani, A. & Yadav, N. Multimodal image fusion and denoising in NSCT domain using CNN and FOTGV. Biomed. Signal Process. Control.71, 103214 (2022). [Google Scholar]
  • 22.Venkatesan, B. & Ragupathy, U. S. Integrated fusion framework using hybrid domain and deep neural network for multimodal medical images. Multidimens. Syst. Signal Process.33(3), 819–834 (2022). [Google Scholar]
  • 23.Bernal, E. A. et al. Deep temporal multimodal fusion for medical procedure monitoring using wearable sensors. IEEE Trans. Multimed.20(1), 107–118 (2018). [Google Scholar]
  • 24.Duan, J. et al. A novel GA-based optimized approach for regional multimodal medical image fusion with superpixel segmentation. IEEE Access9, 96353–96366 (2021). [Google Scholar]
  • 25.Kong, W. et al. Multimodal medical image fusion using gradient domain guided filter random walk and side window filtering in framelet domain. Inform. Sci.585, 418–440 (2022). [Google Scholar]
  • 26.Zhang, X., Liu, G., Huang, L., Ren, Q. & Bavirisetti, D. P. IVOMFuse: An image fusion method based on infrared-to-visible object mapping. Dig. Signal Process.137, 104032 (2023). [Google Scholar]
  • 27.Zhou, X. et al. Re2FAD: A differential image registration and robust image fusion method framework for power thermal anomaly detection. Optik.259, 168817 (2022). [Google Scholar]
  • 28.Gu, X. et al. Infrared-visible synthetic data from game engine for image fusion improvement. IEEE Trans. Games.16, 291–302 (2023). [Google Scholar]
  • 29.Jain, M., Singh, V. & Rani, A. A novel nature-inspired algorithm for optimization: Squirrel search algorithm. Swarm Evol. Comput.44, 148–175 (2019). [Google Scholar]
  • 30.Hashim, F. A., Houssein, E. H., Hussain, K., Mabrouk, M. S. & Al-Atabany, W. Honey Badger algorithm: New metaheuristic algorithm for solving optimization problems. Math. Comput. Simul.192, 84–110 (2022). [Google Scholar]
  • 31.Aghamaleki, J. A. & Ghorbani, A. Image fusion using dual tree discrete wavelet transform and weights optimization. Vis. Comput.39(3), 1181–1191 (2023). [Google Scholar]
  • 32.Jafari, M. & Chaleshtari, M. H. B. Using dragonfly algorithm for optimization of orthotropic infinite plates with a quasi-triangular cut-out. Euro. J. Mech. A/Solids66, 1–14 (2017). [Google Scholar]
  • 33.Mirjalili, S., Mirjalili, S. M. & Lewis, A. Grey wolf optimizer. Adv. Eng. Softw.69, 46–61 (2014). [Google Scholar]
  • 34.Qian, J. et al. Structured illumination microscopy based on principal component analysis. ELight.3(1), 4 (2023). [DOI] [PubMed] [Google Scholar]
  • 35.Trivedi, G. & Sanghavi, R. Fusesharp: A multi-image focus fusion method using discrete wavelet transform and unsharp masking. J. Appl. Math. Inform.41(5), 1115–1128 (2023). [Google Scholar]
  • 36.Hussein, Y. D., Makkey, Y. M. & Abdelrahman, A. S. Hybrid fusion approach for Alzheimer’s disease progression employing IHS and wavelet transform. Menoufia J. Electron. Eng. Res.33(1), 17–23 (2024). [Google Scholar]
  • 37.Xu, H., Ma, J., Jiang, J., Guo, X. & Ling, H. U2Fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell.44(1), 502–518 (2020). [DOI] [PubMed] [Google Scholar]
  • 38.Wang, C., Nie, R., Cao, J., Wang, X. & Zhang, Y. IGNFusion: An unsupervised information gate network for multimodal medical image fusion. IEEE J. Select. Top. Signal Process.16(4), 854–868 (2022). [Google Scholar]
  • 39.Zhou, M., Huang, J., Yan, K., Hong, D., Jia, X., Chanussot, J., & Li, C. A general spatial-frequency learning framework for multimodal image fusion. IEEE Transactions on Pattern Analysis and Machine Intelligence. (2024). [DOI] [PubMed]
  • 40.Sun, Le., Tang, M. & Muhammad, G. CABnet: A channel attention dual adversarial balancing network for multimodal image fusion. Image Vis. Comput.147, 105065 (2024). [Google Scholar]
  • 41.Lin, C., Chen, Y., Feng, S. & Huang, M. A multibranch and multiscale neural network based on semantic perception for multimodal medical image fusion. Sci. Rep.14(1), 17609 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Ghosh, T. & Jayanthi, N. An efficient Dense-Resnet for multimodal image fusion using medical image. Multimed. Tools Appl.2024(83), 68181–68208 (2024). [Google Scholar]
  • 43.Long, Y., Jia, H., Zhong, Y., Jiang, Y. & Jia, Y. RXDNFuse: A aggregated residual dense network for infrared and visible image fusion. Inform. Fusion69, 128–141 (2021). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The data underlying this article are available in https://www.med.harvard.edu/aanlib/home.htm.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES