Abstract.
Purpose
The technique of fusing or integrating medical images collected from single and various modalities is known as medical image fusion. This is done to improve the quality of the images and combine information from several medical images. The whole procedure aids medical practitioners in gaining correct information from single images. Image fusion is one of the fastest-growing research topics in the medical imaging field. Sparse modeling is a popular signal representation technique used for image fusion with different dictionary learning approaches. We propose a medical image fusion with sparse representation (SR) and block total least-square (BLOTLESS) update dictionary learning.
Approach
The domain of dictionary learning is the most significant research domain related to SR. An efficient dictionary increases the effectiveness of sparse modeling. Due to SR being an ongoing interesting research area, the medical image fusion process is done with a modified image fusion framework with recently developed BLOTLESS update dictionary learning.
Results
The experimental results are compared for the image fusion process using other state-of-the-art dictionary learning algorithms, such as simultaneous codeword optimization, method of optimal directions, and K-singular value decomposition. The effectiveness of the algorithm is evaluated based on image fusion quantitative parameters. Results show that the BLOTLESS update dictionary algorithm is a promising modification for the sparse-based image fusion with its applicability in the fusion of images related to different diseases.
Conclusions
The experiments and results show that the dictionary learning algorithm plays an important role in the sparse-based image fusion general framework. The fusion results also show that the proposed improved image fusion framework for medical images is promising compared with frameworks with other dictionary learning algorithms. As an application, it is also used as a tool for the fusion of different modularities of images related to brain tumor and glioma.
Keywords: medical image fusion, sparse representation, deep-learning algorithms, K-singular value decomposition, simultaneous codeword optimization, block total least least-square update
1. Introduction
Image fusion is the sub-domain of information fusion that combines the information from several source images into a single image called the fused image.1 The fused image is an integrated information source that allows users to acquire more details in less time and at a lower cost. The primary goal of the fusion process is to get superior results in terms of contrast, visual experience, and quality. Since the emergence of image fusion in 1985,2 its application domain have spread continuously with various input source images. In military application cameras, visible and infrared image fusion3 is employed. Multi-focus image fusion4,5 is a common feature in digital cameras. Satellite image processing, such as pan-sharpening,6,7 uses remote sensing image fusion. One of the most important applications of digital image fusion is medical image fusion.8,9
In disease diagnosis, an image plays an important role. An accurate diagnosis does not rely on a single type of clinical report. It needs a large amount of information from multi-model imaging reports.10 So the fused images generated by the image fusion process reduce the chance of error in disease diagnosis. Images from several modalities are used to extract enough information for clinical diagnosing for accurate treatment. Medical imaging is human anatomy screening for analyzing and disease diagnosis. There are many medical image techniques, such as x-ray imaging, which is considered to be the best method for identifying ruptures in bones. Cross-sectional layers are created by computed tomography (CT) and magnetic resonance imaging (MRI). Although CT scans are commonly used to examine dense structures, they are not suitable for soft tissue or physiological evaluation. On the other hand, MRI allows for a clearer view of soft tissues and is frequently used to identify tumors and other abnormalities. Both scans (MRI and CT) give the best results when a specific medical diagnosis requires precise images inside the body.11
Positron emission tomography (PET) gives information on blood flow in the body to reveal various cancers and heart diseases. In a PET scan, a radiolabel is injected into a subject’s body with the bio-distribution of the chemical being measured as a function of time to identify physiologic quantities associated with the compound. However, as compared with CT and MRI, PET scans have a lower resolution. Single-photon emission computed tomography (SPECT) discovers clinical information related to human body structure characteristics.10 Different sensors are employed in different imaging techniques to find imaging information of the same region of the body. As the applicability of medical image fusion is concerned, it is used in the detection of different diseases, such as cancer, Alzheimer’s disease, glioma, etc.12 Glioma and glioblastoma are two types of brain tumors.13 High-pixel-intensity cells and uneven border areas characterize glioma tumors. Glioblastoma tumors are low-pixel-intensity cells that may be identified with great accuracy using a variety of traditional techniques. By fusing different modularities of images, the accurate placement of the glioma tumor can be easily detected.13
Image fusion methods may be classified into three primary categories: pixel-level fusion,14,15 feature-level image fusion,16 and decision-level fusion.17 As far as the process domain of fusion is concerned, researchers have developed different techniques to fulfill image fusion tasks. These are categorized into (a) spatial domain techniques and (b) transfer domain techniques.18 A few spatial domain methods are the averaging method, select maximum/minimum method, Brovey transform method, intensity hue saturation method, and principal component analysis method. The spatial domain method directly deals with the pixel values present in images. These methods have many disadvantages, such as artifacts, noise, etc. These disadvantages lead researchers to shift to transfer domain techniques, such as discrete cosine transform, discrete wavelet transform, gradient pyramid, non-sub-sampled contourlet transform (NSCT), etc.1,2,18 Recently, researchers have developed some improved techniques for image fusion. In this direction, compressed sensing plays an important role. Sparse representation (SR)19 is one of the attractive signal processing techniques used for image fusion in which sparse coefficients are fused as a feature-level fusion. The fusion is based on SR having three stages: (a) SR of source images, (b) fusion encoding, and (c) fused image reconstruction.20
Initially, all source images are split into patches of size by the sliding window method. Then all extracted patches are arranged in lexicological order followed by SR coefficients, calculated with some pursuit algorithm21 and a predefined dictionary. A suitable fusion rule is applied to these sparse coefficients followed by reversing the process to get a single fused image. The whole process is collectively called SR-based general fusion.20 In this paper, a sparse-based fusion framework is modified and tested on medical images. This paper is divided into eight sections. Section 2 describes the literature review on digital image fusion with SR and dictionary learning. Section 3 provides a theoretical and mathematical overview of SR. Section 4 provides a thorough review of several benchmark dictionary learning techniques. Section 5 describes the proposed system of image fusion with the new dictionary algorithm and its benefits. Section 6 explains the experimental results from all three stages of the experiment and the analysis.
2. Literature Review
The intensive research continues for many years on digital image fusion, SR, and dictionary learning. For different application areas, a collaborative study on the above topics is also ongoing. Yang and Li22 were the first to propose SR multi-focus image fusion. Yu et al.23 introduced a new image fusion approach in which source images were represented using joint SRs. Yin and Li24 proposed a method of fusion called the joint sparsity model (JSM) for medical images and infrared-visible (IR-VI) images. Yang and Li25 conducted image fusion for various types of image pairs with the help of the simultaneous orthogonal matching pursuit algorithm. Yu et al.26 used SR for medical image fusion with K-singular value decomposition (K-SVD) algorithm,27 whereas Li and Yang28 first performed pan-sharpening (a form of image fusion) based on SR. Yin et al.29 conducted simultaneous image fusion and super-resolution using SR for medical and remote sensing images. Yang et al.30 proposed a color fusion scheme for infrared and low-light-level images with SR used for the fused image of a particular channel. Liu et al.,31 proposed a technique for simultaneous fusion and denoising called adaptive SR. Liu et al.32 conducted fusion of medical images through NSCT and SR in which Laplacian pyramid (LP) and high-pass (HP) coefficients were calculated by NSCT for source images followed by LP bands fused with SR. Using the absolute values of coefficients as activity level measurements, the HP bands were fused. Finally, using inverse NSCT on the combined coefficients, the fused image was constructed. Liu et al.33 proposed a multi-scale (MS) transform and SR framework. Also, Yin,34 learned an MS dictionary that was used with SR for image fusion. Zong and Qiu,35 proposed a medical image fusion method using classified image patches. Aishwarya and Thangammal36 suggest a multimodal medical image fusion based on SR and modified spatial frequency. Huang and Chu37 recently developed the medical image fusion method using guided filtering and SR. Liu et al.38 combined the LP and convolutional SR for medical image fusion to achieve good performance. Maqsood and Javed,39 proposed a fusion scheme based on two-scale image decomposition and SR. Hermessi et al.40 provided a detailed overview of medical image fusion approaches including theoretical background and current advancements in 2021.
Dictionary learning is a crucial part of the medical image fusion domain using SR. Parallel research is continuously done in the area with equal importance. The sparse-based image fusion process’s effectiveness is directly proportional to the effectiveness of the dictionary learning algorithm. An overcomplete dictionary19,27,41,42 generates sparse coefficients for a specific signal, such as an image. Overcomplete means dictionary equations are less than the total variables. Initially, some predefined dictionaries such as Fourier, wavelet, short-time Fourier transform (STFT), and Gabor transforms20,43,44 are used. Randomization is used to create the initial dictionary before the dictionary is to be trained. Over the years, this research area has grown positively, resulting in several benchmark algorithms, such as the method of optimal directions (MOD)45 and K-SVD,27,43 being developed. Many K-SVD variations, such as label consistent-KSVD,46 non-negative K-SVD,27,43,47 deep K-SVD,48 discriminative K-SVD,49 JSM-DKSVD,50 simultaneous codeword optimization (SimCo),51 and others, were designated as milestones in the previous decade. Some dictionary learning algorithms, such as clustering KSVD52 and block total least-squares (BLOTLESS) update dictionary learning,53 have been recently developed. The BLOTLESS update algorithm has only been utilized in image denoising. All of the abovementioned algorithms are used in various signal and image processing tasks in which SR is the key representation technique.
3. Sparse Representation
The term sparse implies many or mostly zero in a vector or matrix. Researchers tried to retrieve an image’s features from the beginning of the image processing era to simplify the tasks performed on images for better representation. At the same time, storage reduction for images is treated as an essential sub-domain. SR has been a solution for both concerns in a practical manner over two decades. SR is the method or tool for representing the image as a linear combination of a few numbers of elementary signals (atoms).27 Reconstruction and mapping of an image signal to the original one are done through the overcomplete dictionary.19,27,41,42
3.1. Mathematical Representation
Let us assume that all mathematical details are for a one-dimensional signal. The original signal that needs to be recovered is represented as vector . The whole system is termed in Eq. (1) as54
| (1) |
where T is a dictionary. In this case, is a spare approximation for the original signal , which is represented in Eq. (2) as
| (2) |
As per Eq. (2), the in RHS implies that the exact recovery of signal is not required if is a sparse approximation of , which is termed its solution.54
The whole under-determined system is summarized, where original signal x is recovered by an overcomplete ( means more unknowns than equations) dictionary , , and . An SR of (which is “”) with very few non-zero entries is called sparse coefficients. The approximate solution of the above under-determined system can be represented as an optimization problem as defined in Eq. (3). The norm is used for the solution; it calculates the total non-zero entries present in vector . This particular system is meant to be an NP-hard problem and needs an approximated solution by approximation techniques.55 Reseachers primarily use the OMP algorithm, which is a greedy pursuit algorithm. Its improved versions are stagewise OMP (StOMP), regularized orthogonal matching pursuit (ROMP), and compressed sensing matching pursuit (CoSaMP).27,56
| (3) |
4. Dictionary Learning
Dictionary learning27 is one of the research topics from the era of signal processing. The purpose of dictionary learning is to seek out the most straightforward training dictionary to sparse the signal. Generally, a dictionary is a two-dimensional (2D) matrix of where each column is called atoms or basis vectors. Initially, a dictionary is a predefined matrix based on signal transform functions, such as wavelet, STFT, contourlet, curvelet, etc. Later, a dictionary trained with a particular set of training signals is generally used. The input source images or a set of natural images are also used as a training set for dictionary training. Suppose that a set of some natural images is used as the training set and a dictionary T exists to represent each signal sparsely as an X vector. The elementary method of dictionary learning is shown in Eq. (4) as57
| (4) |
Here, , and denotes a column with index of vector set .
Many image processing tasks using DL, such as denoising, super-resolution, segmentation, fusion, and so on, are influenced by the efficiency of the DL process. Typically when finding sparse coefficients, dictionary training has to be simultaneously estimated. These two processes run separately, and their solutions iteratively alternate until convergence.57 Generally, dictionary learning algorithms run in two stages. The SR coefficient is computed in the first stage using approximation methods such as MP and OMP.58 The dictionary, on the other hand, is fixed during this procedure. The dictionary must be updated in the second stage.
Many algorithms are based on the above two-stage principle of dictionary learning. As far as this method is concerned, some benchmark algorithms have been proposed in the last two decades, of which the MOD45 algorithm and the K-SVD27 algorithms play a vital role. Research results estimate that KSVD is far better than the MOD algorithm. The first stage is identical for both of them. However, for the second stage, the sparse coefficient matrix is fixed in MOD.
4.1. MOD Dictionary Learning
Engan, in 2006, proposed the state-of-the-art algorithm for dictionary learning termed the “MOD.”45 As per the ideology described in the section above, MOD is also based on the same two-stage principle. The first stage is the sparse coding stage. The second stage is the dictionary update stage. The algorithm finds the sparse coefficient matrix using an approximation algorithm with dictionary T fixed in the first stage. In the second stage, dictionary T is to be updated.59 The MOD algorithms update all dictionary atoms at a time. Initially, the algorithm performance is better compared to other algorithms. But low convergence is a drawback of the algorithm.60
4.2. K-SVD Dictionary Learning
The drawback of the MOD algorithm was recovered in K-SVD.27 As the name implies, K-SVD consists of the joint property of SVD61 and the -means algorithm.62 The only difference between the K-SVD and MOD algorithms is at the second stage. In the second stage of the dictionary update, one column at a time is processed with SVD computation sequentially. In K-SVD, each iteration consists of updating a single atom at a time for dictionary updating.
4.3. SimCo Dictionary Learning
As shown in previous sections, MOD and K-SVD are two benchmark algorithms for dictionary learning. Another milestone added to the dictionary learning research was the development of the SimCo algorithm.51 Later, two versions of the same algorithms, known as primitive and regularized SimCo,63 were released. This algorithm assumed that there are norms present in the dictionary. The improved analysis SimCo was also presented by Dong et al.64
The dictionary problem in the base SimCO algorithm is expressed in Eq. (5) as
| (5) |
where .
Here, the sparse coefficient matrix is defined as shown in Eq. (5), and the dictionary is defined as . The SimCo algorithm assumes that the dictionary having norm columns and their all matrices are represented by (). The non-zero element of sparse coefficient matrix is fixed by constraint . SimCo is processed in two stages. The first stage is the same as in the conventional method. In contrast, optimization methods on manifolds62 are applied in the second stage for dictionary updating. Updating processed on columns of T takes place with unit norm constraints. The sparse coefficient matrix is updated as a modified dictionary . During the process, the positions of non-zero elements remain fixed. As the algorithm’s name implies SimCo, the whole “dictionary update” process is done such that multiple atoms of dictionary are updated with corresponding coefficients at a time.
4.4. Dictionary Learning with Block Total Least-Squares Update
The dictionary algorithm with the BLOTLESS update was proposed by Yu et al.53 The process of simultaneously updating the dictionary blocks and corresponding sparse coefficients is done in this algorithm. In this update, The bilinear nonconvex blocks update problem transfer into a linear least-square problem, which is then efficiently solved in the algorithm.53
4.4.1. Least-square solver
The training samples are generated from in , where is a square matrix or satisfies the condition of with . The whole update issue of the dictionary is represented as a bilinear inverse problem for the sparsity pattern denoted by . The goal is to find T and with constraint being nonconvex, as shown in Eq. (6),
| (6) |
The issue described in Eq. (6) can be expressed as a convex problem for an invertible unknown dictionary matrix T.
Assume ; then . So to find H and X, the goal is set as
| (7) |
or
| (8) |
In this manner, according to Eq. (8), the bilinear problem in Eq. (6) is transformed into an equivalent problem of linear least squares.
The dictionary update problem shown in Eq. (6) has unique solutions if all solutions are in the form of Eq. (9) as
| (9) |
4.4.2. Block total least squares
The dictionary update process is essential in dictionary learning. Various methods are applied for non-overcomplete and overcomplete dictionaries. In non-overcomplete, three potential methods are structure total least squares, parallel total least squares, and iterative total least squares (IterTLS).53 If the dictionary is overcomplete, the only single solution that exists comes with the situation arising when , where is the inverse of an overcomplete dictionary .53 This is a complex situation, and the abovementioned methods for non-overcomplete cannot be applied directly. Thus, as a solution, the entire dictionary is sub-divided into sets called sub-dictionaries (). The updating process is applied from to sub-dictionaries one by one with all other dictionaries and their relative coefficients being fixed. This update process takes place for all sub-dictionaries. The sub-dictionaries are counted as individual blocks. These blocks are updated one by one, as seen in the name “BLOTLESS.” Anyone can use the same approach with a non-overcomplete dictionary update. Dictionary learning based on BLOTLESS converges much faster and needs at least 1/3 less training samples than other benchmark dictionary update methods.53 In this paper, BLOTLESS-IterTLS is used in all experiments.
5. Medical Image Fusion by Sparse-Based Modified Fusion Framework
The SR-based general fusion framework20 has been used for image fusion for many years. It works for all types of input image pairs, so the word “general” is used with its name. A modified version of the SR-based general fusion framework is proposed in this paper. This modification is done using the BLOTLESS update dictionary learning algorithm as per Fig. 1. The basic algorithmic flow of the BLOTLESS update dictionary learning algorithm is also shown in Fig. 2.
Fig. 1.
Modification in the SR-based fusion framework by BLOTLESS update.
Fig. 2.
BLOTLESS update dictionary learning algorithm.
The whole process of image fusion with this framework is divided into five stages as described in the following sections.
5.1. Lexicological Arrangement of Extracted Image Patches
MRI and PET images are the input images for the experiments. Image patches are extracted in an overlapping fashion with a one-pixel step size. If a single patch size is , the total image patches generated are for input image size .20 Then, they are all arranged in lexicological ordering for further processing.
5.2. Dictionary Learning by BLOTLESS Update
The fusion method based on SR is very much affected by the quality of dictionary learning. So compromising with dictionary learning degrades the SR-based fusion method’s quality and exact image recovery condition. For this, many novel and benchmark algorithms have been proposed, some of which are shown in section 4. The framework that is presented in this paper is modified in this stage, with the BLOTLESS update algorithm being used for dictionary learning. First, it divides the dictionary into sub-dictionaries. Then, an algorithm updates the sub-dictionary and its corresponding sparse coefficients using the total least-squares approach.53
5.3. Matrix Generation with Sparse Coefficients
This stage generates sparse coefficient matrices for all input images. It is accomplished using a learned dictionary and a suitable approximation algorithm, such as MP, OMP, StOMP, focal underdetermined system solver (FOCUSS), and others.65
5.4. Sparse Coefficient Fusion
The image fusion rule66 should be applied to the sparse coefficients at this stage. Including a specific fusion rule is critical because other aspects and the efficiency have an impact on the entire fusion process efficiency. The max rule, average method rule, max-abs fusion rule, mean of coefficients fusion rule, weighted average rule, and others are examples of general fusion rules. In the last two decades, various sophisticated fusion rules have been used in addition to the generic fusion rules to improve image fusion outcomes. They are the max weighted multi-norm fusion rule, max-pooling fusion rule, etc.67,68
5.5. Final Image Reconstruction
This stage is termed the last stage of this fusion framework. The fused coefficient matrix is inversely transformed to the lexicological ordering matrix followed by recovery or reconstruction of the final single fused image. This is done through the trained dictionary generated by the BLOTLESS update algorithm.
6. Experiments
In this section, various fusion experiments with findings are shown. The source image pairs of 20 MRI-T1 and PET images are taken from the GitHub resource.69 The dataset of medical image pairs resizes to according to the experiment requirements. The second dataset for Glioma disease is taken from the Whole Brain Atlas.70
6.1. Experiment Setup and Description
The image fusion results with this modified framework by the BLOTLESS update algorithm are compared with results of fusion from other benchmark dictionary learning algorithms.51 The quality parameters for each algorithm are identical for all experiments. These parameters should be followed for fair comparison in the fusion method based on SR in this paper. About 16 various natural images are used for dictionary learning, as shown in Fig. 4. All dictionaries are of dimension with a sparsity level of five. Images are grayscale and of size . The dictionary learning process is run up to 10 iterations in two stages. In the first stage, dictionaries are learned with 16 natural images. In the second stage, dictionaries are learned with the20 pairs of PET-MRI input images, as shown in Fig. 3.
Fig. 4.
The fusion results obtained by different methods from source image pairs 1 to 10.
Fig. 3.
Input natural images used for training data for dictionary learning for BLOTLESS.
Additionally, third stage experiments are conducted for glioma disease using the proposed framework’s fusion of the MRI-T2 and SPECT-T1 image pairs and MRI-PET image pairs. Additionally, the framework is also tested on data of 38 patients for detecting a brain tumor. For all stages, dictionary learning, and sparse coefficient matrix generation, image patches of size are used. The sliding window technique is used with the step size of one pixel to extract the image patches. An Intel(R) Core™ i5-2450 CPU@2.50 GHZ with 8 GB RAM in ideal condition is used for all experiments. The investigation is accomplished on MATLAB version R2018a.
The whole experiment is conducted in three stages. All three stages differ in training images and testing images.
-
Stage 1:
For dictionary training by all four DL algorithms, 16 natural images are used, shown in Fig. 3, and 20 pairs of PET-MRI image pairs are used for testing.
-
Stage 2:
For training and testing purposes, 20 pairs of PET-MRI image pairs are used.
-
Stage 3:
For testing the proposed framework for scan images from different modularities of disease diagnosis, the data scan images as MRI, SPECT, and PET are taken from Harvard Whole Brain Atlas70 and Radiopaedia.org71 for glioma and 38 patients of brain tumors data are taken from GitHub72 having image pairs of MRI-PET data.
6.2. Performance Evaluation Metrics
In evaluating image fusion performance, performance assessment methods are critical. Four quality evaluation metrics are used in the assessment of the proposed modified framework for image fusion. For all four quality evaluation metrics, the better value shows the better performance. The ground truth or reference image is generally unavailable in medical image fusion, so evaluation is conducted with no reference image.
6.2.1. Entropy
Entropy (EN) computes the average number of bits required to quantize the intensities in the image. It is defined in Eq. (10) as:
| (10) |
Here, contains the histogram counts from the image histogram.
6.2.2. Mutual information
Mutual information (MI) is the image fusion parameter based on the intensity distribution of fused image and source image.73 The measurement of MI is based on the Kullback–Leibler74 calculation of the two random variable dependencies, which is calculated as a relationship between images plotted in a 2D histogram as
| (11) |
where is the marginal probability distribution of the A variable. is the marginal probability distribution of the B variable. is the normalized joint distributions.75
The value of is maximum when the value of variable is equals to value of variable .
Generally, image pixel values are shown as random variables, with the image’s histogram values shown as a probability distribution. The joint information typically indicates the quantity of transferred information to the fused image from the source images.
| (12) |
6.2.3. Sum of the correlations of differences
As per the basic terminology of fusion, the fused image contains collective information of all source images. So the sum of the correlations of differences (SCD)75 performance metric uses both source and their impact on the final output fused image rather than comparing the source images with a fused image.73 Thus, according to the SCD metric, the information transfer from the second source image is termed (between fused source and first source image ). This is valid for all other source images, shown as Eqs. (13) and (14). The values of and show the quantity of information transferred to the fused image from source images and , respectively. Then, the correlation values for both differences are calculated by the c(.) function in Eq. (15) as
| (13) |
| (14) |
where the c(.) function calculates the correlation between and and between and as
| (15) |
where and and are average of pixels values of and , respectively.
6.2.4. MS-SSIM
MS-SSIM76,77 is the MS extension of the SSIM parameter that provides more flexibility compared with single-scale SSIM. It calculates the structural distortions between any two images. The single-scale SSIM compares two images concerning contrast, luminance, and structure. But with the advantages, MS-SSIM calculates similarities at each scale for contrast and structure, whereas luminance similarities are calculated at coarsest scale.
| (16) |
7. Experimental Results and Discussion
7.1. Results Stage 1: BLOTLESS Update DL by 16 Natural Images
This section explores the results of twenty medical image pairs, as shown in Figs. 4 and 5. Columns 4 to 7 show output results as fused images for fusion from MOD, KSVD, SimCo, and BLOTLESS, respectively. The quantitative parametric average results for different algorithms on all of the 20 MRI-PET image pairs are shown in Table 1. The results demonstrate that image fusion using a modified framework performs much better for all image pairs. The best outcome parameter is highlighted in bold and accompanied by values in Table 1. When the BLOTLESS update algorithm updates the framework, the outcome in terms of EN parameter is highest. This means that the image fusion was done through the proposed revised framework having a higher degree of detailed information. The MI and SCD parameters represent the quantity of information transferred from source images to the final fused image. These parameters seem good when the BLOTLESS update algorithm is used for dictionary learning. The MS-SSIM quality parameter also indicates the performance of image fusion by the proposed modified framework. This shows that the structural similarity is best in fused images compared with source images when the BLOTLESS update algorithm is used for dictionary learning. All results show the higher effectiveness of a modified sparse-based general fusion framework with the BLOTLESS update dictionary learning algorithm than with other benchmark algorithms.
Fig. 5.
The fusion results obtained by different methods from source image pairs 11 to 20.
Table 1.
Average fusion results of 20 MRI-PET image pairs with 16 natural images used for DL.
| Algorithm | EN | MI | SCD | MS-SSIM |
|---|---|---|---|---|
| MOD | 4.653497 | 9.306994 | 0.547864 | 0.853018 |
| K-SVD | 4.629932 | 9.259863 | 0.534047 | 0.849764 |
| SimCo | 4.632706 | 9.265411 | 0.537765 | 0.850245 |
| BLOTLESS | 4.805529 | 9.611057 | 0.681007 | 0.853615 |
7.2. Results Stage 2: BLOTLESS Update DL by 20 Pairs of Input PET-MRI
This section investigates the results of twenty medical image pairs when the same 20 pairs of PET-MRI input images are used as training images for all four dictionary learning algorithms. When input images are used as training images, the qualitative results for a suggested updated framework with BLOTLESS update DL are significantly superior for all PET-MRI pairs, as shown in Figs. 6 and 7 for image pairs 8 and 19, respectively. Additionally, the quantitative result improves for almost all parameters, as shown in Table 2 and Figs. 8–11. The blue dotted line indicates the parametric results when 16 natural images are used as training images. The green dotted line indicates improved parametric results when input PET-MRI images are used as training images.
Fig. 6.
Fusion results for the eighth image pair after dictionary training by 20 input PET-MRI image pairs. (a) Input MRI image; (b) input PET image; (c) fused image using the MOD algorithm; (d) fused image using the KSVD algorithm; (e) fused image by the SimCo algorithm; and (f) fused image using the BLOTLESS update algorithm.
Fig. 7.
Fusion results for 19th image pair after dictionary training by 20 input PET-MRI image pairs. (a) Input MRI image; (b) input PET image; (c) fused image using the MOD algorithm; (d) fused image using the KSVD algorithm; (e) fused image by the SimCo algorithm; and (f) fused image using the BLOTLESS update algorithm.
Table 2.
Average fusion results of 20 MRI-PET image pairs with input images used for DL by input images.
| Algorithm | Entropy | MI | SCD | MS-SSIM |
|---|---|---|---|---|
| MOD | 4.839575 | 9.67915 | 0.530137 | 0.855467 |
| K-SVD | 4.833474 | 9.666947 | 0.525598 | 0.854797 |
| SimCO | 4.851943 | 9.703886 | 0.53188 | 0.856449 |
| BLOTLESS | 4.903487 | 9.806974 | 0.577226 | 0.864089 |
Fig. 8.

EN variations for both training datasets and all four DL algorithms.
Fig. 9.

MI variations for both training datasets and all four DL algorithms.
Fig. 10.

SCD variations for both training datasets and all four DL algorithms.
Fig. 11.

MS-SSIM variations for both training datasets and all four DL algorithms.
7.3. Results Stage 3: Image Fusion by the Proposed Framework on Medical Image Pairs for Brain Tumor Disease
Stage 3 experiments are done for the disease brain tumor and glioma. For glioma, data of four patients containing MRI-T2 and SPECT-T1 and MRI-PET image pairs are used. For the fifth patient, image pairs are taken from Radiopaedia.org.71,78 Two types of image pairs are taken from this source: a CT and MRI axial TI C+ image pair and a CT and axial T2 image pair. Additionally, for brain tumors, data of 38 patients containing CT-MRI image pairs are used to test the proposed fusion framework. All of the results are manually checked by a regular medical practitioner, and the yellow squares are drawn to indicate glioma or any unknown abnormality detect in fused images.
7.3.1. Details of experiments done on five glioma patients
Patient 1: For patient 1, the six pairs of MRI-T2 and SPECT-T170 are used. The results with the framework modified by the BLOTLESS update dictionary learning algorithm are shown in Fig. 12. The fusion results indicated that, for the glioma disease, the final image contains the information of both scans with tumor or abnormities shown as yellow squares.
Fig. 12.

Fusion results for the proposed framework in glioma disease for patient 1. (a)–(f) Input MRI-T2 images (six slices); (g)–(l) input SPECT-T1 images (six slices); and (m)–(r) fused images with yellow square indicating abnormality or tumor.
Patient 2: Three slice pairs of MRI-T2 and SPECT-T1 are taken for another patient. Fusion is performed on the input three pairs, and the results of fusion are shown in Fig. 13. The results indicated that, for the glioma disease, the final fused image shows scan information of both images with an indication of glioma tumor in the yellow square.
Fig. 13.

Fusion results for the proposed framework in glioma disease for patient 2. (a)–(c) Input MRI-T2 images (three slices); (d)–(f) input SPECT-T2 images (three slices); and (g)–(i) fused images with yellow square indicating glioma tumor.
Patients 3 and 4: For patient 3, the pairs of MRI and PET images from the Whole Brain Atlas are taken for glioma disease. The fusion is performed on input pairs, and the result is shown below in Figs. 14 and 15. As per the Figs. 14 and 15 results columns, the fused images contained input image information of both images with glioma tumor shown in the yellow square.
Fig. 14.

Fusion results for the proposed framework in glioma disease for patient 3. (a) Input MRI image; (b) input PET image; and (c) fused image with yellow square indicating glioma tumor.
Fig. 15.

Fusion results for the proposed framework in glioma disease for patient 4. (a) and (b) Input MRI-T2 images (two slices); (c) and (d) input PET image (two slices); and (e) and (f) fused images with yellow square indicating glioma tumor.
Patient 5: Patient 5 datasets are taken from the public dataset of radiographical images Radiopaedia.org. For this patient, two image scan pairs are taken for glioma disease. As a pair one , CT and MRI axial T1 C+ images are taken while as a pair two CT and MRI axial T2 images are taken shown in Fig. 16. The fusion is done with the proposed fusion framework with a BLOTLESS update dictionary learning algorithm. The results are shown in Fig. 16. As per the results, the fused image shows scan information of both images.
Fig. 16.

Fusion results for the proposed framework in glioma disease for patient 5. (a) and (b) Input CT image; (c) input MRI axial TI C+ image; (d) MRI axial T2 image; and (e) and (f) fused images.
7.3.2. Details of experiments done on 38 brain tumor patients
The proposed fusion framework is used to fuse the MRI-PET dataset of brain tumor patients, and the results state that the final fused images contain the information of both input medical images as shown in Fig. 19.
Fig. 19.
Eight random fusion results from 38 patients CT + MRI image pairs using the proposed framework.
All output images of experiment stage 3 are validated by several image experts, assistant professors, professional doctors, radiologists, and technologists from various medical colleges and image scan labs. This validation is processed in the form of questionnaires79,80 on output images with two parameters (i) the quality of the output fused images and (ii) the amount of information transferred from the source images to fused images for disease detection. The opinions are taken as a rating from five to one with five being excellent results and one being poor results. The results of this process are shown in Figs. 17 and 18. Here, output set number means the patient number in stage 3 experiments.
Fig. 17.

Ratings of subjective evaluations on experiment stage 3 outputs for parameter quality of the output fused image.
Fig. 18.

Ratings of subjective evaluations on experiment stage 3 outputs for parameter “amount of information transfer from source images to fused images for disease detection.
Segmentation and classification of tumor are the main processes for glioma tumors detection80,81 on individual image scans. For these processes, image specialists can use different techniques of image processings, such as edge detection filters, machine-learning (ML) models, and deep-learning (DL) techniques.82,83 However, via medical image fusion, fused images will be more informative than the separate input scans, making them more suitable for tumor segmentation and classification challenges. The high-quality fused images will deliver better automated detection results than single-input scan medical images. On the other hand, if the detection of a tumor is done manually by doctors or radiologists, image fusion techniques will deliver a more informative single image rather than an individual one.
On the other hand, if we compared the proposed method with the state-of-art image fusion techniques, the basic spatial domain and frequency fusion methods are less complex but all have the problems of artifacts, blurring, long running time, enhancing the signal-to-noise ratio, and color distortion.1,2,18 The latest techniques based on ML models and DL have better performances. In these methods, the workload is huge and the cost is expensive, which tends to address the lack of data with over-fitting. Thus, these methods need a rich dataset of training images. DL training is time-consuming with a complicated framework and necessitates a high level of computer system configuration.79,84,85
Thus, sparse-based medical image fusion techniques19 have a better performance with low-cost resource requirements and a better running time. According to their nature, they also require low memory consumption for storage.86
However, the approaches based on sparse-based fusion experienced issues with synchronization of sparsity level to loss of relevant information in the medical image recovery process and the problem of quality of the learned dictionary.87–89
8. Conclusions and Future Work
In this paper, a modified sparsity-based image fusion framework was proposed. First, 16 natural grayscale images were used to train the dictionary of dimension with a five per atom sparsity level. This process was done for all four dictionary learning algorithms. An individual medical image was divided into overlapping patches arranged in lexicological order for the fusion process. Then, a sparse coefficients matrix was generated for fusion, followed by a fusion of coefficient matrices with a specific fusion rule. The reverse process generated final fused images for all 20 pairs of MRI-T1 and PET images. Experimental results on all image pairs demonstrated that the medical image fusion using the proposed sparse-based modified image fusion framework by the BLOTLESS update dictionary learning algorithm performed best compared with fusion using trained dictionaries by other benchmark algorithms. As per the quantitative results, the proposed framework with BLOTLESS update dictionary learning algorithm performed better than the framework with other dictionary learning algorithms.
Additionally, most of the parameters improved when the input images were used for dictionary learning. Also, improvements in qualitative results were seen in the proposed modified framework. In the third stage, the applicability of the proposed framework was tested for medical images in the case of brain tumor and glioma disease. Output images of stage 3 were also validated by multiple experts and professional doctors for glioma disease. Future research is required for comparing the proposed method with other image fusion standard algorithms as quantitative measures. Additionally, the effectiveness of the algorithm and proposed fusion framework may be compared with other medical image fusion methods, and for glioma detection, the outputs of experiments will be further segmented and classified.
Acknowledgments
This research would not have been possible without the support of the authors’ colleagues. The authors would like to thank the faculty of the Information Technology Department for reviewing this manuscript and providing essential suggestion for the paper’s final draft. We are grateful to all the medical image experts, assistant professors, doctors, and image technologists from several renowned medical colleges and hospitals, such as Radiology Department, Sawai Man Singh Medical College Jaipur, LNMC Bhopal, JLN Medical College Ajmer, PHC Thikariya, Sikar, Director Parsadi Lal Memorial Hospital, Nadbai, Bharatpur, Bhausaheb Hire Govt Medical College, Dhule, CT/MRI Center Sms Hospital Jaipur, Madhya Pradesh Medical Science University, Jabalpur, FML hospital, etc.
Biographies
Lalit Kumar Saini is a research scholar at Manipal University Jaipur, Jaipur, India. He received his BE and MTech degrees in computer engineering from the University of Rajasthan and RTU, Kota, respectively. His current research interests include digital image fusion, SR, dictionary learning, and compressed sensing algorithms.
Pratistha Mathur is a professor in the Department of Information Technology at Manipal University Jaipur. She received her MTech degee and PhD in computer science from Banasthali Vidyapith in 1998 and 2012, respectively. Her research interests include digital image processing, soft computing, and ML. She has published more than 40 papers in international and national journals and conferences.
Disclosures
The authors have no relevant financial interests in the paper and no other potential conflicts of interest to disclose.
Contributor Information
Lalit Kumar Saini, Email: lalitkumar20april@hotmail.com.
Pratistha Mathur, Email: pratistha.mathur@jaipur.manipal.edu.
References
- 1.Kaur H., Koundal D., Kadyan V., “Image fusion techniques: a survey,” Arch. Comput. Methods Eng. 28, 4425–4447 (2021). 10.1007/s11831-021-09540-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Huang B., et al. , “A review of multimodal medical image fusion techniques,” Comput. Math. Methods Med. 2020, 8279342 (2020). 10.1155/2020/8279342 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Zuo Y., et al. , “Airborne infrared and visible image fusion combined with region segmentation,” Sensors 17(5), 1127 (2017). 10.3390/s17051127 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Liu Y., et al. , “Multi-focus image fusion: a survey of the state of the art,” Inf. Fusion 64, 71–91 (2020). 10.1016/j.inffus.2020.06.013 [DOI] [Google Scholar]
- 5.Zhang X., “Multi-focus image fusion: a benchmark” (2020).
- 6.Fonseca L., et al. , “Image fusion for remote sensing applications,” in Image Fusion and Its Applications (2011). [Google Scholar]
- 7.Yusuf Y., et al. , “Assessment of pan-sharpening methods applied to image fusion of remotely sensed multi-band data,” Int. J. Appl. Earth Obs. Geoinf. 18, 165–175 (2012). 10.1016/j.jag.2012.01.013 [DOI] [Google Scholar]
- 8.James A. P., Dasarathy B. V., “Medical image fusion: a survey of the state of the art,” Inf. Fusion 19, 4–19 (2014). 10.1016/j.inffus.2013.12.002 [DOI] [Google Scholar]
- 9.Dogra A., et al. , “Medical image fusion: a brief introduction,” Biomed. Pharmacol. J. 11(3), 1209–1214 (2018). 10.13005/bpj/1482 [DOI] [Google Scholar]
- 10.Lovati A. B., et al. , “A comparative study of diagnostic and imaging techniques for osteoarthritis of the trapezium,” Rheumatology 54(1), 96–103 (2015). 10.1093/rheumatology/keu280 [DOI] [PubMed] [Google Scholar]
- 11.Javed U., et al. , “MRI and PET image fusion using fuzzy logic and image local features,” ScientificWorldJournal 2014, 1–8 (2014). 10.1155/2014/708075 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Tan W., et al. , “Multimodal medical image fusion algorithm in the era of big data,” Neural Comput. Appl., 1–21 (2020). 10.1007/s00521-020-05173-2 [DOI] [Google Scholar]
- 13.Selvapandian A., Manivannan K., “Fusion based Glioma brain tumor detection and segmentation using ANFIS classification,” Comput. Methods Programs Biomed. 166, 33–38 (2018). 10.1016/j.cmpb.2018.09.006 [DOI] [PubMed] [Google Scholar]
- 14.Liu Y., et al. , “Deep learning for pixel-level image fusion: recent advances and future prospects,” Inf. Fusion 42, 158–173 (2018). 10.1016/j.inffus.2017.10.007 [DOI] [Google Scholar]
- 15.Li S., et al. , “Pixel-level image fusion: a survey of the state of the art,” Inf. Fusion 33, 100–112 (2017). 10.1016/j.inffus.2016.05.004 [DOI] [Google Scholar]
- 16.Xiao G., et al. , “Feature-level image fusion,” in Image Fusion, Springer, Singapore: (2020). [Google Scholar]
- 17.Wang Y., Chen W., Mao S., “Multi-sensor decision level image fusion based on fuzzy theory and unsupervised FCM,” Proc. SPIE 6200, 62000J (2006). 10.1117/12.681719 [DOI] [Google Scholar]
- 18.Salem Y. B., et al. , “Image fusion models and techniques at pixel level,” in Int. Image Process., Appl. and Syst., Hammamet, Tunisia, pp. 1–5 (2016). 10.1109/IPAS.2016.7880115 [DOI] [Google Scholar]
- 19.Zhang Z., et al. , “A survey of sparse representation: algorithms and applications,” IEEE Access 3, 490–530 (2015). 10.1109/ACCESS.2015.2430359 [DOI] [Google Scholar]
- 20.Saini L. K., Mathur P., “Analysis of dictionary learning algorithms for image fusion using sparse representation,” in Second Int. Conf. Inventive Res. Comput. Appl., pp. 682–89 (2020). 10.1109/ICIRCA48905.2020.9183248 [DOI] [Google Scholar]
- 21.Zhang H., et al. , “A matching pursuit algorithm for backtracking regularization based on energy sorting,” Symmetry 12(2), 231 (2020). 10.3390/sym12020231 [DOI] [Google Scholar]
- 22.Yang B., Li S., “Multifocus image fusion and restoration with sparse representation,” IEEE Trans. Instrum. Meas. 59(4), 884–892 (2010). 10.1109/TIM.2009.2026612 [DOI] [Google Scholar]
- 23.Yu N., et al. , “Image features extraction and fusion based on joint sparse representation,” IEEE J. Sel. Top. Signal Process. 5(5), 1074–1082 (2011). 10.1109/JSTSP.2011.2112332 [DOI] [Google Scholar]
- 24.Yin H., Li S., “Multimodal image fusion with joint sparsity model,” Opt. Eng. 50(6), 067007 (2011). 10.1117/1.3584840 [DOI] [Google Scholar]
- 25.Yang B., Li S., “Pixel-level image fusion with simultaneous orthogonal matching pursuit,” Information Fusion 13(1), 10–19 (2012). 10.1016/j.inffus.2010.04.001 [DOI] [Google Scholar]
- 26.Yu N.-N., et al. , “Medical image fusion based on sparse representation with KSVD,” in World Congr. Med. Phys. and Biomed. Eng. May 26–31, M. Long, Ed., Springer, Beijing, China, pp. 550–53 (2013). [Google Scholar]
- 27.Aharon M., et al. , “K-SVD: an algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Trans. Signal Process. 54(11), 4311–4322 (2006). 10.1109/TSP.2006.881199 [DOI] [Google Scholar]
- 28.Li S., Yang B., “A new pan-sharpening method using a compressed sensing technique,” IEEE Trans. Geosci. Remote Sens. 49(2), 738–746 (2011). 10.1109/TGRS.2010.2067219 [DOI] [Google Scholar]
- 29.Yin H., et al. , “Simultaneous image fusion and super-resolution using sparse representation,” Inf. Fusion 14(3), 229–240 (2013). 10.1016/j.inffus.2012.01.008 [DOI] [Google Scholar]
- 30.Yang S., et al. , “Color fusion method for low-light-level and infrared images in night vision,” in 5th Int. Congr. Image and Signal Process., pp. 534–37 (2012). 10.1109/CISP.2012.6469757 [DOI] [Google Scholar]
- 31.Liu Q., et al. , “Adaptive dictionary learning in sparse gradient domain for image recovery,” IEEE Trans. Image Process. 22(12), 4652–4663 (2013). 10.1109/TIP.2013.2277798 [DOI] [PubMed] [Google Scholar]
- 32.Liu Y., et al. , “Medical image fusion by combining non-subsampled contourlet transform and sparse representation,” Pattern Recognit. 484, 372–81 (2014). 10.1007/978-3-662-45643-9_39 [DOI] [Google Scholar]
- 33.Liu Y., et al. , “A general framework for image fusion based on multi-scale transform and sparse representation,” Inf. Fusion 24, 147–164 (2015). 10.1016/j.inffus.2014.09.004 [DOI] [Google Scholar]
- 34.Yin H., “Sparse representation with learned multiscale dictionary for image fusion,” Neurocomputing 148, 600–610 (2015). 10.1016/j.neucom.2014.07.003 [DOI] [Google Scholar]
- 35.Zong J.-J., Qiu T.-S., “Medical image fusion based on sparse representation of classified image patches,” Biomed. Signal Process. Control 34, 195–205 (2017). 10.1016/j.bspc.2017.02.005 [DOI] [Google Scholar]
- 36.Aishwarya N., Thangammal C. B., “A novel multimodal medical image fusion using sparse representation and modified spatial frequency,” Int. J. Imaging Syst. Technol. 28(3), 175–185 (2018). 10.1002/ima.22268 [DOI] [Google Scholar]
- 37.Huang H., Chu G., “Medical image fusion based on sparse representation and guided filtering,” IOP Conf. Ser. J. Phys. 1302, 022045 (2019). 10.1088/1742-6596/1302/2/022045 [DOI] [Google Scholar]
- 38.Liu F., et al. , “Medical image fusion method by using Laplacian pyramid and convolutional sparse representation,” Concurrency Comput. Pract. Experience 32(17), e5632 (2020). 10.1002/cpe.5632 [DOI] [Google Scholar]
- 39.Maqsood S., Javed U., “Multi-modal medical image fusion based on two-scale image decomposition and sparse representation,” Biomed. Signal Process. Control 57, 101810 (2020). 10.1016/j.bspc.2019.101810 [DOI] [Google Scholar]
- 40.Hermessi H., Mourali O., Zagrouba E., “Multimodal medical image fusion review: theoretical background and recent advances,” Signal Process. Online 183, 108036 (2021). 10.1016/j.sigpro.2021.108036 [DOI] [Google Scholar]
- 41.Guha T., et al. , “Sparse representation-based image quality assessment,” Signal Process. Image Commun. 29(10), 1138–1148 (2014). 10.1016/j.image.2014.09.010 [DOI] [Google Scholar]
- 42.Wright J., et al. , “Sparse representation for computer vision and pattern recognition,” Proc. IEEE 98(6), 1031–1044 (2010). 10.1109/JPROC.2010.2044470 [DOI] [Google Scholar]
- 43.Rubinstein R., Peleg T., Elad M., “Analysis K-SVD: a dictionary-learning algorithm for the analysis sparse model,” IEEE Trans. Signal Process. 61(3), 661–677 (2013). 10.1109/TSP.2012.2226445 [DOI] [Google Scholar]
- 44.Mallat S. G., Zhang Z., “Matching pursuits with time-frequency dictionaries,” IEEE Trans. Signal Process. 41, 3397–3415 (1993). 10.1109/78.258082 [DOI] [Google Scholar]
- 45.Engan K., et al. , “Method of optimal directions for frame design,” IEEE Int. Conf. Acoust., Speech, and Signal Process. Proc. (Cat. No.99CH36258), Vol. 5, pp. 2443–46 (1999). 10.1109/ICASSP.1999.760624 [DOI] [Google Scholar]
- 46.Jiang Z., et al. , “Label consistent K-SVD: learning a discriminative dictionary for recognition,” IEEE Trans. Pattern Anal. Mach. Intell. 35(11), 2651–2664 (2013). 10.1109/TPAMI.2013.88 [DOI] [PubMed] [Google Scholar]
- 47.Aharon M., et al. , “K-SVD and its non-negative variant for dictionary design,” Proc. SPIE 5914, 591411 (2005). 10.1117/12.613878 [DOI] [Google Scholar]
- 48.Scetbon M., et al. , “Deep K-SVD denoising,” IEEE Trans. Image Process. 30, 5944–5955 (2021). 10.1109/TIP.2021.3090531 [DOI] [PubMed] [Google Scholar]
- 49.Zhang Q., Li B., “Discriminative K-SVD for dictionary learning in face recognition,” IEEE Comput. Soc. Conf. Comput. Vision and Pattern Recognit., pp. 2691–98 (2010). 10.1109/CVPR.2010.5539989 [DOI] [Google Scholar]
- 50.Wang Z., et al. , “Joint sparse model-based discriminative K-SVD for hyperspectral image classification,” Signal Process. 133, 144–155 (2017). 10.1016/j.sigpro.2016.10.022 [DOI] [Google Scholar]
- 51.Dai W., Xu T., Wang W., “Simultaneous codeword optimisation (SimCO) for dictionary update and learning,” IEEE Trans. Signal Process. 60(12), 6340–6353 (2012). 10.1109/TSP.2012.2215026 [DOI] [Google Scholar]
- 52.Fu J., et al. , “Clustering K-SVD for sparse representation of images,” EURASIP J. Adv. Signal Process. 2019(1), 47 (2019). 10.1186/s13634-019-0650-4 [DOI] [Google Scholar]
- 53.Yu Q., et al. , “Dictionary learning With BLOTLESS update,” IEEE Trans. Signal Process. 68, 1635–1645 (2020). 10.1109/TSP.2020.2971948 [DOI] [Google Scholar]
- 54.Breen P. “Algorithms for sparse approximation,” Year 4 Project Report, School of Mathematics, University of Edinburgh; (2009). [Google Scholar]
- 55.Elad M., Sparse and Redundant Representations: From Theory to Applications in Signal and Image Processing, Springer-Verlag; (2010). [Google Scholar]
- 56.https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.479.9604&rep=rep1&type=pdf
- 57.Tong T., et al. , “Dictionary learning for medical image denoising, reconstruction, and segmentation,” in Machine Learning and Medical Imaging, Wu G., et al., Eds., Academic Press, pp. 153–181 (2016). [Google Scholar]
- 58.Chen S. S., Donoho D. L., Saunders M. A., “Atomic decomposition by basis pursuit,” SIAM Journal on Scientific Computing 20(1), 33–61 (1998). 10.1137/S1064827596304010 [DOI] [Google Scholar]
- 59.Zuo Y., et al. , “Airborne infrared and visible image fusion combined with region segmentation,” Sensors 17, 1127 (2017). 10.3390/s17051127 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Cai S., et al. , “A dictionary-learning algorithm based on method of optimal directions and approximate K-SVD,” in 35th Chin. Control Conf. (2016). [Google Scholar]
- 61.Nasir H., Stanković V., Marshall S., “Singular value decomposition based fusion for super-resolution image reconstruction,” Signal Process. Image Commun. 27, 180–191 (2012). 10.1016/j.image.2011.12.002 [DOI] [Google Scholar]
- 62.Blömer J., et al. , “Theoretical analysis of the -means algorithm—a survey,” Lect. Notes Comput. Sci. 9220, 81–116 (2016). 10.1007/978-3-319-49487-6_3 [DOI] [Google Scholar]
- 63.Dai W., Xu T., Wang W., “Dictionary learning and update based on simultaneous codeword optimization (SimCO),” in IEEE Int. Conf. Acoust., Speech and Signal Process., IEEE, pp. 2037–2040 (2012). 10.1109/ICASSP.2012.6288309 [DOI] [Google Scholar]
- 64.Dong J., Wang W., Dai W., “Analysis SimCO: a new algorithm for analysis dictionary learning,” in IEEE Int. Conf. Acoust., Speech and Signal Process., Florence, Italy, pp. 7193–7197 (2014). 10.1109/ICASSP.2014.6854996 [DOI] [Google Scholar]
- 65.Yang B., Li S., “Multifocus image fusion and restoration with sparse representation,” IEEE Trans. Instrum. Meas. 59, 884–892 (2010). 10.1109/TIM.2009.2026612 [DOI] [Google Scholar]
- 66.Gorodnitsky I. F., et al. , “Neuromagnetic source imaging with FOCUSS: a recursive weighted minimum norm algorithm,” Electroencephalogr. Clin. Neurophysiol. 95(4), 231–251 (1995). 10.1016/0013-4694(95)00107-A [DOI] [PubMed] [Google Scholar]
- 67.Lewis J. J., et al. , “Pixel- and region-based image fusion with complex wavelets,” Inf. Fusion 8(2), 119–130 (2007). 10.1016/j.inffus.2005.09.006 [DOI] [Google Scholar]
- 68.Indhumathi R., et al. , “Impact of average, maximum and energy fusion rules on the performance of DWT, SWT & NSCT based fusion on bio-medical images,” J. Chem. Pharmaceut. Res. 10(8), 37–42 (2018). [Google Scholar]
- 69.Xu H., “Hanna-Xu/DDcGAN,” 2021, GitHub, https://github.com/hanna-xu/DDcGAN.
- 70.“The Whole-Brain Atlas 2021,” med.harvard, http://www.med.harvard.edu/aanlib/home.html (accessed 8 October 2021).
- 71.Chandrashekar L., Sreedevi A., “Framework for multimodal image fusion for detection of glioblastoma,” in IEEE 17th India Council Int. Conf., pp. 1–6 (2020). 10.1109/INDICON49873.2020.9342422 [DOI] [Google Scholar]
- 72.GitHub, “Multimodal medical image fusion to detect brain tumors,” dataset, 2022, https://github.com/ashna111/multimodal-image-fusion-to-detect-brain-tumors/tree/master/dataset (accessed 24 January 2022).
- 73.Liu Z., et al. , “A feature-based metric for the quantitative evaluation of pixel-level image fusion,” Comput. Vis. Image Underst. 109(1), 56–68 (2008). 10.1016/j.cviu.2007.04.003 [DOI] [Google Scholar]
- 74.Kullback S., “Information theory and statistics,” in Elements of Information Theory, Cover T. M., Thomas J. A., Eds., pp. 347–408, John Wiley and Sons, New York: (1959). [Google Scholar]
- 75.Aslantas V., Bendes E., “A new image quality metric for image fusion: The sum of the correlations of differences,” AEU Int. J. Electron. Commun. 69(12), 1890–1896 (2015). 10.1016/j.aeue.2015.09.004 [DOI] [Google Scholar]
- 76.Wang Z., et al. , “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process. 13(4), 600–612 (2004). 10.1109/TIP.2003.819861 [DOI] [PubMed] [Google Scholar]
- 77.Wang Z., et al. , “Multi-scale structural similarity for image quality assessment,” in Thirty-Seventh Asilomar Conf. Signals, Syst. Comput., Vol. 2, pp. 1398–1402 (2003). 10.1109/ACSSC.2003.1292216 [DOI] [Google Scholar]
- 78.Gaillard F., “Glioblastoma NOS,” Case Study, Radiopaedia.org (accessed 24 January 2022).
- 79.Mlynarski P., et al. , “Deep learning with mixed supervision for brain tumor segmentation,” J. Med. Imaging 6(3), 034002 (2019). 10.1117/1.JMI.6.3.034002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Petrović V., “Subjective tests for image fusion evaluation and objective metric validation,” Inf. Fusion 8(2), 208–216 (2007). 10.1016/j.inffus.2005.05.001 [DOI] [Google Scholar]
- 81.Yousif A., Omar Z. B., Sheikh U. U., “A survey on multi-scale medical images fusion techniques: brain diseases,” J. Biomed. Eng. Med. Imaging 7(1), 18–38 (2020). 10.14738/jbemi.71.7415 [DOI] [Google Scholar]
- 82.Bakas S., et al. , “Overall survival prediction in glioblastoma patients using structural magnetic resonance imaging (MRI): advanced radiomic features may compensate for lack of advanced MRI modalities,” J. Med. Imaging 7(3), 031505 (2020). 10.1117/1.JMI.7.3.031505 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Singh A., Chitalia R., Kontos D., “Radiogenomics in brain, breast, and lung cancer: opportunities and challenges,” J. Med. Imaging 8(3), 031907 (2021). 10.1117/1.JMI.8.3.031907 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Kontos D., Summers R. M., Giger M. L., “Special section guest editorial: radiomics and deep learning,” J. Med. Imaging 4(4), 041301 (2018). 10.1117/1.JMI.4.4.041301 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Peddinti A. S., Maloji S., Manepalli K., “Evolution in diagnosis and detection of brain tumor—review,” J. Phys. Conf. Ser. 2115(1), 012039 (2021). 10.1088/1742-6596/2115/1/012039 [DOI] [Google Scholar]
- 86.Rebollo-Neira L., “A competitive scheme for storing sparse representation of x-ray medical images,” PLoS One 13(8), e0201455 (2018). 10.1371/journal.pone.0201455 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Zhang W., et al. , “Overview of multi-modal brain tumor MR image segmentation,” Healthcare 9(8), 1051 (2021). 10.3390/healthcare9081051 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Faragallah O. S., et al. , “A comprehensive survey analysis for present solutions of medical image fusion and future directions,” IEEE Access 9, 11358–11371 (2021). 10.1109/ACCESS.2020.3048315 [DOI] [Google Scholar]
- 89.Li B., et al. , “Multi-sensor image fusion: a survey of the state of the art,” J. Comput. Commun. 9(6), 73–108 (2021). 10.4236/jcc.2021.96005 [DOI] [Google Scholar]








