Skip to main content
Entropy logoLink to Entropy
. 2021 Feb 19;23(2):239. doi: 10.3390/e23020239

Advanced Driving Assistance Based on the Fusion of Infrared and Visible Images

Yansong Gu 1, Xinya Wang 2,*, Can Zhang 2, Baiyang Li 1
Editor: Raúl Alcaraz
PMCID: PMC7922155  PMID: 33669599

Abstract

Obtaining key and rich visual information under sophisticated road conditions is one of the key requirements for advanced driving assistance. In this paper, a newfangled end-to-end model is proposed for advanced driving assistance based on the fusion of infrared and visible images, termed as FusionADA. In our model, we are committed to extracting and fusing the optimal texture details and salient thermal targets from the source images. To achieve this goal, our model constitutes an adversarial framework between the generator and the discriminator. Specifically, the generator aims to generate a fused image with basic intensity information together with the optimal texture details from source images, while the discriminator aims to force the fused image to restore the salient thermal targets from the source infrared image. In addition, our FusionADA is a fully end-to-end model, solving the issues of manually designing complicated activity level measurements and fusion rules existing in traditional methods. Qualitative and quantitative experiments on publicly available datasets RoadScene and TNO demonstrate the superiority of our FusionADA over the state-of-the-art approaches.

Keywords: advanced driving assistance, infrared and visible image fusion, smart city, generative adversarial network

1. Introduction

Smart cities have become new hot spots for global city development, including smart transportation, smart security, smart communities, and so on. Among them, advanced driving assistance is an indispensable and effective tool playing a pivotal role in smart transportation. The core of a smart city is a high degree of information fusion, so as advanced driving assistance. In the advanced driving assistance scene, there are a large number of information sensing devices to monitor, connect and interact with objects and pedestrians in the environment online [1]. Among the sensors, infrared and visible sensors are generally the most widely used types of sensors whose wavelengths are 300–530 nm and 8–14 μm, respectively.

The peculiarity of combining infrared and visible sensors depends on the fact that visible image captures reflected light to represent abundant texture details, while infrared image captures thermal radiation, which can emphasize thermal infrared targets though in poor lighting conditions or under the severe occlusion [2,3,4]. Based on the strong complementarity between infrared and visible sensors, the fused results can show abundant texture details with salient thermal targets. Therefore, infrared and visible image fusion is undoubtedly a significant and effective application in advanced driving assistance, which is much more beneficial for automatic detection of the system or driver’s visual perception.

In the infrared and visible images fusion, many methods have been proposed in the past few years, and they can be divided into six categories according to corresponding schemes, including pyramid methods [5,6], neural network-based methods [7], wavelet transformation based methods [8], sparse representation methods [9,10], salient feature methods [11,12], and other methods [13]. There are three main parts in these fusion methods, i.e., (i) domain transform, (ii) activity level measurement, and (iii) fusion rule design. The biggest criticism lies in that designing complex activity level measurements and fusion rules manually are usually needed in most existing methods, which leads to additional time consumption and complexity.

The development of the smart city is inseparable from the empowerment of artificial intelligence (AI). Among them, the powerful feature extraction capabilities of deep learning have caught more and more eyes [14,15]. Some detailed exposition about these fusion methods will be discussed later in Section 2.2. These deep learning-based methods have found a new breakthrough for image fusion and also achieved excellent effects. However, this kind of method does not completely break away from the shackles of traditional methods, because the framework based on deep learning is typically only applied to some small parts, e.g., the extraction of features, while the whole fusion process is still based on traditional frameworks.

In addition, both traditional and deep learning-based methods suffer from a common predicament, i.e., information attenuation. Specifically, the extracted (or to be fused) information, including texture details and salient thermal targets, are attenuated to varying degrees due to the weight selection accompanying the fusion process.

To address the above issues and improve the performance of advanced driving assistance, in this paper, we propose a new fusion method that is fully based on deep learning, called FusionADA. For convenience, we abbreviate source visible and infrared images, and fused image as VI, IR and IF, respectively. First of all, in our fusion model, deep learning runs through the whole model, and manually designing complex activity level measurements and fusion rules are not required, thus our FusionADA is a fully end-to-end model. Furthermore, our FusionADA can overcome the predicament of information attenuation, which is reflected in texture details and salient thermal targets, respectively. On the one hand, since the texture details can be characterized by gradient variation, based on the major intensity information, we employ the max-gradient loss to guide the fused image to learn the optimal texture details from source images. On the other hand, with a labeled mask reflecting the domains of salient thermal targets, we establish a specific adversarial framework of two kinds of neural networks, i.e., the generator and the discriminator, based on conditional generative adversarial networks (GAN). Rather than a whole image, the real data only refers to the salient thermal targets from the source infrared image limited by the labeled mask (M), i.e., IRM, while the fake date refers to the corresponding regions of the fused image, i.e., IFM, which forces the fused image to restore the salient thermal targets from the source infrared image. In conclusion, our FusionADA can be trained to generate the fused image with the optimal texture details and salient thermal targets in a fully end-to-end way without information attenuation.

The main contributions of this paper can be summarized into two aspects as follows. (i) In order to improve the performance of the advanced driving assistance, we propose a new fully end-to-end infrared and visible images fusion method, which is achieved without any manual designs of complex activity level measurements and fusion rules. (ii) To overcome the predicament of information attenuation, we employ the max-gradient loss and adversarial learning to learn the optimal texture details and restore the salient thermal targets, respectively.

The rest of this paper is arranged as follows: In Section 2, we present some related works with a conspectus of explanations of the advanced driving assistance and existing deep learning-based fusion methods. The detailed introduction of our FusionADA with the motivation is presented in Section 3. Section 4 shows the fusion performance of our FusionADA on public infrared and visible image fusion datasets RoadScene and TNO, compared with other state-of-the-art methods in terms of both qualitative visual effect and quantitative metrics. Besides, we carry out the ablation experiment of adversarial learning in this section, followed by some conclusions in Section 5.

2. Related Work

In this section, we provide brief explanations of the advanced driving assistance in smart transportation and deep learning-based fusion methods.

2.1. Advanced Driving Assistance

Advanced driving assistance refers to a kind of integrated system that integrates a camera detection module, a communication module and a control module, which is of great benefit for vehicle driving tasks. Specifically, there are different operating principles and levels of assistance to the drivers. The advanced driving assistance can be divided into different classes according to the monitored environment, and the used sensors [16]. These systems will not act completely autonomously, they will only provide relevant information to drivers and assist them when taking key actions. The proposed infrared and visible images fusion method relies on exteroceptive sensors, and the information of fused results is shown on a screen as the visual assistance to the drivers, which can be also incorporated in automatic recognition by smart transportation.

2.2. Infrared and Visible Image Fusion Based on Deep Learning

In the last several years, the breakthroughs in deep learning have driven the vigorous development of artificial intelligence, which also provides new ideas for infrared and visible image fusion. Fusion methods based on deep learning can be roughly divided into two categories: convolutional neural networks (CNN)-based model and GAN-based model [17]. In the methods based on CNN, Liu et al. [18] firstly established a deep convolutional neural network to achieve the generations of both activity level measurement and fusion rule, which are also applied for fusing infrared and visible images. Innovatively, Li et al. [19] used the architecture of dense block to get more useful features from source images in the encoding process, followed by a decoder to reconstruct the fused image. Besides, a novel convolution sparse representation was introduced by Liu et al. [20] for image fusion, where a hierarchy of layers was built by deconvolutional networks. As for the methods based on GAN, Ma et al. [21,22] proposed the FusionGAN to fuse infrared and visible images by adversarial learning, which is also the first time that the GANs are adopted for addressing the image fusion task. Xu et al. [23] achieved fusion via a conditional generative adversarial network with dual discriminators (DDcGAN), in which a generator accompanied by two discriminators is employed to enhance the functional information in IR and texture details in VI.

3. Proposed Method

In this section, combining the characteristics of infrared and visible images and the fusion target, we give a detailed introduction to the proposed method, including our fusion formulation, the network architectures of generator and discriminator, and the definitions and formulations of loss functions.

3.1. Fusion Formulation

The training procedure of our proposed FusionADA is illustrated in Figure 1. The infrared images can distinguish the targets from their background based on the dissimilarity in thermal radiation, but they lack rich texture details. In contrast, the visible images are able to show relatively richer texture details with high spatial resolution, but they fail to highlight the salient targets. Besides, for a certain area corresponding to the two source images, the infrared or visible image may own better texture details. Given an infrared image IR and a visible image VI, the ultimate goal of our FusionADA is to learn a fuse-generator G conditioned on them constrained by a content loss. With the labeled mask M reflecting the domain of salient thermal targets, the fused image IF multiplied by the labeled mask M, i.e., IFM is encouraged to be realistic enough and close enough to real data, i.e., IRM, to fool the discriminator D. Meanwhile, the discriminator aims to distinguish the fake data (IFM) from the real data (IRM). Accordingly, the objective function of adversarial learning can be formulated as follows:

minGmaxDE[logD(IRM)]+E[log(1D(IFM))]. (1)

Figure 1.

Figure 1

The training procedure of FusionADA.

After the continuous optimization of the generator and the adversarial learning of the generator and the discriminator, the fused image will finally possess the optimal texture details and salient thermal targets in a fully end-to-end way.

3.2. Network Architecture

Fuse-Generator G: As shown in Figure 1, the Fuse-Generator can be regarded as an En-decoder structure. In the encoder, for each image, we use a branch to extract information from it. Adopting the idea of DenseNet [19], each layer is directly connected with other layers in a feed-forward manner. Since the information extracted from each source image is not the same, the internal parameters of each branch are also different. There are four convolutional layers in each branch, and each convolutional layer consists of the operations of padding and convolution, and the corresponding activation function, i.e., leaky rectified linear unit (LReLU). In order to avoid the blurring of the image edges caused by “SAME”, the padding mode of all convolution layers is set as “VALID”. The additional padding operation placed before convolution is employed to keep the size of feature maps unchanged and match the size of source images. The kernel sizes of the first two convolutional layers are set to 5, while the kernel sizes of the latter two convolutional layers are set to 3. The strides of all convolutional layers are set to 1. Since the number of output feature maps of each convolutional layer is 16, the number of the final concatenated output feature maps is 128.

The decoder is used for channel reduction and fusion of the extracted information. The kernel sizes in all convolutional layers are uniformly set to 1 with the strides setting to 1, and thus the sizes of feature maps will not change. Therefore, there are no padding operations. The activation function of the last convolutional layer is set as Tanh. Moreover, the specific settings for the number of output channels in all layers are summarized in Table 1.

Table 1.

Input/output channels of all convolutional layers.

Number of Input Channels Number of Output Channels
Encoder Convolutional layer 1 of branch 1/2 1 16
Convolutional layer 2 of branch 1/2 16 16
Convolutional layer 3 of branch 1/2 32 16
Convolutional layer 4 of branch 1/2 48 16
concatenation 128
Decoder Convolutional layer 1 128 64
Convolutional layer 2 64 32
Convolutional layer 3 32 1

Discriminator D: The discriminator is added to form an adversarial relationship with the generator. The input of the discriminator is the real data, i.e., IRM, or the fake data, IFM, and the output is the scalar estimating the probability of the discriminator’s input image from real data rather than fake data. There are only three convolutional layers in the discriminator, which is much simpler compared to the Fuse-Generator. The strides of all convolutional layers are set to 2. After the fully connected layer, the scaler is obtained by the activation function Tanh.

3.3. Loss Functions

The loss functions in our work are composed of the loss of Fuse-Generator LG and the loss of discriminator LD.

3.3.1. Fuse-Generator Loss LG

The Fuse-Generator loss includes content loss LGcon and adversarial loss LGadv, which are used to extract and reconstruct the basic intensity information accompanying the optimal texture details and restore the thermal infrared salient targets. With the weight λ controlling the trade-off between two terms, the Fuse-Generator is defined as follows:

LG=λLGcon+LGadv. (2)

Among them, the content loss LGcon has two parts: basic-content loss LSSIM for extracting and reconstructing the basic intensity information, the max-gradient loss Lgra for obtaining the optimal texture details, which is formulated as follows:

LGcon=LSSIM+ηLgra, (3)

where the η is used to tradeoff the balance of intensity information and gradient variation.

Specifically, the LSSIM is formalized as follows:

LSSIM=ω(1SSIMVI,IF)+1SSIMIR,IF, (4)

where the ω is employed to tradeoff the balance of intensity information and gradient variation. The SSIMX,F is the metric to measure the similarity between two images, including three different factors of brightness, contrast and structure, which is mathematically defined as follows:

SSIMX,F=x,f2μxμf+C1μx2+μf2+C1·2σxσf+C2σx2+σf2+C2·σxf+C3σxσf+C3, (5)

where X and F in our work refer to the source image and fused image, respectively. The x and f mean the image patches of source image X and fused image F, μ and σ are the average values and the standard deviation. C1, C2 and C3 are the parameters to make the metric stable.

Only the basic-content loss LSSIM will cause the issue of information attenuation in texture details. Therefore, we further employ the max-gradient loss Lgra to obtain the optimal texture details. Lgra is mathematically formalized as follows:

Lgra=1HW(Ifgmax)2, (6)

where H and W are the height and width of the source images. (·) refers to the step of calculating the gradient map. The idea of the loss Lgra is to make the gradient map of the fused image (If) and the optimal gradient map of the source images gmax tend to be infinitely similar. The gmax is mathematically defined as follows:

gmax=round(I1+I2|I1+I2+eps|)max(|I1|,|I2|), (7)

where round(·) and max(·) mean the operations of rounding and taking the maximum value. The eps is a very small value to prevent the denominator from being 0.

The adversarial loss LGadv is to further restore the thermal infrared salient targets from the source IR image in the fused image, which is defined as:

LGadv=E[log(1D(IFM))], (8)

where M is the labeled mask reflecting the domains of salient thermal targets. When minimizing LGadv, IFM is encouraged to be realistic enough and close enough to real data, i.e., IRM to fool the discriminator D.

3.3.2. Loss of Discriminator LD

The discriminator loss LD is the term that forms an adversarial relationship with the Fuse-Generator adversarial loss LGadv. The LD is formulated as follows:

LD=E[log(D(IRM))]+E[log(1D(IFM))]. (9)

4. Experimental Results and Analysis

In this section, in order to show the superiority of our proposed FusionADA, we firstly compare it with 7 state-of-the-art fusion methods on the publicly available dataset RoadScene (https://github.com/hanna-xu/RoadScene (accessed on 1 December 2021)) qualitatively. Furthermore, we employ 8 metrics to evaluate their fusion results through qualitative comparisons. In addition, the ablation experiment of the adversarial learning is conducted. Finally, we show the fusion results of our FusionADA on another publicly available dataset TNO (https://figshare.com/articles/TNO_Image_Fusion_Dataset/1008029 (accessed on 1 December 2021)) dataset.

4.1. Experimental Settings

Dataset and Training Details. The training dataset is 45 aligned infrared and visible image pairs with different scenes selected from RoadScene. In order to improve the training performance, the tailoring and decomposition are applied as the expansion strategies before training to obtain a larger dataset. Specifically, the training dataset is uniformly cropped to 4736 patch pairs of size 128×128. There are 30 image pairs for testing. In the test phase, only the trained generator is used to generate the fusion image, and the input source images only need to be of the same resolution size.

Since this work is based on the adversarial learning of the generative adversarial network, we design a training strategy to keep the stability of the generative adversarial network in order to balance the adversarial relationship between the generator and the discriminator. The overall idea lies in that finding the loss value when the generator and the discriminator are in balance, and optimizing the generator or the discriminator to achieve their respective loss values through variable optimization times. The detailed training details of FusionADA are summarized in Algorithm 1. The λ, η, and ω are set to 3, 100 and 1.23 in Equations (2)–(4), respectively.

Algorithm 1: Training details of FusionADA
 Parameter definitions
NG, ND: The numbers of steps for training G, D.
Lmax, Lmin and LGmax are applied to determine a range when training.
Lmax and Lmin mean the adversarial losses of G and D.
LGmax: the total loss of G.
 We set Lmax=1.387, Lmin=1.386, and LGmax=0.1 in the first batch empirically in our work.
  • 1

    Initialize θG for G; θD for D.

  • 2

    For each training iteration:

  • 3
    Train DiscriminatorD:
    • Sample n VI patches {V1,,Vn} and n corresponding IR patches {I1,,In};
    • Acquire generated data {F1,,Fn}
    • Update Discriminator parameters θD by GradientDescentOptimizer to minimize LD in Equation (9); (step I)
    • While LD>LmaxandND<10, repeat step I. NDND+1;
    • Train GeneratorG:
    • Sample n VI patches {V1,,Vn} and n corresponding IR patches {I1,,In};
    • Acquire generated data {F1,,Fn}
    • Update parameters θG by RMSPropOptimizer for minimizing LG in Equation (2); (step II)
    • While LD<LminandNG<10, repeat step II. NGNG+1;
    • While LG>LGmaxandNG<10, repeat step II. NGNG+1;

4.2. Comparison Algorithms and Evaluation Metrics

In order to verify the effectiveness of our FusionADA, we show some intuitive results from our work with 7 other state-of-the-art infrared and visible fusion methods, containing gradient transfer fusion (GTF) [24], fourth-order partial differential equations (FPDE) [25], hybrid multi-scale decomposition (HMSD) [26], DenseFuse [19], proportional maintenance of gradient and intensity (PMGI) [27], unified unsupervised image fusion (U2Fusion) [28], and generative adversarial network with multi-classification constraints (GANMcC) [29]. Among them, GTF, FPDE and HMSD are fusion methods based on the traditional framework, while DenseFuse, PMGI, U2Fusion and GANMcC are deep learning-based fusion methods. Besides the intuitive evaluation, to do a more accurate evaluation of the fused results, we employ eight metrics to evaluate the fusion performance of these eight fusion methods, including standard deviation (SD) [30], spatial frequency (SF) [30], entropy (EN) [31], mean gradient (MG) and edge intensity (EI) [32] that measure the fused image itself, feature mutual information (FMI), the sum of the correlations of differences (SCD) [33], and visual information fidelity (VIF) [33] that measure the correlation between the fused image and source images. Specifically, SD, SF, EN, MG and EI are used to evaluate the contrast, frequency, amount of information, details, gradient amplitude of the edge point in the fused image, respectively. FMI is used to evaluate the amount of feature information that is transferred from source images to the fused image. SCD and VIF are used to measure the sum of the correlations of differences and information fidelity, respectively.

4.3. Qualitative Comparisons

There are four representative and intuitive fusion results of eight methods on infrared and visible images from the RoadScene dataset in Figure 2, Figure 3, Figure 4 and Figure 5. Compared to the existing seven other comparative fusion methods, our fused results show three obvious advantages. First, The salient thermal targets can be characterized clearly in our fused images, such as the pedestrians in Figure 2, Figure 3 and Figure 5, and the driver and passenger who got off the bus halfway in Figure 4 (all shown in the green boxes). The targets in the fused results of other methods all look dimmer compared to our results. Due to the salient thermal targets in our fused images, the drivers and machines can identify targets more easily and accurately, which facilitates the subsequent operations. Second, the scenes in our fused results show richer texture details, such as the schoolbag in Figure 2, the signs in Figure 3 and Figure 4 and the pavement marking in Figure 5 (all shown in the enlarged red boxes). Some scenes in the results of other methods seem fuzzier. The rich texture details are more conducive to scene understanding for the drivers and machines. Last but not the least, our results look cleaner than others without redundant fog or noise compared with the results of other methods.

Figure 2.

Figure 2

Qualitative comparison of FusionADA with corresponding seven state-of-the-art methods on the “building” image pair from the RoadScene dataset. From left to right, from top to bottom: source VIS image, the fused results of gradient transfer fusion (GTF), fourth-order partial differential equations (FPDE), hybrid multi-scale decomposition (HMSD) and DenseFuse. source IR image, the fused results of PMGI, U2Fusion, GANMcC and our FusionADA.

Figure 3.

Figure 3

Qualitative comparison of FusionADA with corresponding seven state-of-the-art methods on the “crossroad 1” image pair from the RoadScene dataset.

Figure 4.

Figure 4

Qualitative comparison of FusionADA with corresponding seven state-of-the-art methods on the “crossroad 2” image pair from the RoadScene dataset.

Figure 5.

Figure 5

Qualitative comparison of FusionADA with corresponding seven state-of-the-art methods on the “road” image pair from the RoadScene dataset.

4.4. Quantitative Comparisons

To have a more comprehensive and objective evaluation of the experimental results. We selected 30 test pairs of infrared and visible images randomly to further perform quantitative comparisons of our FusionADA with the competitors on eight fusion metrics. Each test image pair is aligned with the same resolution. The results of their values are summarized in Table 2. It is worth noting that our FusionADA can almost reach the optimal or suboptimal mean values on the eight metrics. For the metrics MG, EI, FMI and SCD, our FusionADA can also achieve comparable results with the suboptimal average values, only following behind a certain method by a narrow margin. It can be concluded that our results contain stronger contrast, richer texture details, more information and are closer to source images with less distortion.

Table 2.

Quantitative comparison of our FusionADA for infrared and visible image fusion with 5 other comparative methods. The average and standard deviation values of eight metrics for different methods are provided; red: optimal average values, blue: suboptimal average values.

SD SF EN MG EI FMI SCD VIF
GTF 0.1823±0.0712 0.0295±0.0117 7.3142±0.5122 0.0118±0.0088 0.1083±0.0321 0.8863±0.0452 0.9835±0.0425 0.5313±0.1532
FPDE 0.1337±0.0712 0.0393±0.0821 6.9404±0.5213 0.0200±0.0400 0.1668±0.3251 0.8659±0.0635 1.0919±0.6356 0.4991±0.2985
HMSD 0.1741±0.1029 0.0497±0.0212 7.3015±2.3251 0.0210±0.1254 0.2423±0.1254 0.8701±0.2563 1.4882±0.5632 0.8572±0.2153
DenseFuse 0.1733±0.1325 0.0398±0.0215 7.2950±0.3261 0.0192±0.0071 0.1625±0.0512 0.8749±0.0421 1.6316±0.2123 0.7597±0.7235
PMGI 0.1519±0.0852 0.0399±0.0212 7.1091±0.4213 0.0188±0.0057 0.1655±0.0564 0.8718±0.0432 1.2878±0.8421 0.7424±0.5231
U2Fusion 0.1625±0.0721 0.0493±0.0164 7.2328±0.5231 0.0237±0.0085 0.2116±0.1021 0.8716±0.0421 1.4837±0.4351 0.9337±0.9013
GANMcC 0.1702±0.0632 0.0341±0.0123 7.2345±1.2362 0.0163±0.0632 0.1506±0.0965 0.8605±0.3526 1.5819±0.3215 0.6832±0.1321
FusionADA 0.1863±0.0753 0.0509±0.0301 7.3456±0.2365 0.0221±0.0102 0.1851±0.1212 0.8801±0.0462 1.5825±0.3251 0.9619±0.6324

In addition, we also provide the mean and standard deviation of runtime for eight methods in Table 3. Although our FusionADA does not achieve optimal efficiency, it still plays a comparable role.

Table 3.

The average and standard deviation of running time for eight methods. (unit: second).

GTF FPDE HMSD DenseFuse PMGI U2Fusion GANMcC FusionADA
Running Time 2.72±1.20 0.95±0.38 0.59±0.15 0.29±0.04 0.14±0.03 0.98±0.45 0.295±0.32 0.10±0.01

4.5. Ablation Experiment of Adversarial Learning

The adversarial learning with a labeled mask is further employed in our FusionADA to restore the salient thermal targets from the source infrared image. To show the effect of adversarial learning, the following comparative experiments are conducted: (a) the adversarial learning is not applied; (b) the adversarial learning is applied. The experimental settings in other parts of the ablation experiments are the same. As can be seen from Figure 6, the fused results with adversarial learning own more salient thermal targets, which is more beneficial to the drivers and machines to identify the targets. Therefore, we can conclude that adversarial learning plays an important part in the fusion process.

Figure 6.

Figure 6

Results on whether the adversarial learning exists. From left to right: source VIS image, source IR image, the fused results without adversarial learning, the fused results of our FusionADA (with adversarial learning).

4.6. Generalization Performance

Our FusionADA also performs well on other datasets. To show the performance of our FusionADA on other datasets, we choose the TNO dataset to carry out the experiments without retraining the fusion methods. In particular, we choose two state-of-the-art methods HMSD and GANMcC that perform well on RoadScene to be the comparative methods with our FusionADA. The intuitive fusion results are presented in Figure 7. By contrast, with less noise, the salient thermal targets can be characterized more clearly and the richer texture details are contained in our results, which can be concluded that our FusionADA has good generalization performance and obtains excellent fusion results on other datasets.

Figure 7.

Figure 7

Fusion results on the TNO dataset.

5. Conclusions

In this paper, we propose a novel end-to-end fusion model for advanced driving assistance to obtain key and rich visual information under sophisticated road conditions, called FusionADA. Specifically, we achieve our FusionADA by fusing the infrared and visible images from infrared and visible sensors. For drivers and machines, based on the scenes, the salient thermal targets and rich texture details are indispensable for identifying targets easily and accurately. Therefore, in our model, we guide the generator to generate the optimal texture details from source images. Meanwhile, we constitute an adversarial framework with a labeled mask to further restore the salient thermal targets from the source infrared image. In addition, our FusionADA is achieved in a fully end-to-end way, which avoids manually designing complicated activity level measurements and fusion rules. The adequate experimental results reveal that our FusionADA not only presents better visual performance compared with other state-of-the-art methods, but also preserves the maximum or approximate maximum amount of features from source images.

Author Contributions

All authors have made great contributions to the work. Y.G. and X.W. designed the research and analyzed the results. Y.G. and X.W. performed the experiments and wrote the manuscript. X.W., C.Z. and B.L. gave insightful suggestions to the work and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 62075169.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Ziebinski A., Cupek R., Erdogan H., Waechter S. A survey of ADAS technologies for the future perspective of sensor fusion; Proceedings of the International Conference on Computational Collective Intelligence; Halkidiki, Greece. 28–30 September 2016; pp. 135–146. [Google Scholar]
  • 2.Ma J., Ma Y., Li C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion. 2019;45:153–178. doi: 10.1016/j.inffus.2018.02.004. [DOI] [Google Scholar]
  • 3.Huang X., Qi G., Wei H., Chai Y., Sim J. A novel infrared and visible image information fusion method based on phase congruency and image entropy. Entropy. 2019;21:1135. doi: 10.3390/e21121135. [DOI] [Google Scholar]
  • 4.Ma J., Xu H., Jiang J., Mei X., Zhang X.P. DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. 2020;29:4980–4995. doi: 10.1109/TIP.2020.2977573. [DOI] [PubMed] [Google Scholar]
  • 5.Yin S., Wang Y., Yang Y.H. A Novel Residual Dense Pyramid Network for Image Dehazing. Entropy. 2019;21:1123. doi: 10.3390/e21111123. [DOI] [Google Scholar]
  • 6.Li X., Zhang X., Ding M. A sum-modified-Laplacian and sparse representation based multimodal medical image fusion in Laplacian pyramid domain. Med. Biol. Eng. Comput. 2019;57:2265–2275. doi: 10.1007/s11517-019-02023-9. [DOI] [PubMed] [Google Scholar]
  • 7.Teng J., Wang S., Zhang J., Wang X. Neuro-fuzzy logic based fusion algorithm of medical images; Proceedings of the International Congress on Image and Signal Processing; Yantai, China. 16–18 October 2010; pp. 1552–1556. [Google Scholar]
  • 8.Zhao F., Xu G., Zhao W. CT and MR Image Fusion Based on Adaptive Structure Decomposition. IEEE Access. 2019;7:44002–44009. doi: 10.1109/ACCESS.2019.2908378. [DOI] [Google Scholar]
  • 9.Liu Y., Yang X., Zhang R., Albertini M.K., Celik T., Jeon G. Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter. Entropy. 2020;22:118. doi: 10.3390/e22010118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Jiang W., Yang X., Wu W., Liu K., Ahmad A., Sangaiah A.K., Jeon G. Medical images fusion by using weighted least squares filter and sparse representation. Comput. Electr. Eng. 2018;67:252–266. doi: 10.1016/j.compeleceng.2018.03.037. [DOI] [Google Scholar]
  • 11.Xu Z. Medical image fusion using multi-level local extrema. Inf. Fusion. 2014;19:38–48. doi: 10.1016/j.inffus.2013.01.001. [DOI] [Google Scholar]
  • 12.Jiang F., Kong B., Li J., Dashtipour K., Gogate M. Robust visual saliency optimization based on bidirectional Markov chains. Cogn. Comput. 2020:1–12. doi: 10.1007/s12559-020-09724-6. [DOI] [Google Scholar]
  • 13.Tian X., Chen Y., Yang C., Ma J. Variational Pansharpening by Exploiting Cartoon-Texture Similarities. IEEE Trans. Geosci. Remote Sens. 2021 doi: 10.1109/TGRS.2020.3048257. [DOI] [Google Scholar]
  • 14.Ma J., Zhang H., Yi P., Wang Z. SCSCN: A Separated Channel-Spatial Convolution Net with Attention for Single-view Reconstruction. IEEE Trans. Ind. Electron. 2020;67:8649–8658. doi: 10.1109/TIE.2019.2950866. [DOI] [Google Scholar]
  • 15.Ma J., Wang X., Jiang J. Image Super-Resolution via Dense Discriminative Network. IEEE Trans. Ind. Electron. 2020;67:5687–5695. doi: 10.1109/TIE.2019.2934071. [DOI] [Google Scholar]
  • 16.Shopovska I., Jovanov L., Philips W. Deep visible and thermal image fusion for enhanced pedestrian visibility. Sensors. 2019;19:3727. doi: 10.3390/s19173727. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Huang J., Le Z., Ma Y., Mei X., Fan F. A generative adversarial network with adaptive constraints for multi-focus image fusion. Neural Comput. Appl. 2020;32:15119–15129. doi: 10.1007/s00521-020-04863-1. [DOI] [Google Scholar]
  • 18.Liu Y., Chen X., Cheng J., Peng H., Wang Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolution Inf. Process. 2018;16:1850018. doi: 10.1142/S0219691318500182. [DOI] [Google Scholar]
  • 19.Li H., Wu X.J. DenseFuse: A fusion approach to infrared and visible images. IEEE Trans. Image Process. 2018;28:2614–2623. doi: 10.1109/TIP.2018.2887342. [DOI] [PubMed] [Google Scholar]
  • 20.Liu Y., Chen X., Ward R.K., Wang Z.J. Image fusion with convolutional sparse representation. IEEE Signal Process. Lett. 2016;23:1882–1886. doi: 10.1109/LSP.2016.2618776. [DOI] [Google Scholar]
  • 21.Ma J., Yu W., Liang P., Li C., Jiang J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion. 2019;48:11–26. doi: 10.1016/j.inffus.2018.09.004. [DOI] [Google Scholar]
  • 22.Ma J., Liang P., Yu W., Chen C., Guo X., Wu J., Jiang J. Infrared and visible image fusion via detail preserving adversarial learning. Inf. Fusion. 2020;54:85–98. doi: 10.1016/j.inffus.2019.07.005. [DOI] [Google Scholar]
  • 23.Xu H., Liang P., Yu W., Jiang J., Ma J. Learning a generative model for fusing infrared and visible images via conditional generative adversarial network with dual discriminators; Proceedings of the International Joint Conference on Artificial Intelligence; Macao, China. 10–16 August 2019; pp. 3954–3960. [Google Scholar]
  • 24.Ma J., Chen C., Li C., Huang J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion. 2016;31:100–109. doi: 10.1016/j.inffus.2016.02.001. [DOI] [Google Scholar]
  • 25.Bavirisetti D.P., Xiao G., Liu G. Multi-sensor image fusion based on fourth order partial differential equations; Proceedings of the International Conference on Information Fusion; Xi’an, China. 10–13 July 2017; pp. 1–9. [Google Scholar]
  • 26.Zhou Z., Wang B., Li S., Dong M. Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters. Inf. Fusion. 2016;30:15–26. doi: 10.1016/j.inffus.2015.11.003. [DOI] [Google Scholar]
  • 27.Zhang H., Xu H., Xiao Y., Guo X., Ma J. Rethinking the Image Fusion: A Fast Unified Image Fusion Network based on Proportional Maintenance of Gradient and Intensity; Proceedings of the AAAI Conference on Artificial Intelligence; Hilton New York Midtown, NY, USA. 7–12 February 2020; pp. 12797–12804. [Google Scholar]
  • 28.Xu H., Ma J., Jiang J., Guo X., Ling H. U2fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 2020 doi: 10.1109/TPAMI.2020.3012548. [DOI] [PubMed] [Google Scholar]
  • 29.Ma J., Zhang H., Shao Z., Liang P., Xu H. GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans. Instrum. Meas. 2020;70:1–14. [Google Scholar]
  • 30.Eskicioglu A.M., Fisher P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995;43:2959–2965. doi: 10.1109/26.477498. [DOI] [Google Scholar]
  • 31.Roberts J.W., van Aardt J.A., Ahmed F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008;2:023522. [Google Scholar]
  • 32.Yang Z., Chen Y., Le Z., Fan F., Pan E. Multi-source medical image fusion based on Wasserstein generative adversarial networks. IEEE Access. 2019;7:175947–175958. doi: 10.1109/ACCESS.2019.2955382. [DOI] [Google Scholar]
  • 33.Aslantas V., Bendes E. A new image quality metric for image fusion: The sum of the correlations of differences. AEU-Int. J. Electron. Commun. 2015;69:1890–1896. doi: 10.1016/j.aeue.2015.09.004. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES