Next Article in Journal
Effects of Forest Gaps on Forest Floor Microbial Community Composition in Pinus tabulaeformis Forests in a Rocky Mountain Area, Beijing, China
Previous Article in Journal
Characteristics of Vegetation Change and Its Climatic and Anthropogenic Driven Pattern in the Qilian Mountains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints

College of Forestry, Central South University of Forestry and Technology, Changsha 410004, China
*
Author to whom correspondence should be addressed.
Forests 2023, 14(10), 1952; https://doi.org/10.3390/f14101952
Submission received: 2 June 2023 / Revised: 8 September 2023 / Accepted: 19 September 2023 / Published: 26 September 2023
(This article belongs to the Section Natural Hazards and Risk Management)

Abstract

:
Aimed at addressing deficiencies in existing image fusion methods, this paper proposed a multi-level and multi-classification generative adversarial network (GAN)-based method (MMGAN) for fusing visible and infrared images of forest fire scenes (the surroundings of firefighters), which solves the problem that GANs tend to ignore visible contrast ratio information and detailed infrared texture information. The study was based on real-time visible and infrared image data acquired by visible and infrared binocular cameras on forest firefighters’ helmets. We improved the GAN by, on the one hand, splitting the input channels of the generator into gradient and contrast ratio paths, increasing the depth of convolutional layers, and improving the extraction capability of shallow networks. On the other hand, we designed a discriminator using a multi-classification constraint structure and trained it against the generator in a continuous and adversarial manner to supervise the generator, generating better-quality fused images. Our results indicated that compared to mainstream infrared and visible image fusion methods, including anisotropic diffusion fusion (ADF), guided filtering fusion (GFF), convolutional neural networks (CNN), FusionGAN, and dual-discriminator conditional GAN (DDcGAN), the MMGAN model was overall optimal and had the best visual effect when applied to image fusions of forest fire surroundings. Five of the six objective metrics were optimal, and one ranked second-to-optimal. The image fusion speed was more than five times faster than that of the other methods. The MMGAN model significantly improved the quality of fused images of forest fire scenes, preserved the contrast ratio information of visible images and the detailed texture information of infrared images of forest fire scenes, and could accurately reflect information on forest fire scene surroundings.

1. Introduction

Forest fire surroundings are affected by terrain, meteorology, light, smoke, fire spread, and other factors, making it difficult to judge the internal conditions of the forest fire scene, which can easily lead to casualties if rescuers make errors in judgment during operations [1]. Therefore, firefighters need to quickly and accurately understand the conditions inside the fire and make avoidance decisions. Forest fire surroundings have obvious visible and infrared image feature information. Visible images of forest fire scenes can clearly express texture detail information, but they are difficult to adapt to changes in forest fire surroundings; infrared images of forest fire scenes are less affected by light and smoke and can better reflect the changes in forest fire surroundings, but it is difficult to reflect texture detail features in them.
Due to the complexity of forest fire scene surroundings, relying only on single-sensor data often does not provide enough information to support firefighters’ decision-making. Image fusion technology has important advantages in forest fire scene rescue. By fusing visible and infrared images of forest fires, the advantageous features of both sources can be fully utilized to provide more comprehensive and accurate fire scene information. Visible light images can provide clear texture details to help firefighters recognize objects and terrain, but they are affected by factors such as light and smoke. Infrared images, on the other hand, can better reflect the distribution of hot spots and fire expansion, and are relatively insensitive to light and smoke. Therefore, the study of visible and infrared image fusion in the scene of forest fires can make up for the limitations of each of the images, provide more comprehensive and accurate information about the fire scene, improve the efficiency and safety of rescuers, and thus significantly reduce the casualties caused by forest fires. Image fusion technology has gained widespread attention for its convenient, fast and economical features, assisting firefighters in making accurate decisions and providing an efficient method for fire rescue. At the same time, the application of image fusion technology can also provide a reference and inspiration for image processing and analysis in other fields and promote the development and application of imaging technology.
In addition, many studies have shown that image fusion techniques have the potential for practical applications in forest fire scene rescue. For example, Dios et al. used image fusion techniques to improve the accuracy of fire source localization, fire monitoring, and fire boundary identification [2]. Nemalidinne et al. used visible and infrared image fusion techniques to improve fire detection [3]. These findings suggest that image fusion techniques have great potential to improve the efficiency and safety of firefighters in forest fire rescues.
The scene of forest fire image fusion is a special form of image fusion field. Current visible and infrared image fusion methods can be classified into traditional methods and deep learning-based methods according to their principles. Traditional image fusion methods mainly include spatial domain methods and transform domain methods. Representative methods include logical filtering methods [4], color space methods [5], multi-scale transform methods [6], principal component analysis methods [7], sparse representation methods [8], and wavelet transform fusion methods [9]. However, traditional fusion methods have large limitations in terms of feature extraction and fusion rules. On the one hand, the diversity of source images makes the manual design of feature extraction and fusion rules increasingly complex; on the other hand, the generalizability of traditional methods is limited, and further improvement of fusion performance faces difficulties.
In recent years, the rapid development of deep learning has driven great progress in the field of image fusion. Deep learning-based methods utilize the powerful nonlinear fitting ability of neural networks to make fused images with desired distributions and further improve the performance of fused images. Convolutional neural networks (CNN) construct deep structural neural networks by training on natural samples, which enables a deeper image feature extraction to maintain the integrity of structural information and the preservation of detailed information in fused images [10]. CNNs have achieved good performance in multifocal images, but their computation is time-consuming, and the fusion performance depends on the characteristics of the training samples [11]. Ma et al. innovatively introduced generative adversarial networks (GAN) into the field of image fusion, combining adversarial learning and content-specific loss bootstrapping to achieve the preservation of significant contrast and texture details in fused images in an unsupervised situation [12]. However, the aforementioned methods do not sufficiently consider the extraction of information from shallow networks when extracting source image information, so there is still room for further improvement in fusion performance. If we can improve the degree of information utilization in the training process of GANs and achieve a good balance in maintaining visible and infrared image information, we are bound to further improve the performance of image fusion.
The superior performance of existing deep learning-based methods relies on a large number of labeled (reference image) datasets. However, labeled images are difficult to obtain for forest fire scene image fusion tasks. In order to make full use of the small amount of labeled data and a large amount of unsupervised data, this paper proposes the visible and infrared image fusion of forest fire scenes using a method based on a multi-level and multi-classification GAN, referred to as MMGAN, which divides the input of the generator into a gradient path and intensity path, constructs primary and secondary information loss functions, and improves the depth of the convolutional layer of the generator, enhancing the information extraction capability of the shallow network. A multi-classifier is used in the discriminator, which determines the probability that the input is a visible image and an infrared image. By constraining these two probabilities through successive standoff learning, the generator can align the probability distributions of both IR and visible images. This effectively prevents the dilution of dominant information, resulting in fused images with pronounced contrast and intricate texture details. The comparative experimental results show that the proposed method can better accomplish the fusion of effective information from visible and infrared images of the forest fire scene, and the proposed method has better fusion performance compared with other, similar fusion methods. In addition, the proposed method has good generalizability and can be extended to any visible and infrared image fusion dataset.

2. Materials and Methods

2.1. Data and Pre-Processing

The visible and infrared image data of the forest fire scene in this study were acquired in real time from forest fires in Xintian County, Yongzhou City, Hunan Province, through a visible and infrared binocular camera integrated into a single firefighter’s helmet. Xintian County is located in the southern part of Hunan Province, a region rich in forest resources and characterized geographically by mountainous, hilly, and forest-covered areas. The forest fire encompasses multiple ignition points and areas, yielding a wealth of visible and infrared image data. The acquired image data captured the characteristics of the fire smoke, burned areas and surroundings, and recorded the different stages and locations of the fire, providing a valuable data resource for research.
The TNO dataset comprises multispectral images of various scenes captured by different multi-band camera systems, including Athena, DHV, FEL and TRICLOBS. It is a widely used public research dataset in the field of infrared and visible image fusion, and its images have been subjected to strict image alignment [13].
In this paper, the forest fire scene’s visible and infrared images and the TNO public dataset were selected as the data for the comparison experiments. The number of image pairs used for testing in the forest fire domain was 50 pairs, and 4000 pairs of images were cropped using a sliding window of 8 × 10; the number of image pairs used for testing in the TNO dataset was 100 pairs, and 8000 pairs of images were cropped using a sliding window of 8 × 10. All of these images maintain a uniform size of 2.6 × 3.5 cm, accounting for various factors such as data acquisition and processing.
To address the problems arising from the acquisition and transmission of forest fire scene images, the acquired images are preprocessed, including (1) a particle swarm optimization-based approach to grayscale forest fire scene images [14,15], based on the grayscale histogram distribution law of high pixel values in a single channel of a color image, particle swarm search is used to select the optimal threshold, as well as to automatically generate the weights of components for grayscale processing of forest fire scene images. (2) An image denoising method based on wavelet transform is used to remove the effect of noise on forest fire scene images to improve the image quality [16], the wavelet transform analysis method is established based on the idea of short-time Fourier transform localization, the image signal is segmented and refined according to multiple levels, and the wavelet coefficients obtained from the wavelet transform are appropriately processed, which can be used for suppressing the noise randomly generated by the sensor in the imaging process. (3) A frequency domain-based image enhancement method is employed to enhance the visual quality of the images, highlighting finer details in the forest fire scene images [17]. This involves processing the image signal with a specialized high-pass filter to achieve enhancement. Gaussian filtering uses a Gaussian convolution kernel to apply different Gaussian weights to pixel gray values at various locations in the image, smoothing the forest fire scene image. (4) Based on the above data preprocessing, the SURF alignment method is used to align the visible and infrared images of the forest fire scene for the problem of non-overlap, which leads to errors in image fusion [18], extract the feature points of the forest fire scene image, use integral images to simplify the operation process, construct the Hessian matrix to extract the image feature points, construct the scale space to form the three-dimensional scale space of the feature points; obtain the direction of the feature points by calculating the harr wavelet response; form the feature description subvector based on the main direction of the feature points; add the scale factor in eliminating the erroneous feature points, choose the appropriate similarity metric to match the similar feature points, and finally obtain the completely overlapped visible and infrared images of the forest fire scene.

2.2. Generative Adversarial Network Fundamentals

Generative adversarial networks (GAN) were proposed by Goodfellow et al. in 2014 and consist of a generator and a discriminator [19]. The generator is the target network that aims to generate false data that match the target distribution; the discriminator is a classifier that is responsible for accurately distinguishing between real data and the fake data generated by the generator. Thus, there is an adversarial relationship between the generator and the discriminator. In other words, the generator wants to generate forged data that the discriminator cannot identify, while the discriminator wants to distinguish the forged data accurately. The generator and discriminator enhance their capabilities through continuous iterations until the discriminator can no longer differentiate between real data and the forged data generated by the generator. At this point, the generator has the ability to generate data that match the target distribution. Next, the above adversarial learning process is described formally.
Suppose the generator is denoted as G and the discriminator is denoted as D. The random data input to the generator are denoted as Z = (z1, z2,…, zn), and the target data as X = (x1, x2,…, xn)~Px. Then, the generator is committed to estimate the distribution of the target data XP X and generates data G(z) matching that distribution as much as possible, while the discriminator D needs to distinguish accurately between the real data X and the generated pseudo data G(z). In summary, the purpose of generating adversarial networks is to make the distribution of the faked data PG continuously approximate the target data distribution PX during continuous adversarial training. Therefore, the objective function of the GAN is defined as Equation (1):
min G   max D   E x i ~ P X log D x i + E z i ~ P z log 1 D G z i
In continuous iterative training, the generator and the discriminator promote each other as an adversarial relationship, continuously improving their falsification or discriminatory ability. When the distance between the two distributions becomes sufficiently small, the discriminator will no longer be able to distinguish between real and fake data. At this point, we can say that the generator has successfully estimated the distribution of the training data.

2.3. MMGAN Structural Design

The overall network structure consists of a generator network and a discriminator network.

2.3.1. Network Structure of Generator

The multilayer convolutional structure has a powerful feature representation capability as well as a hierarchical learning capability, and with the increase in the number of convolutional layers, the model can gradually capture higher-level features as well as understand the global and local information of the image, improving the efficiency of image analysis and processing. Hence, by increasing the depth of the convolutional layers, the generator network architecture enhances the accuracy of model training. We used a pre-trained network on ImageNet as the base network and fine-tuned it for the fusion task. For both the gradient path and the intensity path, we employed 8 convolutional layers, each utilizing a 3 × 3 convolutional kernel. As an example, one of these layers had dimensions 64 × 3 × 3 × 128, where 64 represents the number of input channels, 3 × 3 is the convolutional kernel size, and 128 is the number of output channels. Each convolutional layer includes a Leaky ReLU activation function layer and a batch normalization (BN) layer. In addition, the input of the 4th convolutional layer is feature-connected to the output of the 2nd and 3rd convolutional layers, and the input of the 6th convolutional layer is feature-connected to the output of the 4th and 5th convolutional layers. This structure helps to preserve the characteristics of the shallow network, thus effectively reducing the loss of image information. Finally, the two feature maps are interconnected along the channels to achieve full fusion of information. The step size of the last convolutional layer is set to 1, and the output uses a Tanh activation function and a 1 × 1 convolutional kernel to obtain the fusion result of the network. The structure of the generator network is shown in Figure 1.
By designing the gradient path and intensity path, the extraction of contrast information of the visible image and texture information of the infrared image is realized in the generator network structure. The gradient path aims to extract the texture information of the image, while the visible image is rich in texture detail information, so two visible images and one infrared image are coupled to one channel as the input of the gradient path; the intensity path aims to extract the contrast information of the image, while the infrared image is rich in contrast information, so two infrared images and one visible image are coupled to one channel as the input of the intensity path. This serves as the input to the intensity path, enabling the balanced extraction of gradient and intensity information and facilitating complementary information between both images.

2.3.2. Network Structure of Discriminator

The discriminator is designed using a multi-level multi-classification constraint structure capable of distinguishing visible image features, infrared image features, and fused image features generated by the generator. It estimates the probability of each class within the input image, resulting in a 1 × 2 probability vector as output. We accomplish this by employing a pre-trained network from ImageNet as the base network and fine-tuning it for the discriminator task. The discriminator consists of four convolutional layers, and one linear layer. Four convolutional layers are used to process the input image features, all of size 3 × 3 convolutional kernels, all activation functions are Leaky ReLU functions, and the last three convolutional layers are set with a step size of 2 and are batch normalized by adding BN layers. The last linear layer outputs a 1 × 2 two-dimensional classification probability based on the image features extracted from the first four convolutional layers as input, indicating the probability of the input feature being a visible image feature Pvis and the probability of the input feature being an infrared image feature Pir, respectively. The generator learns a reasonable fusion strategy at that time. The discriminator network structure is shown in Figure 2.

2.3.3. General Framework of MMGAN Model

Compared with the overall GAN framework, the generator input path of MMGAN is divided into gradient as well as intensity. Firstly, for the gradient path of the generator, two visible images and one infrared image are combined along the channel dimension; secondly, for the intensity path of the generator, two infrared images and one visible image are combined along the channel dimension; then, with specific loss functions and network design, the generator can extract gradient and intensity information in a balanced way, so that the primary gradient and secondary intensity information of the visible image is collected, and the primary intensity and secondary gradient information of the infrared images complement each other. As the discriminator assesses the fused images, the generator anticipates that these images are a combination of both visible and infrared components, resulting in an expectation for the discriminator’s output to yield a high probability. The discriminator’s task is to accurately judge the fused image as a false image, i.e., it expects the output probability to be both small, which excludes the possibility that the generated image is a visible and infrared image, i.e., the output image can only be a fused image. Therefore, the generator and the discriminator are continuously trained against each other. When both of the two-dimensional probabilities Pvis and Pir of the discriminator output are small, the generator can independently generate information-balanced fusion images. The general framework of GAN is shown in Figure 3, and the general framework of MMGAN is shown in Figure 4.
Through the aforementioned design, this paper’s method can produce superior fusion results compared to traditional generative adversarial networks. It accomplishes this by preserving not only significant visible image contrast information but also by incorporating rich texture details from infrared images.

2.3.4. Loss Function Design

The loss functions include generator loss function LG and discriminator loss function LD, where the discriminator loss function consists of visible light discriminator loss function L D v i s and infrared discriminator loss function L D i r and forest fire discriminator loss function L D f u s e d .
(1)
Generator loss function
The generator loss function consists of the generator content loss function L G c o n for constraint information extraction and the generator adversarial loss function L G a d v for constraint information balancing. The specific formula is shown in Equation (2):
L G = γ L G c o n + L G a d v
where γ is the regularization parameter responsible for maintaining the balance between the two items and is set to 100.
The L G c o n content loss function consists of four parts: the main gradient loss, the main intensity loss, the auxiliary intensity loss and the auxiliary gradient loss, and the content loss function is shown in Equation (3):
L G c o n = L i n t e n s i t y m a i n + L g r a d m a i n + L g r a d a u x + L i n t e n s i t y a u x = δ 1 I fused - I ir F 2 + δ 2 I fused - I vis F 2 + δ 3 I fused - I ir F 2 + δ 4 I fused - I vis F 2
where δ 1 , δ 2 , δ 3 , and δ 4 are set to 1, 2, 3, and 4, respectively.
The balance between the image information can be achieved by adding the adversarial loss with discriminator to the generator’s loss function, and the adversarial loss function definition L G a d v is shown in Equation (4):
L G adv = D I fused n [ 1 ] - a 2 + D I fused n [ 2 ] a 2
where a denotes the discriminator’s probabilistic labeling for the fused image, and a is set to 1. The first term in the vector D(.) [1] denotes the probability that the discriminator determines that the generated fused image of the forest fire scene is a visible light image, while the second term D(.) [2] then denotes the probability that the discriminator determines that the generated forest fire fusion image is an infrared image.
(2)
Discriminant loss function
The discriminator is a multi-level multi-classifier that estimates the probability of each class of the input image. The loss function must continuously improve the discriminator discriminative power, which drives it to effectively and accurately discriminate whether the fused image is a visible image or an infrared image. Therefore, the discriminator loss function LD consists of three components to determine the loss L D v i s of the visible image, the loss L D i r of the infrared image and the loss L D f u s e d of the fused image, and the discriminator loss function is shown in Equation (5):
L D = L D vis + L D ir + L D fused
The discriminator outputs 1 × 2 vectors including P v i s = D x [ 1 ] and P i r = D x [ 2 ] .
In order to improve the discriminator’s ability to recognize visible images, the predicted P v i s needs to be close to 1 and P i r close to 0 when the input image is a visible image. The loss function L D v i s corresponding to the visible light image is shown in Equation (6):
L D vis = 1 2 N i = 1 N P vis I vis n - a 1 2 + P ir I vis n - a 2 2
where a1 and a2 are probability labels corresponding to visible light images, a1 is set to 1, and a2 is set to 0.
In order to improve the discriminator’s ability to recognize infrared images, when the input image is an infrared image, the prediction Pir needs to be close to 1 and Pvis close to 0. The loss function L D i r corresponding to the infrared image is defined as shown in Equation (7):
L D ir = 1 2 N i = 1 N P vis I ir n - b 1 2 + P ir I ir n - b 2 2
where b1 and b2 are the probability labels corresponding to the infrared images, b1 is set to 0, and b2 is set to 1.
Finally, in order to improve the discriminator discrimination performance, when the input image is a forest fire fusion image, both the prediction Pir and Pvis need to be close to 0 in order to represent the ability of the forest fire fusion image to be neither a visible image nor an infrared image.
The loss function L D f u s e d corresponding to the fused image is defined as shown in Equation (8):
L D fused = 1 2 N i = 1 N P vis I fused n - c 1 2 + P ir I fused n - c 2 2
where c1 and c2 are the probability labels corresponding to the fused images of the forest fire, c1 is set to 0, and c2 is set to 0.

2.4. Training Process and Parameter Setting

2.4.1. Training Process

(1)
Initialize the network parameters of the generator and the discriminator, input the visible and infrared images of the forest fire scene into the corresponding discriminator, and update the parameters of the discriminator according to the output results.
(2)
Link the visible and infrared images of the forest fire scene and input them into the generator network, obtain the fused images of the forest fire scene generated by the generator, and update the parameters of the generator network by combining the corresponding generator loss. Repeat this process up to 20 times.
(3)
Input the visible image and infrared map of the forest fire scene and the fusion image generated by the generator into the corresponding discriminator, respectively, and update the parameters of the discriminator according to the output results. Repeat this process up to 5 times.
(4)
Repeat processes (2) and (3) until the network reaches equilibrium.

2.4.2. Parameter Setting

Throughout the generator and discriminator training process, we set the Leaky ReLU activation function layer parameter to 0.01, configured a batch size of 48 in the batch grouping layer, and employed Adam as the optimizer to adjust the network’s weight parameters. These parameters were updated by calculating the gradient to minimize the loss function. The number of training iterations was set to 10, and the initial learning rate was set to 0.0001.

2.5. Evaluation System

In this paper, the performance of the fusion method is evaluated from two aspects: subjective evaluation and objective evaluation. The subjective evaluation approach relies on the observer’s visual perception, and a successful fusion result should encompass both the pronounced contrast of the infrared image and the abundant texture of the visible image.
Six quantitative metrics that are widely used in the field of image fusion were chosen for the objective evaluation, including information entropy (IEN) [20], standard deviation (STD) [21], average gradient (AG) [22], space frequency (SF) [23], feature mutual information (FMI) [24] and structural similarity (SSIM) [25].
IEN = l = 0 L 1 p l   log 2   p l
STD = 1 M × N i = 1 M j = 1 N F i , j F ¯ 2
AG = 1 ( M 1 ) ( N 1 ) i = 1 M 1 j = 1 N 1 [ F ( i , j ) F ( i + 1 , j ) ] 2 + [ F ( i , j ) F ( i + 1 , j ) ] 2 2
SF = R F 2 + C F 2
F M I F A B = F M I ( A , F ) + F M I ( B , F )
SSIM = SSI M A , F + SSI M B , F
In Equations (9)–(14), pl is the frequency of the occurrence of points with gray value l, L is the number of gray levels of the fused image, and the larger value of STD represents the higher contrast of the image and the better quality of the image. Where, M denotes the width of the fused image, N denotes the height of the fused image, F(i,j) is the gray value of image F at pixel (i,j), F ¯ is the mean gray value, RF denotes the spatial frequency in the horizontal direction, CF denotes the spatial frequency in the vertical direction, F is the fused image, A and B are the source images, SSIMA,f is the brightness loss, and SSIMB,f is the correlation loss. The larger the value of all metrics, the better.

3. Results

Five image fusion methods were selected for experimental comparison with the methods in this paper: (1) anisotropic diffusion-based image fusion method (ADF) [26]; (2) guided filtering-based image fusion method (GFF) [27]; (3) convolutional neural network-based image fusion method (CNN) [28]; (4) generative adversarial network-based image fusion method (FusionGAN) [12]; and (5) image fusion method based on dual discriminator generative adversarial network (DDcGAN) [29]. The proposed methods in this paper are all pre-trained on the dataset.

3.1. TNO Dataset Fusion Results and Comparative Analysis

3.1.1. Subjective Assessment

To demonstrate the superiority of the MMGAN model proposed in this paper, comparative experiments were first conducted on the publicly available dataset TNO.
Figure 5 shows three sets of representative subjective evaluation results to demonstrate the performance of various methods. In the subjective evaluation comparison results, the ADF method has low contrast and does not highlight significant target features; the GFF fusion method has severe distortion, and the details in the local zoomed-in image have become illegible; the CNN method focuses too much on preserving the structural texture and neglects the preservation of the thermal radiation target; the FusionGAN method loses many texture details, and the boundary produces artifacts that make the boundary of the target widen and the whole blurred; while the DDcGAN method better maintains the saliency of the thermal radiation target in the infrared image, our proposed method excels in preserving the sharpness of the thermal target’s edges and presents a superior visual effect. For example, in the first and third sets of results, the method in this paper is able to better preserve the house structure with clearly visible edges, while other methods result in blurred contours due to edge diffusion.

3.1.2. Objective Assessment

In order to objectively evaluate the fusion effectiveness of ADF, GFF, CNN, FusionGAN, DDcGAN, and MMGAN, we conducted a quantitative analysis on 20 test im-ages from the TNO dataset, and the results are presented in Table 1. According to the data in the table, our proposed method in this paper achieved the top rank in four metrics: IEN, AG, FMI, and SSIM, and it secured the second position in STD and SF metrics, just after DDcGAN. These objective findings indicate that our method outperforms others in terms of information richness, minimal introduction of pseudo-information, high correlation with the source image, and superior contrast.
Moreover, the quantitative results are in line with the visual assessment shown in Figure 5. The fusion outcomes obtained by our method exhibit a clearer and more detailed representation of the forest fire scene compared to other methods. The enhanced information richness and improved contrast in the fused images can significantly assist firefighters in comprehending the fire scene environment and making critical decisions during rescue operations.
In conclusion, our proposed method demonstrates superior performance in objective evaluation metrics on the TNO dataset compared to existing fusion techniques. We believe that the results presented in this study will have a meaningful impact on advancing forest fire image fusion techniques and contribute to more effective and safer forest firefighting efforts.

3.2. Fusion Results and Comparative Analysis of Forest Fire Scene Datasets

In order to further demonstrate the superiority of the MMGAN algorithm proposed in this paper, a comparative analysis of subjective and objective evaluations of forest fire scene image fusion results was performed.

3.2.1. Subjective Assessment

Three groups of typical fusion results are selected to demonstrate the performance of each method, as shown in Figure 6. Observing the fusion result plots, we can find that the ADF method contains rich texture detail information, but the noise is more obvious. The GFF has a great defect in the contrast display of the image, the texture information and contrast information in the plots are defocused, a certain number of artifacts are at the edges, and the clarity is poor. The CNN method does a good job in preserving the tree texture, but the brightness is too high, resulting in insufficient thermal radiation information. While the FusionGAN fusion method provides rich contrast information, it suffers from significant loss of infrared texture detail. DDcGAN retains more infrared texture information compared to FusionGAN, but it is slightly lacking compared to MMGAN. It is obvious from the figure that MMGAN can retain the salient targets in the IR image very accurately with almost no loss of thermal radiation intensity, and the edges remain sharp; meanwhile, it can also retain the texture details in the visible image very well. Therefore, MMGAN not only retains rich texture detail information of visible and infrared images, but also maintains rich contrast information. Overall, the method in this paper has obvious advantages in terms of sharpness and contrast, and can achieve a good balance of visible and infrared image information.

3.2.2. Objective Assessment

In order to objectively evaluate the fusion effects of ADF, GFF, CNN, FusionGAN, DDcGAN and MMGAN, the fusion results for forest fire scenes were quantitatively analyzed on 20 forest fire scene images, and the results are shown in Table 2. The data in the table show that the MMGAN proposed in this paper obtained the highest average value in five objective quality evaluation metrics, namely, IEN, STD, AG, FMI, and SSIM; among them, although the SF value had only the second-highest value, it can be seen that it significantly exceeded the SF value of other methods, especially FusionGAN, which indicates that MMGAN can be better than FusionGAN in maintaining a good balance of visible and infrared image information. In addition, the SF values obtained by MMGAN were much larger than those of all other methods, thus indicating that the method in this paper can better fuse the contrast and texture detail information of visible and infrared images, and the fused image information is richer.
In summary, MMGAN transmits the most information from the source image to the fused image, introduces the least pseudo-information, and maintains the edges best during the fusion process. The resulting fused images contain the most information, exhibit the highest contrast, and boast the richest overall texture structure. Therefore, the method in this paper is also quantitatively advantageous compared to the other fusion methods.

3.3. Comparative Analysis of Fusion Efficiency

Operational efficiency is one of the important metrics to assess the performance of a method. To evaluate operational efficiency, we calculated the average running times of various methods on both the TNO dataset and the forest fire scene dataset and conducted a comparison. The results are shown in Table 3. The method proposed in this paper achieves the fastest average running speed on both datasets, which is more than five times faster than other compared methods. This means that our method is able to complete the image fusion task in a shorter time, improving the overall efficiency.

4. Discussion

This study improved the model learning as well as generation capabilities by improving the design of the generator, discriminator, and loss function. Specifically, by dividing the input path of the generator of the traditional GAN into gradient and intensity, the depth of the convolutional layer of the generator is increased to further improve the information extraction capability of the shallow network, enabling the generator to better capture the detailed features of visible and infrared images, thus improving the quality and accuracy of the fused images. At the same time, the discriminator is designed by using multiple classifiers that can discriminate the input images as visible, infrared, and fused images simultaneously and output two-dimensional probabilities, so that the discriminator improves the ability to distinguish different image types and helps to better evaluate the authenticity and quality of the fused images. Through the design of the loss function, the discriminator constrains the generator and performs inverse parameter updates, aiding the generator in gradually optimizing the quality of the generated images during continuous adversarial training. This process leads to the generation of fused images.
The study addresses the issue of potential computational slowdown due to increased model depth by modifying the network structure through the inclusion of batch normalization, residual connections, and depth-wise separable convolution. This modification reduces the number of parameters and the computational complexity of the network, enhancing the model’s computational efficiency and speed. Consequently, it accelerates the training and inference processes of MMGAN, providing more timely and accurate environmental awareness and decision support for forest fire scenes.
The application of image fusion technology to the forest fire scene in our research yields significant practical benefits. By fusing images of the forest fire scene, forest firefighters can quickly understand the surrounding status of the fire, including fire expansion, smoke distribution, burning hot spots, and other key information, and can more accurately assess the danger level of the fire and make timely risk avoidance decisions based on changes in the fire environment. This is important for protecting the safety of the firefighters and improving the efficiency of fire rescue and helping reduce the casualties and property losses caused by forest fire accidents.
Current research on visible and infrared image fusion for forest fire scenes focuses on static images, while the changing situation of fires often requires the processing and analysis of continuous video data. In the future, research can delve deeper into applying visible and infrared image fusion techniques to forest fire scene video data using multi-level, multi-classification generative adversarial networks. This approach can enable more precise monitoring of fire dynamics by analyzing and fusing continuous video frames. It can also aid forest firefighters in making timely decisions and taking appropriate rescue actions, ultimately leading to a significant reduction in firefighter casualties due to forest fires.

5. Conclusions

To tackle the issue where traditional GANs often neglect visible contrast ratio information and detailed infrared texture details, this study introduced a multi-level and multi-classification GAN-based approach for fusing forest fire scene images from the visible and infrared spectra, and validated its accuracy. In terms of visual results, the fused images generated by MMGAN had a more significant contrast ratio and richer detailed texture information, as MMGAN could retain both the contrast ratio of the visible images and the detailed texture information of the infrared images. In terms of objective evaluation metrics, compared to the sub-optimal image fusion method, the evaluation results based on IEN, AG, FMI, and SSIM for MMGAN improved by 15.6%, 14.1%, 3.77%, and 5.47%, respectively, when applied to the TNO dataset; the evaluation results based on IEN, STD, AG, FMI, and SSIM improved by 10.23%, 0.383%, 7.46%, 0.82%, and 1.94%, respectively, when applied to the forest fire scene dataset. Furthermore, based on the same dataset, MMGAN’s image fusion speed was more than five times faster than that of the sub-optimal fusion method.
In conclusion, the proposed method has better interpretability, which means it can generate more reasonable fusion strategies self-adaptively based on input images. In addition, MMGAN can retain both rich textures in detail and a significant contrast ratio of source images, effectively avoiding the weakening of beneficial information during the fusion process. The model’s computational speed is optimized while enabling better performance in the fusion of visible and infrared images. Therefore, MMGAN image fusion technology applied to forest fire scenes can quickly provide more accurate and comprehensive information on forest fire surroundings for firefighters, which helps firefighters and disaster management personnel to better understand the real-time situation inside the fire, including the location of the fire source, the degree of combustion, and the expansion of the fire, so as to provide a more accurate and comprehensive description of the fire environment. It also serves as a scientific foundation for firefighters to precisely identify and locate the fire source and other critical targets.
Finally, MMGAN image fusion technology lays a vital foundation for enhancing the intelligence and automation of forest fire scenes by integrating deep learning with image fusion technology. In the future, with the continuous development of artificial intelligence and image processing technology, MMGAN image fusion technology is expected to play an increasingly important role in forest fire rescue and disaster management, and make greater contributions to people’s safety and property security.

Author Contributions

G.Z. and Z.Y. conceived and designed the study. Q.J. wrote the first draft and performed the data analysis. Q.J. collected all study data. S.T., H.X. and Y.W. provided critical insights in editing the manuscript. X.W. was responsible for project management. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation Project of China (Grant No. 32271879), the Science and Technology Innovation Platform and Talent Plan Project of Hunan Province (Grant No. 2017TP1022), the National Natural Science Foundation of China Youth Project (Grant No. 32201552) and Changsha City Natural Science Foundation (Grant No. kq2202274).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Feng, L.; Xiao, H.; Yang, Z.; Zhang, G. A Multiscale Normalization Method of a Mixed-Effects Model for Monitoring Forest Fires Using Multi-Sensor Data. Sustainability 2022, 14, 1139. [Google Scholar] [CrossRef]
  2. Martinez-de Dios, J.R.; Arrue, B.C.; Ollero, A.; Merino, L.; Gómez-Rodríguez, F. Computer vision techniques for forest fire perception. Image Vis. Comput. 2008, 26, 550–562. [Google Scholar] [CrossRef]
  3. Nemalidinne, S.M.; Gupta, D. Nonsubsampled contourlet domain visible and infrared image fusion framework for fire detection using pulse coupled neural network and spatial fuzzy clustering. Fire Saf. J. 2018, 101, 84–101. [Google Scholar] [CrossRef]
  4. Li, H.; Li, L.; Zhang, J. Multi-focus image fusion based on sparse feature matrix decomposition and morphological filtering. Opt. Commun. 2015, 342, 1–11. [Google Scholar] [CrossRef]
  5. Ma, J.; Fan, X.; Yang, S.X.; Zhang, X.; Zhu, X. Contrast Limited Adaptive Histogram Equalization Based Fusion in YIQ and HSI Color Spaces for Underwater Image Enhancement. Int. J. Pattern Recognit. Artif. Intell. 2017, 32, 1854018. [Google Scholar] [CrossRef]
  6. Xiang, Z.; Wu, H.; Zhou, D. Metallic debossed characters industrial online non-segmentation identification based on improved multi-scale image fusion enhancement and deep neural network. IET Image Process. 2022, 16, 852–868. [Google Scholar] [CrossRef]
  7. Wu, C.; Chen, L. Infrared and visible image fusion method of dual NSCT and PCNN. PLoS ONE 2020, 15, e0239535. [Google Scholar] [CrossRef]
  8. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  9. Quan, Y.; Bu, L.; Wu, W. Improved weighted average of threshold value of HSV and wavelet transform image fusion. J. Liaoning Tech. Univ. (Nat. Sci.) 2016, 35, 65–69. [Google Scholar]
  10. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  11. Zheng, Y.; Zhang, G.; Tan, S.; Yang, Z.; Wen, D.; Xiao, H. A forest fire smoke detection model combining convolutional neural network and vision transformer. Front. For. Glob. Chang. 2023, 6, 1136969. [Google Scholar] [CrossRef]
  12. Ma, J.; Wei, Y.; Liang, P.; Chang, L.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  13. Toet, A. The TNO Multiband Image Data Collection. Data Brief 2017, 15, 249–251. [Google Scholar] [CrossRef]
  14. Hasnat, A.; Halder, S.; Bhattacharjee, D.; Nasipuri, M. A Proposed Grayscale Face Image Colorization System using Particle Swarm Optimization. Int. J. Virtual Augment. Real. 2017, 1, 72–89. [Google Scholar] [CrossRef]
  15. Kanmani, M.; Narsimhan, V. An image contrast enhancement algorithm for grayscale images using particle swarm optimization. Multimed. Tools Appl. 2018, 77, 23371–23387. [Google Scholar] [CrossRef]
  16. Qian, Y.; Huang, Z.; Fang, H.; Zuo, Z. WGLFNets: Wavelet-based global-local filtering networks for image denoising with structure preservation. Opt. Z. Licht- Elektron. J. Light-Electronoptic 2022, 261, 169089. [Google Scholar] [CrossRef]
  17. Gao, L.; Qi, L.; Guan, L. The Property of Frequency Shift in 2D-FRFT Domain with Application to Image Encryption. IEEE Signal Process. Lett. 2021, 28, 185–189. [Google Scholar] [CrossRef]
  18. Gool, T. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar]
  19. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  20. Roberts, J.W.; Van Aardt, J.A.; Ahmed, F.B. Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J. Appl. Remote Sens. 2008, 2, 023522. [Google Scholar]
  21. Rao, Y.J. In-fibre Bragg grating sensors. Meas. Sci. Technol. 1997, 8, 355. [Google Scholar] [CrossRef]
  22. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  23. Li, S.; Yang, B. Multifocus image fusion using region segmentation and spatial frequency. Image Vis. Comput. 2008, 26, 971–979. [Google Scholar] [CrossRef]
  24. Qu, G.; Zhang, D.; Yan, P. Information measure for performance of image fusion. Electron. Lett. 2002, 38, 313–315. [Google Scholar] [CrossRef]
  25. Zhou, W.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar]
  26. Bavirisetti, D.P.; Dhuli, R. Fusion of Infrared and Visible Sensor Images Based on Anisotropic Diffusion and Karhunen-Loeve Transform. IEEE Sens. J. 2015, 16, 203–209. [Google Scholar] [CrossRef]
  27. Sasidharan, R.; Siji, P.; Davis, A.; Image Fusion with Guided Filtering. JETIR. 2015. Available online: https://api.semanticscholar.org/CorpusID:212503287 (accessed on 1 June 2023).
  28. Liu, Y.; Chen, X.; Cheng, J.; Peng, H.; Wang, Z. Infrared and visible image fusion with convolutional neural networks. Int. J. Wavelets Multiresolut. Inf. Process. 2018, 16, 1850018. [Google Scholar] [CrossRef]
  29. Ma, J.; Xu, H.; Jiang, J.; Mei, X.; Zhang, X.P. DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 2020, 29, 4980–4995. [Google Scholar] [CrossRef]
Figure 1. Network structure of generator.
Figure 1. Network structure of generator.
Forests 14 01952 g001
Figure 2. Network structure of discriminator.
Figure 2. Network structure of discriminator.
Forests 14 01952 g002
Figure 3. Framework of GAN model.
Figure 3. Framework of GAN model.
Forests 14 01952 g003
Figure 4. Framework of MMGAN model.
Figure 4. Framework of MMGAN model.
Forests 14 01952 g004
Figure 5. Fusion results of each model based on TNO dataset. (a) Group A visible and infrared fusion images; (b) Group B visible and infrared fusion images; (c) Group C visible and infrared fusion images. The red box shows an enlarged view of the green box detail.
Figure 5. Fusion results of each model based on TNO dataset. (a) Group A visible and infrared fusion images; (b) Group B visible and infrared fusion images; (c) Group C visible and infrared fusion images. The red box shows an enlarged view of the green box detail.
Forests 14 01952 g005aForests 14 01952 g005b
Figure 6. Fusion results of each model based on forest fire scene dataset. (a) Group A visible and infrared fusion images; (b) Group B visible and infrared fusion images; (c) Group C visible and infrared fusion images. The red box shows an enlarged view of the green box detail.
Figure 6. Fusion results of each model based on forest fire scene dataset. (a) Group A visible and infrared fusion images; (b) Group B visible and infrared fusion images; (c) Group C visible and infrared fusion images. The red box shows an enlarged view of the green box detail.
Forests 14 01952 g006aForests 14 01952 g006b
Table 1. Objective evaluation results for assessing the quality of fused images based on the TNO dataset.
Table 1. Objective evaluation results for assessing the quality of fused images based on the TNO dataset.
MetricIENSTDAGSFFMISSIM
Method
ADF6.75034.2596.83612.71012.2610.612
GFF6.78140.1477.85511.10413.3890.608
CNN7.10148.1136.65714.89514.0070.610
FusionGAN7.43151.5347.47215.34814.0990609
DDcGAN7.64252.0368.05917.89614.3250.622
MMGAN8.83051.2659.19817.76414.8660.656
Bold indicates the method of this paper; highlight bold indicates the optimal result.
Table 2. Objective evaluation results for assessing the quality of fused images based on the forest fire scene dataset.
Table 2. Objective evaluation results for assessing the quality of fused images based on the forest fire scene dataset.
MetricIENSTDAGSFFMISSIM
Method
ADF6.25935.7447.02514.25912.5170.612
GFF6.75335.1577.72415.32213.270.606
CNN6.45233.517.64713.86512.9030.664
FusionGAN7.73050.9167.57215.76514.1700.638
DDcGAN7.84952.4738.98317.60014.8360.669
MMGAN8.86252.6749.65217.34714.9580.680
Bold indicates the method of this paper; highlight bold indicates the optimal result.
Table 3. Average running time(s) of each model based on each dataset.
Table 3. Average running time(s) of each model based on each dataset.
MethodForest Fire SceneTNO
ADF9.4767.456
GFF5.3023.259
CNN0.5620.499
FusionGAN0.1960.360
DDcGAN0.6130.613
MMGAN0.1430.066
Bold indicates the method of this paper; highlight bold indicates the optimal result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, Q.; Tan, S.; Zhang, G.; Yang, Z.; Wen, Y.; Xiao, H.; Wu, X. Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints. Forests 2023, 14, 1952. https://doi.org/10.3390/f14101952

AMA Style

Jin Q, Tan S, Zhang G, Yang Z, Wen Y, Xiao H, Wu X. Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints. Forests. 2023; 14(10):1952. https://doi.org/10.3390/f14101952

Chicago/Turabian Style

Jin, Qi, Sanqing Tan, Gui Zhang, Zhigao Yang, Yijun Wen, Huashun Xiao, and Xin Wu. 2023. "Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints" Forests 14, no. 10: 1952. https://doi.org/10.3390/f14101952

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop