Next Article in Journal
Volatile Compound Chemistry and Insect Herbivory: Pinus edulis Engelm. (Pinaceae) Seed Cone Resin
Previous Article in Journal
The Relationship between Stand Structure and Tree Growth Form—Investigating the Effects of Selection Cuttings in Mountainous Mixed Beech Forests
Previous Article in Special Issue
UAV Multispectral Imagery Predicts Dead Fuel Moisture Content
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Defogging Learning Based on an Improved DeepLabV3+ Model for Accurate Foggy Forest Fire Segmentation

1
College of Information Science and Technology, Nanjing Forestry University, Nanjing 210037, China
2
School of Computer Science and Engineering, Changshu Institute of Technology, Changshu 215500, China
*
Author to whom correspondence should be addressed.
Forests 2023, 14(9), 1859; https://doi.org/10.3390/f14091859
Submission received: 28 June 2023 / Revised: 3 August 2023 / Accepted: 8 September 2023 / Published: 13 September 2023

Abstract

:
In recent years, the utilization of deep learning for forest fire detection has yielded favorable outcomes. Nevertheless, the accurate segmentation of forest fires in foggy surroundings with limited visibility remains a formidable obstacle. To overcome this challenge, a collaborative defogging learning framework, known as Defog DeepLabV3+, predicated on an enhanced DeepLabV3+ model is presented. Improved learning and precise flame segmentation are accomplished by merging the defogging features produced by the defogging branch in the input image. Furthermore, dual fusion attention residual feature attention (DARA) is proposed to enhance the extraction of flame-related features. The FFLAD dataset was developed given the scarcity of specifically tailored datasets for flame recognition in foggy environments. The experimental findings attest to the efficacy of our model, with a Mean Precision Accuracy (mPA) of 94.26%, a mean recall (mRecall) of 94.04%, and a mean intersection over union (mIoU) of 89.51%. These results demonstrate improvements of 2.99%, 3.89%, and 5.22% respectively. The findings reveal that the suggested model exhibits exceptional accuracy in foggy conditions, surpassing other existing models across all evaluation metrics.

1. Introduction

Forests are one of the most vital ecosystems on Earth, providing essential ecosystem services such as oxygen and ecological balance [1], as well as serving as habitat and food sources for animals and plants. In addition, forests are important resources for society due to mitigating climate change and reducing carbon dioxide emissions [2]. However, forest fires represent a significant risk to these ecosystems, causing extensive damage and environmental losses globally. With over 35,000 forest fires occurring annually, millions of hectares of forest are consumed, leading to significant economic and environmental impacts [3,4,5]. Therefore, effective monitoring and evaluation of forest fires are crucial for us to minimize the loss of natural resources.
Traditional methods for detecting forest fires, such as manual monitoring, thermal imaging, and radio detection, often have low identification rates and require significant time and effort [6]. While deep neural networks have made substantial advances in forest fire detection [7], detecting forest fires in hazy images remains a challenging task. Foggy conditions can cause images to become blurry, making it more difficult to distinguish the forms, ranges, and spectra of fires [8,9]. More research is needed to create new algorithms and strategies to increase the accuracy and usefulness of forest fire recognition in foggy forests.
Defogging image processing has seen significant advancements in recent years, driven in part by the use of the physical scattering model to formulate the fogging process [10,11]. The model describes the interaction of light with atmospheric particles to simulate the scattering and absorption of light in foggy conditions.
Two types of single-picture defogging techniques have emerged in the last decade. The first category is based on foggy priors, such as dark channels [12], color attenuation [13], and haze lines [14]. However, these methods frequently rely on complicated statistical hypotheses and physical models that may not precisely assess physical characteristics and apply to various situation [15]. The second group includes techniques based on deep learning, including AOD-Net [16], which recovers images by reformulating the physical scattering mode, end-to-end networks designed by Cai et al. [17] and Chen et al. [18], which do not rely on foggy priors, and the DCPDN model suggested by Zhang et al. [19]. Dong et al. [20] also incorporated enhancement strategies into their network to gradually recover clean images. These methods have demonstrated remarkable performance in defogging.
More and more deep learning techniques are being used to solve the issue of forest fires in hazy photos. For example, Yang et al. [21] designed a smoke detection model that leverages dark channel-assisted hybrid attention, while Merve Balki Tas et al. [22] created a smoke detection idea specifically for foggy wildfires. Huang et al. [23] also designed a GXTD detection model with a defogging function. These approaches, however, may have drawbacks such as indistinct edges, poor adaptation to fluctuating fog concentrations, and a high false alarm rate. On the other hand, there have been effective solutions developed for defogging segmentation of other objects. For instance, IRDCLNet provides a fog segmentation solution for ships that uses an interference reduction module and a dynamic contour learning module [24]. Another illustration is the work of Zhu et al., which enhances the defogging technique by using a dual attention mechanism and an SOS acceleration module [25]. He et al. [26] also propose using residuals and attention mechanisms to address the problem of different fog concentrations. These methods have shown promising results in their respective domains. It is also worth mentioning Chen et al.’s semi-supervised joint defogging learning framework, which effectively identifies vehicles in foggy weather by fusing feature layers [27]. This approach has demonstrated superior performance in accurately recognizing vehicles in foggy conditions.
In our study, we present an improved joint defogging semantic segmentation model named Defog DeepLabV3+. It is built upon DeepLabV3+ [28] and is designed to enhance the stability and accuracy of the segmentation of forest fires. A branch for dehazing and a branch for segmenting forest fires make up the suggested model. The dehazing branch uses pyramid feature enhancement to further hone the dehazed image and a two-stage feature extraction approach to improve the quality of the input image. On the other hand, the forest fire segmentation branch utilizes clean image dehazing features acquired from the dehazing branch. The segmentation branch integrates Atrous Spatial Pyramid Pooling (ASPP) [29] to gather contextual data at multiple scales, and an attention mechanism to increase the features’ ability to discriminate, thus achieving better forest fire segmentation results. By sharing fused features, the model enhances the defogging and segmentation modules and aims to enhance the overall effectiveness of detecting forest fires in cloudy photos.
The following are the primary contributions of this paper:
  • We design a smart attention mechanism module in our model, consisting of two independent branches that use different attention strategies. The module also uses residual structures to enhance the model’s feature representation capabilities.
  • Our model proposes a new training framework that unifies defogging and segmentation networks. The joint defogging learning framework preserves defogging features for forest fire segmentation, enabling the model to cope with poor visibility.
  • Because no dataset exists specifically for foggy weather forest fire segmentation, we restructured existing benchmarks and created the Foggy FLAME Dataset (FFLAD). This dataset is intended to aid research in this area by serving as a baseline for evaluating the efficacy of defogging and segmentation models in forest fire detection under low visibility situations.
  • Our experimental findings show that our Defog DeepLabV3+ model outperforms several existing CNN-based algorithms in the segmentation of foggy forest fire instances.
This work is organized as follows. Section 2 describes the experimental material and methods for detecting fogged forest fires, including the joint training framework of the defogging network and the segmentation network. The experimental data and analyses are presented in Section 3. The debate is presented in Section 4. Section 5 concludes with some last thoughts.

2. Materials and Procedures

This paragraph describes the FFLAD dataset and production procedure. We also designed a joint training model for the defogging and segmentation networks called Defog DeepLabV3+. The first challenge addressed in this model is deep feature defogging. The second task is forest fire semantic segmentation, with an emphasis on improving forest fire segmentation accuracy.

2.1. Dataset

2.1.1. Data Source

Because of the scarcity of fogged forest fire data and the scarcity of mature fogged forest fire datasets, we created a comprehensive dataset called FFLAD by selecting suitable forest fire detection datasets that were applicable for fogging. Our data sources include the FLAME dataset [30], the Bowfire dataset [31], and the ERA dataset [32]. The FLAME dataset consists of fire aerial images of combusting waste in pine forests in Arizona. The Bowfire dataset contains numerous negative samples that can be easily mistaken for flames, including forest fire images in various scenarios. We extracted forest fire scenes from multiple unmanned aerial vehicle (UAV) videos in the ERA dataset to construct our dataset. These high-quality datasets have been annotated by researchers and serve as a suitable foundation for our Fog Flame Dataset (FFLAD). Figure 1 showcases some representative images.
In addition, we obtained forest fire photos from Internet search engines as an additional test set to assess the generalization performance of our model. The selected fire images had diverse scenes and contained many interfering objects.
The dataset consisted of foggy and clear images in RGB format with a bit depth of 24. Each image had a resolution of 3840 × 2160 or 1024 × 680 pixels and a horizontal and vertical resolution of 96 dpi. The input image size for the model was set to 512 × 512 pixels. The label images were grayscale images with a bit depth of 8, where the label for the fire had a pixel value of 1. The size of the label images was consistent with that of the input images. The dataset had a size of 16.54 GB.

2.1.2. Data Fogging and Enhancement Processing

For our experiments, we utilized the FLAME dataset and a subset of the Bowfire dataset as the original dataset and preprocessed them as needed. Specifically, we created a depth map for each clear map using the Monodepth2 self-supervised depth estimation model [33]. In the case of a homogeneous atmosphere, the single-point transmittance (t) and the scene depth information (d) are related by the scattering factor in the atmosphere (β), according to the following equation [12,34]:
t ( x ) = e β d ( x )
Equation (1) shows that the amount of radiation in the scene decreases exponentially as distance increases, where t is less than or equal to 1. To model the effect of atmospheric scattering, we used the equation described by reference [10,11]:
I ( x ) = J ( x ) t ( x ) + A ( 1 t ( x ) )
J ( x ) = I ( x ) A t ( x ) + A
where I(x) represents the observed blurred image, J(x) is the scene brightness to be recovered, t(x) is the global atmospheric light and transmission maps, and A represents the atmospheric light.
I ( x ) = J ( x ) e β d ( x ) + A ( 1 e β d ( x ) )
We generated the fog map using Equation (4), which is derived from Equations (1) and (2). In Equation (4), the constant A represents atmospheric light intensity and may vary depending on fog concentration and weather conditions. Typically, A falls within the range of 10 to 100 [11]. The fog maps are saved in “.jpg” format with the same dimensions as the original image. Ultimately, the FFLAD dataset comprises original images, two types of depth maps, fog maps, and label maps (Figure 2).
There were 5001 total photos in our dataset, which were split into three groups: a training set, a validation set, and a test set. In deep learning segmentation research on forest fires, the 80/10/10 data split is a commonly used practice, and we also adopted this approach. In addition, we conducted comparative experiments using data splits of 90/5/5, 70/15/15, and 70/20/10. The results showed that compared to the 80/10/10 data split, the other data splits led to overfitting, where the model performed well on the training set but poorly on the test set. The model performed well on both the training and validation sets, and its performance on the test set was similarly dependable, thanks to the very balanced 80/10/10 data split. These sets account for 80%, 10%, and 10% of the dataset, respectively (Table 1).
In addition to DeepLabV3+’s standard data enhancement techniques of random scaling, random cropping, and random color perturbation, our model employs additional data enhancement techniques such as random Gaussian blur, random rotation, and random color gamut translation. These techniques significantly enhance the model’s performance and resilience, with only marginal overhead increases [35].

2.1.3. Model Training

The hyperparameter choices for applying DeepLabV3+ for image segmentation that impacts the model are the backbone network selection, batch size, image size, initial learning rate, decoder, number of iterations, and optimizer strategy. Based on DeepLabV3+ experiments, our hyperparameter selection strategy in the experiment is shown in Table 2:

2.2. Defog DeepLabV3+ Architecture

2.2.1. DeepLabV3+

The DeepLabV3+ network architecture is designed for semantic segmentation tasks. It has an encoder–decoder structure that leverages DeepLabV3 as the encoder to obtain detailed contextual information. As an improvement in the DeepLab series, DeepLabV3 inherits the overall sequential architecture of BackBone and DeepLab Head [29,36,37]. The Backbone is used to extract deep image features, and ResNet is one of the commonly used choices due to its strong feature extraction capability brought by the residual structure. The DeepLab Head is used to map the features extracted by the Backbone to pixel-level segmentation results, including the Atrous Spatial Pyramid Pooling (ASPP) module and decoder module. The ASPP module is a core component of DeepLabv3, which uses atrous convolutions with variable sampling rates to handle multi-scale features to get various scales’ information. In addition, the ASPP module also includes a global pooling layer for handling global information of the entire image, thereby better capturing background information. The decoder module restores the feature map to the original image resolution through upsampling and deconvolution operations. However, ASPP itself introduces noise, and the large convolution kernels of atrous convolution cause the weight values to disperse, making the model sensitive to noise.
In DeepLabV3+, the decoder module is responsible for recovering object boundary information and extracting features from the encoder using an extended convolution of arbitrary resolution. Additionally, the encoder’s backbone network employs a deeply separable convolution strategy [38] and an improved initial architecture [39] to reduce computational costs while maintaining accuracy.
The DeepLabV3+ network integrates the ASPP module in addition to the encoder–decoder structure to collect multi-scale contextual data. The ASPP module extracts features at various scales while reducing the number of parameters by using atrous convolutions with various dilation rates [29]. This makes it possible for the network to efficiently gather both local and global contextual data, which are crucial for tasks requiring semantic segmentation.
Figure 3 describes the network architecture of DeepLabV3+. By combining the encoder–decoder structure, deeply separable convolutions, an improved initial architecture, and the ASPP module, DeepLabV3+ has displayed cutting-edge performance in semantic segmentation tasks on several benchmark datasets. This makes the architecture a powerful tool for various applications that require accurate semantic segmentation.
For image segmentation studies of forest fires, effective advanced models have been proven to include UNet, DeepLabV3+, FCN, PSPNet, and the recently proposed MaskSU R-CNN [40]. However, UNet typically loses a lot of spatial information during downsampling, resulting in a small receptive field. Additionally, UNet only considers local context and has difficulty capturing global information, making it prone to overfitting. FCN’s redundant fully convolutional structure has high memory consumption and results in blurred object boundaries in the segmentation output, without considering context [41]. While PSPNet to some extent overcomes the insufficiency of context information with its pyramid pooling structure, the shallow feature fusion is insufficient, resulting in poor segmentation details [42]. MaskSU R-CNN well inherits the high accuracy of MS R-CNN, but the model design is more complex and the scale is larger, with poor real-time performance [40]. In comparison, the DeepLabV3+ model can effectively enlarge the receptive field and collect the global information through atrous convolutions and the ASPP module, thereby performing excellently in segmentation details. DeepLabV3+ also controls the resolution of the segmentation output by changing the parameters of atrous convolution, thereby improving the segmentation speed while ensuring accuracy. Additionally, the DeepLabV3+ encoder–decoder structure achieves cross-layer feature passing, which is conducive to the joint design and use of the defogging framework.
Related studies tested the first 4 models’ performance on the FLAME dataset [43], with UNet and DeepLabV3+ performing similarly with an accuracy of around 91%, while FCN and PSPNet achieved relatively lower accuracy at 85% and around 91% respectively. This once again validates the rationality of our model selection. Specifically, DeepLabV3+ achieved the highest mean IoU on the FLAME dataset, outperforming UNet, FCN, and PSPNet. The better performance can be attributed to DeepLabV3+’s ability to capture multi-scale contextual information through the ASPP module and effective decoder design. Compared to UNet’s simple encoder–decoder architecture, DeepLabV3+ exploits richer semantic information from inputs through atrous separable convolutions and global context modeling, enabling more precise segmentation of flames and flame-like objects in forest fire images.

2.2.2. Defog DeepLabV3+

As described in Figure 4, our architecture comprises two main branches: the forest fire segmentation branch and the defogging branch. The collective feature sharing module (CFSM), which is shared by both branches during training, makes sure that the fog-free features produced by this module may be applied to either branch of our joint defogging learning architecture. In the fog removal branch, a two-stage recover module further processes the obtained depth information into dehazing feature and the pyramid enhancement block generates the final fog-free results. In the forest fire division branch, the deep feature module is composed of the designed fusion attention residual module and ASPP. ASPP is the key module designed in DeepLabV3+. During inference, only the CFSM and segmentation modules need to be executed. This architecture significantly improves the performance of forest fire segmentation in foggy weather without adding a computational burden during the inference phase.

2.3. Defogging Branch

The forest fire segmentation branch performs better in foggy conditions thanks to the defogging branch, which is intended to improve the quality of the common features extracted by the CFSM. This branch uses the CFSM and the module for fog-free picture reconstruction to achieve its goal.

2.3.1. Shared Fused Feature Extraction Network

The CFSM is designed for extracting features from the input images that contain crucial information for joint learning of fog removal and forest fire segmentation. To keep the architecture simple and avoid adding computational burden to the network, the CFSM was designed based on the DCNN part of the module in the DeepLabV3+ network. While the features derived from the deeper layers of the network contain more high-level information, the features retrieved from the shallower levels of the network contain more spatial and low-level information that assists in the defogging process [20]. As a result, the segmentation branch of the proposed CFSM has shallow convolutional layers. The fog-free picture reconstruction module receives the characteristics that the CFSM retrieved after performing defogging.

2.3.2. Fog-Free Image Reconstruction Module

The features extracted by the CFSM called FC may be degraded by fog, resulting in a limited performance of forest fire segmentation. To reconstruct the FC shared with the segmentation branch during joint learning, we propose the fog-free image reconstruction module FIRM, which is composed of the following architecture. First, the extracted features FC are subjected to a convolutional and two residual blocks in order to obtain more precise characteristics. Secondly, a deconvolution operation is carried out twice to upsample these features to match the input resolution as the dimensionality of these features is lowered in the preceding layers. The upsampled features are then concatenated with the input image and passed through the pyramid enhancement module (PEM) to produce the final, fog-free outcome. The PEM module can extend the network’s characterization capability by extracting features based on many learning scales and various receptive areas. Our operation is based on pyramid pooling [44].

2.4. Image Segmentation Branch

For the forest fire segmentation branch, we utilized DeepLabV3+ as the backbone, with the DCNN shallow convolutional layer serving as the CFSM that is used for feature extraction by the defogging branch. However, the segmentation performance is suboptimal because the original ASPP module in DeepLabV3+ cannot obtain sufficient contextual information when dealing with smaller targets. To overcome this challenge, we designed a novel attention mechanism called DARA, which was connected in parallel to the original ASPP module. This design greatly enhanced the performance of our product.

2.4.1. Dual-Attention Residual Module

Our dual-fusion attention residual feature attention mechanism (DARA) first passes the depth features through two dilated convolutions to create new incremental depth features with different centers of gravity in the position attention [45] and channel attention [46] modules.
Obtaining a distinguishing feature representation is crucial for semantic segmentation, and that may be performed by accumulating long-term contextual data [36]. The location–attention module encodes rich contextual information into local features, which are represented as the correlation influence matrix between pixel points, thereby improving the local feature representation. The location attention module’s and the computational equations’ final output features are listed as follows:
S j i = exp ( B i T · C j ) i = 1 N exp ( B i T · C j ) ( S R N × N )
E j 1 = α i = 1 N S j i D i + A j ( E 1 R C × H × W )
where is the deeper feature B , C R C × N , D R C × H × W obtained by convolving the input deep feature; S j i is a measurement of how the pixel at the place I affects position j and N = H × W .
In computer vision, high-level semantic features can be represented as channel graphs, which are specific to each object class. However, these semantic responses are not independent of each other. To address this, we suggest a channel attention module that simulates the relationships between different channels. This module uses a channel attention feature map to represent the relationships between different semantic responses. The final output features of the channel attention module are denoted as X R C × C and can be computed using the following equations:
X j i = exp ( A i · A j ) i = 1 C exp ( A i · A j )
E j 2 = β i = 1 C ( X j i A i ) + A j
where the deeper features A R C × H × W obtained by the convolution of the input depth features X j i represent the influence of channel I on channel j. Our final feature map E R C × H × W is described as:
E j = E j 1 + E j 2 + A j
To improve the characterization of raw features in a semantic segmentation task, we propose to use both raw depth features and learned depth features. We fuse these features and incorporate them into the DeepLabV3+ network architecture. Specifically, we connect the fused depth features to the encoder in parallel with the ASPP module, which improves the network’s capacity for feature identification. Additionally, this inclusion of depth features enriches the long-term contextual semantic information captured by the network. Figure 5 shows the enhanced network design.

2.4.2. Loss Function

In the training stage, the loss function curve gives a graphic representation of the model’s convergence. For Defog DeepLabV3+, the total loss is composed of two aspects. One is the loss of the defogging branch, which removes the fog from foggy images. The other is the loss generated by the segmentation branch, which splits the input image into patches for processing. The total loss can be expressed as the sum of these two components. Specifically, the mathematical equation for the total loss is:
L t o t a l = L D F + L S G
The aim of minimizing this loss during training is to improve the performance of the Defog DeepLabV3+ model on the task of defogging. The defogging branch’s loss function may be represented as follows:
L D F = 1 Q i = 1 Q J i J i G T 2
where 2 denotes the l2 parametric number. J i and J i G T denote the predicted fog-free image I and the corresponding ground truth values in a batch, respectively.
The split branch’s loss function may be stated as follows:
L S G = 1 N j n y j ln G α ( X j Z ) + ( 1 y j ) ln ( 1 G α ( X j Z ) )
where G α ( X j Z ) is the label probability with ai pixel j and is the GT label.

2.5. The Process of Working with the Model

During the model training phase, it is necessary to run both the dehazing branch and the image segmentation branch simultaneously. However, during the forest fire detection phase, the model only needs to run the image segmentation branch. This working mode ensures the quality of the dehazing segmentation while significantly improving the efficiency of model operation (Figure 6).

3. Results

As described in this section, we tested our proposed deep learning framework for defogging and forest fire segmentation in various conditions. The hardware consists of an Intel(R) Xeon(R) Gold 6330 CPU and an NVIDIA RTX 3090 GPU with 24 GB of RAM (Intel and NVIDIA Corporation, Santa Clara, CA, USA).

3.1. Defogging Branch

3.1.1. Evaluation of the Effectiveness of Defogging

To verify the usefulness of our defogging module in boosting our model’s segmentation performance, we conducted the experiment comparing our Defog DeepLabV3+ to a baseline segmentation model called BaseLine, which was trained without the defogging branch. We used two different backbone models, mobilenetv2 [38] and xception [39] in our experiments. Furthermore, we tested the impact of two additional modules on our model’s performance: the DARA module and the PEM module. To ensure fairness in our experiment, we used mean pixel accuracy (mPA) as the evaluation metric. The total number of properly categorized pixels is divided by the total number of pixels in the image to compute mPA.
P A = i = 0 k p i i i = 0 k j = 0 k p i j
where P i j Indicates the jth pixel of row. This metric is calculated by dividing the total number of pixels in the image by the number of pixels that have been properly classified.
In Table 3 and Figure 7, our experiment shows that incorporating the defogging branch into the deep learning framework can greatly enhance segmentation performance, especially for the mobilenetv2+ architecture. By adding the defogging branch, the mPA increases to 93.33%. Furthermore, incorporating the pyramid enhancement module PEM into the defogging branch further improves the mPA to 94.26%. These results highlight the effectiveness of our framework in improving segmentation accuracy, especially in challenging foggy environments.

3.1.2. Visualization Analysis

To gain a more intuitive understanding of the effectiveness of our defogging module, we selected the first five images from our dataset to showcase the defogging effect of our model under different scenarios, fog concentrations, and fire sizes. For each image, we arranged the fog map, defogging effect map, and original map for reference.
Figure 8 presents the results of our experiment, which proves that the defogging module can effectively adapt to the changing scenarios of forest fire defogging. Our model can clearly reveal the flames of various shapes and sizes from the fog, and the comparison with the original image still shows strong recognition. These results highlight the ability of our proposed framework to improve the visibility of foggy images.

3.2. Forest Fire Segmentation

3.2.1. Assessment Indicators

In addition to the mean pixel accuracy (mPA) metric, we also evaluated the performance of our Defog DeepLabV3+ model using the intersection over union (IoU) metric. This statistic calculates the percentage of overlap between anticipated segmentation results and their related ground truth labels. A value closer to 1 indicates better segmentation performance.
To ensure fairness in our evaluation, we computed the mean IoU to quantify our model. The equation for computing mean IoU is as follows:
m I o U = 1 N i = 1 N P i G i P i G i
where Pi and Gi denote the predicted result of the i-th image and the corresponding ground truth label, respectively. This metric provides a more comprehensive evaluation of our model’s segmentation performance, which is critical for applications like forest fire detection and prevention.
In our experiments, the target is considered a positive sample if the IoU is 0.5 or above, and a negative otherwise. In addition, an additional metric from recall can be used, denoted as:
Re c a l l = T P T P + F P
where TP represents true positive cases and FN represents false negative cases. It represents the ratio of right forecasts to total positive case predictions.
To ensure fairness in our evaluation, we computed the mean recall on the test set as another measure of our model’s performance. The equation for computing mean recall is similar to that of mean IoU. This metric provides a complementary evaluation of our model’s segmentation performance, as it focuses on the ability of our model to correctly identify positive cases.

3.2.2. Performance Assessment

The enhanced forest fire semantic segmentation model, built upon DeepLabV3+, integrates defogging optimization to generate defogging maps and forest fire segmentation simultaneously. In Figure 9, the green box highlights the area with flames, which is the target for segmentation. Image a shows the original input images used in the foggy forest fire segmentation task. The experimental results highlighted in red in Image b demonstrate the commendable performance of the model. When compared with the ground truth labels in Figure 9c, we observe that our model can accurately capture small, hidden small flames, such as flame areas 2, 3, 5 and 6. The flame edge segmentation is also very detailed, such as flame areas 1, 4, 7 and 8.
In the visual representation of images, flames have the characteristics of high luminance and high contrast [47]. This is similar to lamps, the sun, etc. Our model should be able to accurately identify flames and distinguish these flame-like objects. In Figure 10, Image a from top to bottom is lanterns, the sun, the chandelier, and the fire hydrant. Although our model has a very small amount of misjudged pixels in Image b, such as area 1, 2 and 3, the similar objects are well distinguished from the flames. Specifically, our model achieves a high precision of 99.7% for flame segmentation, demonstrating superior performance in distinguishing flames from similar objects.
Our experiments used five-fold cross-validation, which is consistent with the data splitting ratio we agreed upon. First, the dataset was randomly divided into five equal-sized folds. Then, for each fold i, the i-th fold was used as the test set, and the other four folds were used as the training set. In this way, five models were obtained, and each model used different training and test sets. Table 4 and Figure 11 present the experimental findings. All models achieved high performance in terms of mPA, mIoU, and mRecall, with an average performance of 94.23%, 89.51%, and 94.04%, respectively. Furthermore, the model’s performance on the two test sets was not much different. This demonstrates the strong generalization performance of our model.
The confusion matrix for the model 1 test set is shown in Figure 12, and the ROC curve drawn by moving the threshold is shown in Figure 13, with the AUC calculated as 0.97 (±0.015) according to the curve. The model performs well with strong classification ability.
Furthermore, Figure 14 shows the loss curves of our model during training for 100 epochs on the training and validation sets. It reflects that in model 1, the performance of the test set and the training set change with iterations. The overall smooth and downward trend, and gradually converge after about 80 epochs. This demonstrates how well our model can gradually improve the segmentation results by learning the properties of the input image.
To prove the superior performance of the Defog DeepLabV3+ model for forest fire segmentation, we compared our model with many cutting-edge deep-learning-based models, such as UNet, PSPNet [42], and DeepLabV3+. We used the FFLAD dataset and configuration to train all models to guarantee it to be persuasive.
We chose four representative images and display them in Figure 15. Most models produced relatively accurate segmentations for obvious targets that were distinguishable from the surrounding area and had low haze. However, for forest fires with small targets or high haze, shown with the green boxes in Figure 15a,b, most models had a little under-segmentation, except UNet-agg, DeepLabv3+, and our Defog DeepLabV3+. For small and hidden flames such as flame areas 2, 3, and 5, PSPNet could not correctly identify them. Image h of flame area 5 shows that DeepLabV3+ is not very fine-grained for flame edges. Both PSPNet and DeepLabV3+ have serious false detections (marked in blue boxes). Our model outperformed other segmentation models, especially on images with large haze and complex scenarios.
Table 5 and Figure 16 present the quantitative analysis of the compared models, where evaluation metrics such as mPA, mIoU, and mRecall were used to assess their performance. Our proposed model achieved the second-highest mPA and mIoU scores among the compared models, while also achieving the highest mRecall score in both of the test sets. These results suggest that the proposed model is able to accurately identify objects in the dataset, while also capturing the overall structure and context of the scene. Overall, the proposed model shows promise as an effective approach for semantic segmentation on this particular dataset. The combination of high mPA, mIoU, and mRecall scores indicates that the proposed model can achieve both accurate and comprehensive segmentation results, making it a valuable tool for applications that require precise object detection and segmentation.
To demonstrate the importance of introducing the attention mechanism in the segmentation branch of our model, we conducted a simple but necessary ablation experiment. Table 6 displays the experiment’s outcomes, and they indicate that the introduction of the Dense Attention Refinement Module (DARA) significantly improves the mean intersection over union (mIoU) by about 1.5% in both cases. And Figure 17 fully demonstrates the superiority of Model 3. This suggests that by carefully exploiting the inherent information of the features, our attention strategy may improve model accuracy.
In addition, we analyzed the importance of different pixel features after adding the DARA module through a heatmap. As shown in the Figure 18, Figure 18b was generated by the model and it shows the result of combining the feature map with the original image. The color of the heatmap represents the importance of different regions of the image to the model’s judgment. In this image, we can see that some areas are red or yellow, which means that these areas play an important role in the model and have a greater impact on the final prediction results. And some areas appear blue, which means that these areas have less influence on the final prediction results. The flame feature is more important than other parts in our model. This suggests that the DARA module can effectively capture the flame features. This analysis provides valuable insights for further improving the model and optimizing the segmentation results.
We estimated the model size and inference time of our proposed framework in addition to analyzing the segmentation performance of the comparable approaches. Compared to the original DeepLabV3+ model, our improved model (Defog DeepLabV3+) is slightly larger in terms of the total number of parameters, with an increase of about 46%. This increase is mainly due to the addition of a new defogging branch in our model.
In Table 7 and Figure 19, we can see that despite the increase in model size, the inference time of our improved model-B is almost the same as that of DeepLabV3+. This is because only the computation of the split branch is involved in the inference process. Therefore, the modest increase in computational overhead is worth the improvement in segmentation performance achieved by our proposed framework.
The evaluation of both model size and inference time is important for practical applications, as it provides information about the efficiency and feasibility of deploying the model in resource-constrained environments. Our evaluation’s findings show that the suggested structure successfully strikes a compromise between computational effectiveness and the efficiency of segmentation. This is essential for real-world applications like forest fire detection and prevention.
Finally, we performed sensitivity analysis on the model input parameters, including the batch size, initial learning rate, and number of iterations. The model performance was evaluated by the loss function on the test set, as shown in Figure 20. As we can see from the figure, the batch size fluctuated around 0.07 with a range of 0.01, the initial learning rate was between 0.07 and 0.15, and the number of iterations converged to 0.073. It can be observed that these three parameters had little sensitivity to the model performance.

4. Discussion

Forest fires are complex and dynamic events that constantly change over time, making it difficult to accurately capture and describe features such as flames and smoke. The shape of the fire is also altered by the terrain, further complicating the detection and segmentation of forest fires [48]. Additionally, foggy weather reduces visibility and change the color and brightness of images, further challenging the extraction of detailed information from images [11].
To address these challenges, determining the extent of forest fires on foggy days has become an important and challenging problem that needs to be solved. In previous research on defogging semantic segmentation, many methods have adopted a two-stage processing strategy, where defogging image extraction is followed by semantic segmentation [24,25]. However, this approach may result in information loss, as important semantic information may be lost during the defogging image extraction process. Furthermore, the two-stage processing strategy requires the application of two models in the inference stage, which may require more data and computational resources.
In our study, we propose a new joint defogging method that improves the existing semantic segmentation model. Our approach combines two tasks of defogging with forest fire segmentation and employs a smart attention mechanism (DARA) to improve the display of feature channels. We investigated clean basic information and deep complex meanings to accomplish accurate forest fire segmentation. To assess the efficacy of our suggested technique, we first examined the joint defogging branch and showed that our approach was significantly effective. We then compared our model with existing state-of-the-art semantic segmentation models, which use traditional defogging methods. Our results demonstrate that our model outperformed the comparison methods in both recall and mIoU metrics, indicating that our model had good performance in solving the forest fire detection problem in foggy weather.
We also examined the effectiveness of the DARA attention module and showed that it was effective in extracting depth image features. Additionally, we compared our proposed Defog DeepLabV3+ model with the original DeepLabV3+ model and showed that our method delivered a modest improvement with an acceptable increase in overhead.
Compared to the previous multi-stage processing strategies adopted by researchers, such as the GXTD detection model with defogging capabilities, the GXTD detection model improves image quality and fire detection accuracy by removing haze from images [23]. However, this approach requires additional computational cost, which may result in higher execution time, especially when processing a large number of images. On the other hand, the smoke detection model that utilizes dark channel-aided mixed attention does not require additional computational cost, but the performance of the model is much inferior to ours due to the prior-based approach [21].
Despite the high accuracy achieved by our Defog DeepLabV3+ model in fogged forest fire segmentation, there are still some shortcomings that need to be addressed. In further research, we plan to streamline the learning module for deep segmentation features and explore semi-supervised training schemes to bridge the domain gap between real and synthetic datasets [27]. Overall, our proposed method shows promise in improving the accuracy and robustness of forest fire detection and prevention, which is crucial for protecting our environment and communities.

5. Conclusions

In this study on precise segmentation of forest fires in foggy weather, our main achievements were as follows:
  • We proposed the Defog DeepLabV3+ joint defogging and forest fire segmentation model to address foggy weather challenges in detecting forest fires.
  • We designed the DARA attention mechanism to enhance feature channel representation and improve forest fire segmentation accuracy.
  • We constructed the FFLAD dataset with synthetic and real foggy images to train our model.
  • Our experimental results show that Defog DeepLabV3+ outperforms state-of-the-art methods with 94.26% accuracy, 94.04% recall, and 89.51% mIoU.
Each component of our proposed model contributes to the overall performance, indicating the potential to improve forest fire detection and prevention in foggy weather. This is crucial for environmental and community protection.

Author Contributions

T.L. devised the programs and drafted the initial manuscript. W.C. contributed to the writing and the experiments. X.L. and Y.M. helped with the data collection and data analysis. J.H. helped to improve the manuscript and modified the model in the later stage. D.G. revised the initial manuscript. J.X. designed the project and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of Jiangsu Province (BK20211357), the Qing Lan Project of Jiangsu Province.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Raihan, A.; Begum, R.A.; Said, M.N.M. A meta-analysis of the economic value of forest carbon stock. Geogr. Malays. J. Soc. Space 2021, 17, 321–338. [Google Scholar] [CrossRef]
  2. Xu, Y.; Xiao, F. Assessing Changes in the Value of Forest Ecosystem Services in Response to Climate Change in China. Sustainability 2022, 14, 4773. [Google Scholar] [CrossRef]
  3. Boer, M.M.; Resco de Dios, V.; Bradstock, R.A. Unprecedented burn area of Australian mega forest fires. Nat. Clim. Chang. 2020, 10, 171–172. [Google Scholar] [CrossRef]
  4. Clarke, H.; Nolan, R.H.; De Dios, V.R.; Bradstock, R.; Griebel, A.; Khanal, S.; Boer, M.M. Forest fire threatens global carbon sinks and population centres under rising atmospheric water demand. Nat. Commun. 2022, 13, 7161. [Google Scholar] [CrossRef]
  5. Bal, Y.; Wang, B.; Wu, Y.-d.; Liu, X.-d. A review of global forest fires in 2021. Fire Sci. Technol. 2022, 41, 705–709. [Google Scholar]
  6. Sadi, M.; Zhang, Y.; Xie, W.-F.; Hossain, F.A. Forest fire detection and localization using thermal and visual cameras. In Proceedings of the 2021 International Conference on Unmanned Aircraft Systems (ICUAS), Athens, Greece, 15–18 June 2021; pp. 744–749. [Google Scholar]
  7. Chai, J.; Zeng, H.; Li, A.; Ngai, E.W. Deep learning in computer vision: A critical review of emerging techniques and application scenarios. Mach. Learn. Appl. 2021, 6, 100134. [Google Scholar] [CrossRef]
  8. Lertsinsrubtavee, A.; Sarambage Jayarathna, K.G.; Mekbungwan, P.; Kanabkaew, T.; Raksakietisak, S. SEA-HAZEMON: Active Haze Monitoring and Forest Fire Detection Platform. In Proceedings of the 17th Asian Internet Engineering Conference, Hiroshima, Japan, 19–21 December 2022; pp. 88–95. [Google Scholar]
  9. He, L.; Gong, X.; Zhang, S.; Wang, L.; Li, F. Efficient attention based deep fusion CNN for smoke detection in fog environment. Neurocomputing 2021, 434, 224–238. [Google Scholar] [CrossRef]
  10. McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons, Inc.: New York, NY, USA, 1976. [Google Scholar]
  11. Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233. [Google Scholar] [CrossRef]
  12. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  13. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar]
  14. Li, P.; Tian, J.; Tang, Y.; Wang, G.; Wu, C. Deep retinex network for single image dehazing. IEEE Trans. Image Process. 2020, 30, 1100–1115. [Google Scholar] [CrossRef]
  15. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color balance and fusion for underwater image enhancement. IEEE Trans. Image Process. 2017, 27, 379–393. [Google Scholar] [CrossRef]
  16. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  17. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [PubMed]
  18. Chen, D.; He, M.; Fan, Q.; Liao, J.; Zhang, L.; Hou, D.; Yuan, L.; Hua, G. Gated context aggregation network for image dehazing and deraining. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1375–1383. [Google Scholar]
  19. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3194–3203. [Google Scholar]
  20. Dong, H.; Pan, J.; Xiang, L.; Hu, Z.; Zhang, X.; Wang, F.; Yang, M.-H. Multi-scale boosted dehazing network with dense feature fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2157–2167. [Google Scholar]
  21. Yang, L.; Gong, X.; Wu, Z.; Han, Y.; He, L.; Li, F. Dark-Channel Mixed Attention Based Neural Networks for Smoke Detection in Fog Environment. In Proceedings of the Adjunct Proceedings of the 2021 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2021 ACM International Symposium on Wearable Computers, New York, NY, USA, 21–26 September 2021; pp. 691–696. [Google Scholar]
  22. Balki Tas, M.; Tas, Y.; Balki, O.; Aydin, Z.; Tasdemir, K. Camera-based wildfire smoke detection for foggy environments. J. Electron. Imaging 2022, 31, 053033. [Google Scholar]
  23. Huang, J.; He, Z.; Guan, Y.; Zhang, H. Real-time forest fire detection by ensemble lightweight YOLOX-L and defogging method. Sensors 2023, 23, 1894. [Google Scholar] [CrossRef] [PubMed]
  24. Sun, Y.; Su, L.; Luo, Y.; Meng, H.; Zhang, Z.; Zhang, W.; Yuan, S. IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6029–6043. [Google Scholar] [CrossRef]
  25. Zhu, Z.; Luo, Y.; Qi, G.; Meng, J.; Li, Y.; Mazur, N. Remote sensing image defogging networks based on dual self-attention boost residual octave convolution. Remote Sens. 2021, 13, 3104. [Google Scholar] [CrossRef]
  26. He, S.; Chen, Z.; Wang, F.; Wang, M. Integrated image defogging network based on improved atmospheric scattering model and attention feature fusion. Earth Sci. Inform. 2021, 14, 2037–2048. [Google Scholar] [CrossRef]
  27. Chen, W.-T.; Chen, I.-H.; Yeh, C.-Y.; Yang, H.-H.; Ding, J.-J.; Kuo, S.-Y. Sjdl-vehicle: Semi-supervised joint defogging learning for foggy vehicle re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; pp. 347–355. [Google Scholar]
  28. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany 8–14 September 2018; pp. 801–818. [Google Scholar]
  29. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  30. Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P.; Blasch, E. The FLAME Dataset: Aerial Imagery Pile Burn Detection Using Drones (UAVs); IEEE DataPort: New York, NY, USA, 2020. [Google Scholar]
  31. Chino, D.Y.; Avalhais, L.P.; Rodrigues, J.F.; Traina, A.J. Bowfire: Detection of fire in still images by integrating pixel color and texture analysis. In Proceedings of the 2015 28th SIBGRAPI Conference on Graphics, Patterns and Images, Salvador, Bahia, Brazil, 26–29 August 2015; pp. 95–102. [Google Scholar]
  32. Mou, L.; Hua, Y.; Jin, P.; Zhu, X.X. Era: A data set and deep learning benchmark for event recognition in aerial videos [software and data sets]. IEEE Geosci. Remote Sens. Mag. 2020, 8, 125–133. [Google Scholar] [CrossRef]
  33. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
  34. Wang, Y.; Jeong, W.; Moon, Y.S. Single Image Dehazing Based on Depth Map Estimation via Generative Adversarial Networks. J. Korean Soc. Internet Inf. 2018, 19, 43–54. [Google Scholar]
  35. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation policies from data. arXiv 2018, arXiv:1805.09501. [Google Scholar]
  36. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  37. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  38. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  39. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  40. Guan, Z.; Miao, X.; Mu, Y.; Sun, Q.; Ye, Q.; Gao, D. Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model. Remote Sens. 2022, 14, 3159. [Google Scholar] [CrossRef]
  41. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  42. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  43. Wang, Z.; Peng, T.; Lu, Z. Comparative Research on Forest Fire Image Segmentation Algorithms Based on Fully Convolutional Neural Networks. Forests 2022, 13, 1133. [Google Scholar] [CrossRef]
  44. Chen, G.; Li, C.; Wei, W.; Jing, W.; Woźniak, M.; Blažauskas, T.; Damaševičius, R. Fully convolutional neural network with augmented atrous spatial pyramid pool and fully connected fusion path for high resolution remote sensing image segmentation. Appl. Sci. 2019, 9, 1816. [Google Scholar] [CrossRef]
  45. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  46. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  47. Lentile, L.B.; Holden, Z.A.; Smith, A.M.; Falkowski, M.J.; Hudak, A.T.; Morgan, P.; Lewis, S.A.; Gessler, P.E.; Benson, N.C. Remote sensing techniques to assess active fire characteristics and post-fire effects. Int. J. Wildland Fire 2006, 15, 319–345. [Google Scholar] [CrossRef]
  48. Seydi, S.T.; Saeidi, V.; Kalantar, B.; Ueda, N.; Halin, A.A. Fire-Net: A deep learning framework for active forest fire detection. J. Sens. 2022, 2022, 8044390. [Google Scholar] [CrossRef]
Figure 1. Images of possible forest fire. (a,b) Small target fire in the snow; (c) smoky larger fire; (d) dense area fire; (e,f) other objects similar to fire.
Figure 1. Images of possible forest fire. (a,b) Small target fire in the snow; (c) smoky larger fire; (d) dense area fire; (e,f) other objects similar to fire.
Forests 14 01859 g001
Figure 2. Fire samples of FFLAD: (a) clear images of wildfires; (b,c) depth map with three-channel and single-channel; (d) foggy images of wildfires; and (e) ground truth.
Figure 2. Fire samples of FFLAD: (a) clear images of wildfires; (b,c) depth map with three-channel and single-channel; (d) foggy images of wildfires; and (e) ground truth.
Forests 14 01859 g002
Figure 3. Model structure of DeepLabV3+.
Figure 3. Model structure of DeepLabV3+.
Forests 14 01859 g003
Figure 4. The architecture of our joint dehazing learning network for forest fire segmentation. The network consists of two branches: the forest fire division branch and the fog removal branch. Two branches share collective features simultaneously by Collective Feature Sharing Module.
Figure 4. The architecture of our joint dehazing learning network for forest fire segmentation. The network consists of two branches: the forest fire division branch and the fog removal branch. Two branches share collective features simultaneously by Collective Feature Sharing Module.
Forests 14 01859 g004
Figure 5. DARA: fusion attention residual module.
Figure 5. DARA: fusion attention residual module.
Forests 14 01859 g005
Figure 6. The flow chart of the detailed work with the whole model.
Figure 6. The flow chart of the detailed work with the whole model.
Forests 14 01859 g006
Figure 7. The efficacy of the suggested cooperative defogging learning method. Model 1~6, respectively, represent the 6 models from top to bottom in Table 3.
Figure 7. The efficacy of the suggested cooperative defogging learning method. Model 1~6, respectively, represent the 6 models from top to bottom in Table 3.
Forests 14 01859 g007
Figure 8. Visualization on the FFLAD dataset. (a) the fogged image; (b) the defogged affected image; and (c) the clear image.
Figure 8. Visualization on the FFLAD dataset. (a) the fogged image; (b) the defogged affected image; and (c) the clear image.
Forests 14 01859 g008
Figure 9. Results of Defog DeepLabV3+ for forest fire segmentation. All flame areas are highlighted with a green box with a serial number. (a) Raw images; (b) predicted images; and (c) ground truth.
Figure 9. Results of Defog DeepLabV3+ for forest fire segmentation. All flame areas are highlighted with a green box with a serial number. (a) Raw images; (b) predicted images; and (c) ground truth.
Forests 14 01859 g009
Figure 10. Flame discrimination experiment in other environments involving other similar object types. (a) Raw image with other similar objects; (b) predicted result. Flame recognition areas are highlighted with a green box with a serial number.
Figure 10. Flame discrimination experiment in other environments involving other similar object types. (a) Raw image with other similar objects; (b) predicted result. Flame recognition areas are highlighted with a green box with a serial number.
Forests 14 01859 g010
Figure 11. The average model performance of 5-fold cross-validation. Test1 represents the split test set and Test2 represents the additional test set.
Figure 11. The average model performance of 5-fold cross-validation. Test1 represents the split test set and Test2 represents the additional test set.
Forests 14 01859 g011
Figure 12. The confusion matrix.
Figure 12. The confusion matrix.
Forests 14 01859 g012
Figure 13. The ROC curve. The blue line represents the result of a random guess.
Figure 13. The ROC curve. The blue line represents the result of a random guess.
Forests 14 01859 g013
Figure 14. Loss curves of Defog DeepLabV3+.
Figure 14. Loss curves of Defog DeepLabV3+.
Forests 14 01859 g014
Figure 15. Results on testing images. Flame recognition areas are highlighted with a green box with a serial number. Flame segmentation details are highlighted with blue boxes. (a) Raw images; (b) ground truth. Segmented images by (c) PSPNet-mobilenet, (d) PSPNet-resnet50, (e) UNet-vgg, (f) DeepLabv3+, (g) our Defog DeepLabV3+, and (h) segmentation details. Numbers (1~7) represent the order of fire points, and from left to right are the predicted results generated by the corresponding ground truth, PSPNet-mobilenet, PSPNet-resnet50, UNet-vgg, DeepLabv3+, and our Defog DeepLabV3+ label.
Figure 15. Results on testing images. Flame recognition areas are highlighted with a green box with a serial number. Flame segmentation details are highlighted with blue boxes. (a) Raw images; (b) ground truth. Segmented images by (c) PSPNet-mobilenet, (d) PSPNet-resnet50, (e) UNet-vgg, (f) DeepLabv3+, (g) our Defog DeepLabV3+, and (h) segmentation details. Numbers (1~7) represent the order of fire points, and from left to right are the predicted results generated by the corresponding ground truth, PSPNet-mobilenet, PSPNet-resnet50, UNet-vgg, DeepLabv3+, and our Defog DeepLabV3+ label.
Forests 14 01859 g015
Figure 16. Comparison of different deep learning segmentation methods on FFLAD. (a) The split test set and (b) the additional test set.
Figure 16. Comparison of different deep learning segmentation methods on FFLAD. (a) The split test set and (b) the additional test set.
Forests 14 01859 g016
Figure 17. Ablation comparison for the DARA module. All models are based on Defog DeepLabV3+. Model 1 and 2 represent xception with DARA or not. Models 3 and 4 represent mobilenetv2 with DARA or not.
Figure 17. Ablation comparison for the DARA module. All models are based on Defog DeepLabV3+. Model 1 and 2 represent xception with DARA or not. Models 3 and 4 represent mobilenetv2 with DARA or not.
Forests 14 01859 g017
Figure 18. Heat map analysis of image features. Flame recognition areas are highlighted with a green box. (a) The original image; (b) the heat map of the fusion of the intermediate feature map and the original image.
Figure 18. Heat map analysis of image features. Flame recognition areas are highlighted with a green box. (a) The original image; (b) the heat map of the fusion of the intermediate feature map and the original image.
Forests 14 01859 g018
Figure 19. Model running efficiency comparison. Alphabet A represents Backbone xception and alphabet B represents Backbone mobilenetv2.
Figure 19. Model running efficiency comparison. Alphabet A represents Backbone xception and alphabet B represents Backbone mobilenetv2.
Forests 14 01859 g019
Figure 20. Sensitivity analysis on the model input parameters. (a) The impact of Batch size on the model performance; (b) The impact of initial Learning Rate on the model performance; (c) The impact of Number of iterations on the model performance.
Figure 20. Sensitivity analysis on the model input parameters. (a) The impact of Batch size on the model performance; (b) The impact of initial Learning Rate on the model performance; (c) The impact of Number of iterations on the model performance.
Forests 14 01859 g020
Table 1. The detailed information of the dataset FFLAD.
Table 1. The detailed information of the dataset FFLAD.
DatasetNumber of ImagesProportion
Training set356480%
Validation44510%
Testing set44510%
Additional test set547-
Total5001100%
Table 2. Training parameters of our model.
Table 2. Training parameters of our model.
Training ParametersDetails
Batch size16
Initial learning rate0.01
DecoderASPP
Number of iterations100
Optimizer strategySGD
Image size512 × 512
Table 3. The efficacy of the suggested cooperative defogging learning method. “BaseLine” indicates that the improved presentation of DeepLabV3+ does not have joint defogging learning. DARA and PEM indicate whether to use the corresponding module. We use the sign ‘√’ to indicate whether to enable them and display the highest value of the mPA in bold.
Table 3. The efficacy of the suggested cooperative defogging learning method. “BaseLine” indicates that the improved presentation of DeepLabV3+ does not have joint defogging learning. DARA and PEM indicate whether to use the corresponding module. We use the sign ‘√’ to indicate whether to enable them and display the highest value of the mPA in bold.
ModuleDCNNDARAPEM mPA (%)
Baselinexception 89.95
BaseLinemobilenetv2 91.27
Defog DeepLabV3+ xception 92.16
Defog DeepLabV3+ xception93.18
Defog DeepLabV3+mobilenetv2 93.33
Defog DeepLabV3+mobilenetv294.26
Table 4. The result of 5-fold cross-validation. Test1 represents the split test set and Test2 represents the additional test set.
Table 4. The result of 5-fold cross-validation. Test1 represents the split test set and Test2 represents the additional test set.
ModelTest SetmPA (%)mIoU (%)mRecall (%)
Model 1Test 194.1089.3293.98
Test 294.0089.2594.03
Model 2Test 194.2289.4894.15
Test 294.1289.4394.16
Model 3Test 194.3589.6794.27
Test 294.2589.6294.29
Model 4Test 194.1889.4294.08
Test 294.0889.3794.10
Model 5Test 194.3089.6194.17
Test 294.2089.5694.19
AverageTest 194.2389.5194.04
Test 294.1389.4594.15
Table 5. Comparison of different deep learning segmentation methods on FFLAD. Test1 represents the split test set and Test2 represents the additional test set.
Table 5. Comparison of different deep learning segmentation methods on FFLAD. Test1 represents the split test set and Test2 represents the additional test set.
MethodologyTest SetmPA (%)mIoU (%)mRecall (%)
PSPNet-mobilenetv2Test 183.1571.6477.93
Test 283.1071.5577.85
PSPNet-resnet50Test 180.7075.4887.68
Test 280.8075.5587.75
UNet-vggTest 195.0088.9592.86
Test 295.2089.1292.96
DeepLabV3+-mobilenetv2Test 191.2784.2990.15
Test 291.4084.4890.33
OursTest 194.2689.5194.04
Test 294.2089.4994.02
Table 6. Ablation experiment for the DARA module. We use the sign ‘√’ to indicate whether to enable the DARA module.
Table 6. Ablation experiment for the DARA module. We use the sign ‘√’ to indicate whether to enable the DARA module.
ModelDCNNDARAmPA (%)mIoU (%)mRecall (%)
Defog DeepLabV3+ xception93.1888.0593.03
Defog DeepLabV3+ xception 92.1086.7292.61
Defog DeepLabV3+mobilenetv294.2689.5194.04
Defog DeepLabV3+mobilenetv2 93.3088.0093.09
Table 7. Model running efficiency comparison.
Table 7. Model running efficiency comparison.
ModuleDCNNTotal Number of ParametersInference Time (ms)
DeepLabV3+xception9,827,4663357.82
DeepLabV3+mobilenetv25,849,0703016.42
Defog DeepLabV3+xception12,984,0293829.34
Defog DeepLabV3+mobilenetv29,596,7453339.61
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, T.; Chen, W.; Lin, X.; Mu, Y.; Huang, J.; Gao, D.; Xu, J. Defogging Learning Based on an Improved DeepLabV3+ Model for Accurate Foggy Forest Fire Segmentation. Forests 2023, 14, 1859. https://doi.org/10.3390/f14091859

AMA Style

Liu T, Chen W, Lin X, Mu Y, Huang J, Gao D, Xu J. Defogging Learning Based on an Improved DeepLabV3+ Model for Accurate Foggy Forest Fire Segmentation. Forests. 2023; 14(9):1859. https://doi.org/10.3390/f14091859

Chicago/Turabian Style

Liu, Tao, Wenjing Chen, Xufeng Lin, Yunjie Mu, Jiating Huang, Demin Gao, and Jiang Xu. 2023. "Defogging Learning Based on an Improved DeepLabV3+ Model for Accurate Foggy Forest Fire Segmentation" Forests 14, no. 9: 1859. https://doi.org/10.3390/f14091859

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop