Next Article in Journal
Evaluation of Vegetation Indexes and Green-Up Date Extraction Methods on the Tibetan Plateau
Next Article in Special Issue
Modeling and Estimating the Land Surface Temperature (LST) Using Remote Sensing and Machine Learning (Case Study: Yazd, Iran)
Previous Article in Journal
A Compensation Method of Saturated Waveform for Space-Borne Laser Altimeter
Previous Article in Special Issue
Correction: Marchese et al. Mt. Etna Paroxysms of February–April 2021 Monitored and Quantified through a Multi-Platform Satellite Observing System. Remote Sens. 2021, 13, 3074
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model

College of Information Science and Technology, Nanjing Forestry University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(13), 3159; https://doi.org/10.3390/rs14133159
Submission received: 26 May 2022 / Revised: 28 June 2022 / Accepted: 29 June 2022 / Published: 1 July 2022

Abstract

:
In recent years, forest-fire monitoring methods represented by deep learning have been developed rapidly. The use of drone technology and optimization of existing models to improve forest-fire recognition accuracy and segmentation quality are of great significance for understanding the spatial distribution of forest fires and protecting forest resources. Due to the spreading and irregular nature of fire, it is extremely tough to detect fire accurately in a complex environment. Based on the aerial imagery dataset FLAME, this paper focuses on the analysis of methods to two deep-learning problems: (1) the video frames are classified as two classes (fire, no-fire) according to the presence or absence of fire. A novel image classification method based on channel domain attention mechanism was developed, which achieved a classification accuracy of 93.65%. (2) We propose a novel instance segmentation method (MaskSU R-CNN) for incipient forest-fire detection and segmentation based on MS R-CNN model. For the optimized model, the MaskIoU branch is reconstructed by a U-shaped network in order to reduce the segmentation error. Experimental results show that the precision of our MaskSU R-CNN reached 91.85%, recall 88.81%, F1-score 90.30%, and mean intersection over union (mIoU) 82.31%. Compared with many state-of-the-art segmentation models, our method achieves satisfactory results on forest-fire dataset.

Graphical Abstract

1. Introduction

Forest fires have caused substantial economic losses, air pollution, environmental degradation, and other challenges all over the world, wreaking havoc on human life, animals, and plants [1,2,3]. According to incomplete statistics, over 200,000 forest fires occur worldwide, destroying approximately 10 million hectares of forest land [4]. Furthermore, the risk of forest fires in China is increasing due to factors such as the wide distribution of forests, the complex topography of forest areas, and the backward monitoring technology [5]. As a result, rapid detection of forest fires can reduce damage to ecosystems and infrastructure. In particular, incipient forest-fire recognition technology can provide forest firefighters with more accurate data on fire behavior, thereby preventing the spread of fires [6].
Traditional forest-fire recognition methods include setting up watchtowers in forest areas for manual monitoring or using infrared instruments carried by helicopters for forest fire detection. These methods are time-consuming, laborious, and generally inefficient in recognition. Thanks to breakthroughs in the field of artificial neural networks, deep neural networks (DNNs) have recently emerged as the most advanced technology in some computer vision challenges [7]. The current success of these types of architectures in very complicated tasks has broadened the scope of their prospective applications and paved the way for their application to real-world problems [8,9]. Nonetheless, recognizing forest fires using aerial imagery is challenging due to fire’s various shapes, scopes, and spectral overlaps [10,11].
At the moment, computer vision-based fire-recognition algorithms are primarily concerned with two aspects: (1) flame detection using color and motion features; and (2) feature extraction and classification using deep neural networks [12]. For example, [13] explored the trajectory and irregularity of flames for fire detection in differently colored spaces. [14] separated the chrominance component from luminance using YCbCr and specific rules. This approach detects fires with high accuracy but is only effective for targets with large fire points. [15] tried to detect fires using a multi-channel framework composed of pixels, complex features, and a Bayes classifier. Considering the translucency of smoke, [16] proposed a semantic segmentation model based on a 3D parallel Fully Convolutional Network (FCN) [17] for recognizing smoke regions in videos. The method eliminates background interference to a certain extent and realizes real-time detection. [18] designed several neural networks to recognize forest-fire images with different backgrounds at night and day and focused on analyzing the model’s performance under different combinations of network structures. [19] aimed to avoid false alarms from forest fires using recurrent neural networks (RNNs). In contrast to other convolutional neural networks (CNNs), RNN-based models achieve higher fire-detection accuracy, but at the cost of a slower detection speed and higher equipment requirements. At larger scales, satellite technology is routinely utilized to assess forest fires globally, but, usually, at a relatively low resolution, and real-time fire images are limited by satellite trajectory [20,21]. In addition, with the advantages of all-weather, wide field of view, flexibility, and fast deployment, [22] achieved incipient forest-fire monitoring in forest areas using an Unmanned Aerial Vehicle (UAV) equipped with a miniature edge computing platform and visible light lenses.
In specific scenarios, the fire-recognition models mentioned above have obtained good results with respect to accuracy and speed. However, most of the datasets used in these studies are composed of images in the middle and later stages of fires and images taken by ground cameras, characterized by large fires and concentrated fires. Therefore, the recognition effect of incipient forest fires is poor [23,24,25]. To our knowledge, there are very few aerial imagery datasets for forest-fire recognition, and these are urgently needed in the development of fire-behavior analysis tools for aerial fire forecasting. Note that UAV-based images show different characteristics, such as dynamic blur and top-view perspective, which are fundamentally different from those taken by ground-based cameras [26,27].
In this paper, we designed specific solutions to two computer-vision problems using the FLAME dataset [28] consisting of video frames captured by UAVs: (1) The binary classification of images, i.e., fire and no-fire. In view of the fact that forest fires are characterized by small targets and complex shapes, the module with an attention mechanism is incorporated into the structure of the classification network ResNet [29], aiming to focus on the label-related regions instead of the background; and (2) fire detection using instance segmentation approaches to accurately distinguish the fire individual and pixel information. An improved model based on Mask Scoring R-CNN (MS R-CNN) [30] for forest fire segmentation is proposed.
The main contributions of this paper are summarized as follows:
  • We design a novel attention mechanism module, which consists of two independent branches for learning semantic information between different channels to enrich feature representation capability;
  • We utilize a U-shaped network to reconstruct the MaskIoU branch of MS R-CNN with the aim of correcting forest-fire edge pixels and reducing segmentation errors; and
  • Experimental results show that the proposed MaskSU R-CNN outperforms many existing CNN-based models on forest-fire instance segmentation.
The remainder of this paper is organized as follows. Section 2 introduces experimental materials and methods for two problems, namely, fire classification and fire instance segmentation. Section 3 provides the experimental results and analysis. Section 4 presents the discussion. Finally, Section 5 makes a few concluding remarks.

2. Materials and Methods

This section details the data source and the annotation of the training set, and then two approaches are presented to solve the different challenges. The first challenge is fire and no-fire classification using deep-learning (DL) method. The second challenge is fire instance segmentation, which is complementary to the first problem, and we will further identify fire regions and distinguish pixel classes on images classified as fire-containing.

2.1. Dataset

2.1.1. Data Source

The data obtained through wireless sensor networks and infrared technology has been widely used in the detection, monitoring, and evaluation of forest fires. Furthermore, the aerial superiority of drones allows us to better understand the forest topography structure and the location of the fire [31]. As a result, we selected the FLAME dataset as our data source, which can be obtained from the website (https://ieee-dataport.org/open-access/flame-dataset-aerial-imagery-pile-burn-detection-using-drones-uavs, accessed on 16 December 2021). The FLAME dataset was gathered by drones during the burning of deposits in Arizona pine forests, and it includes video frames and heatmaps taken by infrared cameras, such as the WhiteHot and GreenHot palettes. Figure 1 shows some representative images from fire, no-fire, and thermal videos.

2.1.2. Data Collection and Annotation

To meet the requirements of our experiment, four researchers preprocessed the original FLAME dataset, including video-frame extraction and image annotation. The study is separated into two parts: forest-fire classification and case segmentation.
For the image-classification task, we extracted 8000 RGB images from the video, which were divided into two classes (fire and no-fire) and saved as the ‘.jpg’ format with the size of 254 × 254. The dataset consists of three components: the training set (6400, 80%), validation set (800, 10%), and testing set (800, 10%). To observe the different effects of the training set on the model’s performance, we employed four different proportions of training set to train the model. The validation and testing set in these experiments continue to use the same images, and both are at a proportion of 10%. The implementation details are recorded in Table 1.
For the instance segmentation task, it can be defined as a pixel-level binary classification problem, in which each pixel is labeled as fire or no-fire (background). To complete the segmentation of the forest fires, the images labeled as fire from Table 1 are considered as a training set. In addition, to train the instance segmentation model and guarantee the quality of the segmentation, we extracted the ground truth of each image through Labelme software. Figure 2 shows the annotation result of a training sample.

2.2. Fire Image Classification Using DSA-ResNet

The principle of image classification using deep neural networks is different from traditional digital image processing techniques [14,32,33]. Traditional methods mostly use mathematical modeling or shallow networks for processing and then recognition, which often fail to break the recognition rate bottleneck and have the problem of missed and false detections in practical applications [26]. However, training a CNN to realize this image classification task aids in learning elements unrelated to the fire. Among CNNs, ResNet [29], with many residual blocks, allows for smooth gradient-flow and improves classification accuracy. Furthermore, the attention mechanism provides new momentum for the advancement of CNNs to extract more useful information [34,35]. Experiments show that some attention mechanisms based on channel domain or spatial domain, such as SENet [36] and CBAM [37], can significantly improve network recognition ability. Forest-fire recognition is a challenging task due to the interference of smoke and the translucent nature of the flames. Considering the complexity of forest-fire characteristics, we further exploit multi-scale information based on the SE module to enhance the representation of the model. Unlike previous studies, we propose a novel module using attention mechanism for convolution kernels, which can dynamically select and fuse feature maps from different scales of convolution kernels, termed the Dual Semantic Attention (DSA) module. To be more specific, we implement the DSA module via three operators—Separate, Fuse, and Select, as shown in Figure 3.
Separate: An input feature map X R H × W × C is converted by two transformations with a series of operations, such as grouped convolutions, batch normalization (BN), and Rectified Linear Unit (ReLU) activation function, to achieve the output, denoted as F ˜ = X U ˜ R H × W × C and F ^ = X U ^ R H × W × C . Note that for the purpose of learning the weight relations of different branches, we define convolution kernel size of 3 × 3 and 5 × 5 for feature extraction.
Fuse: The purpose of fusion is to learn the channel weights between different feature streams by adaptively adjusting the convolutional kernels (neurons). Firstly, we fuse the output of F ˜ and F ^ using element-wise summation:
U = U ˜ + U ^ ,
then we obtained the global representation s R C by global average pooling:
s c = F g p ( U c ) = 1 H × W i = 1 H j = 1 W U c ( i , j ) ,
where U c denotes the feature map of c-th channel. Furthermore, a fully connected (FC) layer is applied to achieve a compact tensor z R d and reduce computational effort:
z = F f c ( s ) = δ ( B N ( W s ) ) ,
where δ is the ReLU activation function, B N ( ) represents the batch normalization, W R C r × C is a linear mapping, and r is the descending ratio, which is set to 16 in our experiments.
Select: Two independent FC layers are used to embed the attention information, followed by normalization using softmax function:
a t t i = s o f t m a x ( F f c i ( z ) ) ,
where a t t i R C denotes i-th softmax attention, and i = { 1 , 2 } . Finally, the output of DSA module V is obtained via the channel-based attention weights and their corresponding feature maps, denoted as:
V = a t t 1 U ˜ + a t t 2 U ^ .
Note that the above formula is implemented in two branches and one can easily derive the case with more branches by extending Equations (1), (4), and (5).
On the basis of ResNet, the above-mentioned DSA module is integrated into the model, and the network structure of DSA-ResNet with 50 layers is shown in Figure 4. Similar to other CNNs, our DSA-ResNet50 model consists of three primary components: (1) the input feature matrix, (2) the feature extraction layers, and (3) the output layer. During the training phase, the input matrix X is firstly resized to 230 × 230 × 3, which is dependent on the image size and its channels. Then, the data is augmented using random rotation, horizontal flip, and other techniques to improve the generalization ability of the model and to avoid overfitting. The feature extraction layers consist of a large number of convolution blocks (DSA-Residual module), and each block follows identity mapping and a ReLU activation function [38]. The batch normalization helps to accelerate the convergence of the loss function and enables the model to learn different distributions of the data by normalization. The output of the last feature extraction layer is 7 × 7 × 2048, which is later adjusted to 1 × 1 × 2048 by global pooling. Due to the fact that fire classification is a binary classification problem, we use the sigmoid activation function to output its probabilities (fire, no-fire), denoted as:
P ( f i r e ) = σ ( f i r e | φ ( θ ) ) = 1 1 + e φ ( θ ) ,
where φ ( θ ) represents the output of the FC layer, which is obtained using the input matrix X , pixel values for each channel, and all weights across the entire feature extraction layer, and θ is the weight for the last layer. The output is the probability of fire-recognition with a threshold set to 0.5. To train our DSA-ResNet50 model, a loss function is used to improve network accuracy and find the best weight matrix, which is defined as a binary cross-entropy:
L ( y , y ^ ) = 1 N i = 1 N ( y i l o g p ( y ^ i ) + ( 1 y i ) l o g ( 1 p ( y ^ i ) ) ) ,
where N represents the total number of samples used for each epoch, y is the ground truth label for each image labeled as fire ( y = 1 ) or no-fire ( y = 0 ) in the training phase, and p ( y ^ ) represents the predicted result of an image classified as the fire class. In addition, training is carried out by Adam optimizer [39] to update gradient flow of the network, with the L 2 regularization set to 1 × 10 4 .

2.3. Fire Instance Segementation Using MaskSU R-CNN

As an improvement of Mask R-CNN [40], MS R-CNN is the most advanced instance segmentation method at present. Mask R-CNN regards classification confidence as a criterion for segmentation quality. However, experimental evidence demonstrates that there is no significant correlation between predicted mask quality and classification confidence. To solve this problem, MS R-CNN is obtained by adding a MaskIoU branch to the Mask R-CNN, which is employed to learn and predict the segmentation results, i.e., the segmentation confidence. The MaskIoU branch utilizes related feature map and the mask branch’s result as input (14 × 14 × 257) and then calculates the Intersection Over Union (IoU) between the predicted mask and its corresponding ground truth label in the image. Consequently, MS R-CNN is capable of achieving more precise segmentation results than Mask R-CNN.
UNet [41] is a semantic segmentation model based on FCN, which mainly includes three parts: down-sampling, up-sampling, and concatenation operation. Down-sampling consists of convolution-pooling blocks and is used to compress the number of channels. Each block has two convolutional layers, one pooling layer, and ReLU activation function. In addition, up-sampling doubles the size of the feature map and halves the number of channels by transposed convolution. After that, the output is connected with the feature map with the same size obtained by down-sampling. Throughout the entire process, the concatenation of feature map helps to integrate information in both shallow and deep networks.
Inspired by the UNet structure, we reconstructed the MS R-CNN’s MaskIoU branch using a U-shaped network called MaskSU R-CNN in this paper, and the model structure is presented in Figure 5.

2.3.1. Feature Extraction Network

Given the outstanding performance of the attention mechanism on the image classification task, we adopt DSA-ResNet50 as the backbone of our MaskSU R-CNN. In addition, the introduction of the feature pyramid network (FPN) [42] for multi-scale fusion contributes to extracting more effective features.

2.3.2. Region Proposal Network (RPN) and Region of Interest (RoI) Align

The RPN [43] is composed of two convolution blocks (3 × 3, 1 × 1): The 3 × 3 convolution block extracts features from the output of the backbone network using 32 convolution kernels; the 1 × 1 convolution block is used to adjust the parameters of each anchor box and determine whether there is an object in it. During the training phase, the RPN generates nine anchor boxes for each pixel on the feature map, and then a series of initial proposal regions are screened using Non-Maximum Suppression (NMS).
Unlike the RoI Pooling in the work of [43], RoI Align [40] is used here to scale the RPN output to the same size. RoI Align utilizes bilinear interpolation to calculate grid point coordinates, which effectively preserves the edge pixels of the object to obtain predicted masks with high quality.

2.3.3. Multi-branch Prediction for Classes, Bounding Boxes, and Masks

The multi-branch prediction network contains three branches: the R-CNN branch for classification and bounding-box regression, the mask branch for generating predicted masks, and the MaskIoU branch for segmentation evaluation. The R-CNN branch is consistent with most object detection methods, using the softmax function for classification and SmoothL1 loss for bounding-box regression.
In Mask R-CNN model, the segmentation quality is equivalent to the classification confidence, which is unscientific in practical situations. To address this problem, MS R-CNN introduces the MaskIoU branch, which evaluates the quality of segmentation by concatenating the feature map (14 × 14 × 256) with the output of the mask branch (14 × 14 × 1). During the process of convolution, the features in the shallow network will cause a certain loss. To better perform feature fusion and reduce feature loss, a U-shaped network is adopted in this paper to reconstruct the MaskIoU branch. The novel MaskIoU branch (Table 2) consists of eight convolutional layers, including down-sampling for channel compression, up-sampling for feature expansion, and concatenation operation. Finally, the IoU value between the predicted mask and corresponding ground truth is calculated by three FC layers. The feature concatenation in the U-shaped network integrates the information between different feature maps, which is significant for segmenting the edge pixels of fire.

2.3.4. Model Training and Loss Function

As a large image dataset in deep learning, COCO [44] has more than 220 k images with 80 categories. Before training our MaskSU R-CNN model, a COCO-based pretrained model was applied using transfer learning [45] to train deep neural networks to be more stable and efficient. During training of the model, the method used in the work of [40] has been adopted. Thirty-two anchors are randomly selected from each image in the batch, and the loss is generated based on the positional relationship with the ground truth label. If the IoU value is greater than 0.5, the RoI is considered as a positive sample; otherwise, it is a negative sample, and the ratio of positives and negatives is 1:3.
For generating the regression target in MaskIoU branch, the predicted mask is binarized with a threshold of firstly 0.5. Then we regress the MaskIoU between the predicted mask and its corresponding ground truth by l 2 loss. In addition, the mask loss L m a s k and the MaskIoU loss L m a s k i o u are defined on positive samples only.
During the training phase, the curve of loss function visually reflects the convergence of the model. The total loss of MaskSU R-CNN is composed of two components: (1) the loss in RPN; and (2) the loss generated by the multi-branch prediction network, which can be described as:
L t o t a l = L r p n + L m u l b r a n c h ,
where the RPN loss L r p n is composed of the classification loss (softmax loss) and the bounding-box regression loss (SmoothL1 loss), which is used to generate many proposals (the output of RPN), including the identification of whether or not there are real objects in the anchor and the parameter adjustment of the anchor position. L r p n is computed as follows:
L r p n = 1 N c l s 1 i L c l s ( p i , p i * ) + λ 1 1 N r e g 1 i p i * L r e g ( t i , t i * ) ,
where L m u l b r a n c h is generated by different branches, including the classification loss (softmax loss), the bounding-box regression loss (SmoothL1 loss), the mask loss L m a s k and the MaskIoU loss L m a s k i o u . The formula is expressed as follows:
L m u l b r a n c h = L ( p i , p i * , t i , t i * , s i , s i * ) = 1 N c l s 2 i L c l s ( p i , p i * ) + λ 2 1 N r e g 2 i p i * L r e g ( t i , t i * ) + γ 2 1 N m a s k i L m a s k ( s i , s i * ) + 1 2 N m a s k i L m a s k i o u ( s i , s i * )
where the classification term is normalized by the mini-batch size (i.e., N c l s = 256) and regression term is normalized by the number of anchor locations (i.e., N r e g ~2400); λ * and γ * are hyperparameters used to balance the loss of anchors or bounding-boxes regression and the loss of mask generation during the training phase, which are set to 10 and 1 in our implementation. The classification loss L c l s , regression loss L r e g , mask loss L m a s k , and MaskIoU loss L m a s k i o u are listed as followed:
L c l s ( p i , p i * ) = l o g p * p ,
L r e g ( t i , t i * ) = i s m o o t h L 1 ( t i t i * ) w h e r e   s m o o t h L 1 ( x ) = { 0.5 x 2     i f   | x | < 1 | x | 0.5     o t h e r w i s e
L m a s k ( s i , s i * ) = ( s i * l o g ( s i ) ( 1 s i * ) l o g ( 1 s i ) ) ,
L m a s k i o u ( s i , s i * ) = ( s i s i * s i s i * ) 2 ,
where p i represents the probability that the predicted result of anchor i is the ground truth. Since RPN is used to detect the presence of the target (foreground or background) instead of classification, the value of p i * is 1 when anchor i is a positive sample, otherwise it is 0. t i = ( t i x , t i y , t i w , t i h ) represents the regression parameters of anchor i , including the center coordinates of the bounding box ( x , y ) , the width w , and height h . t i * indicates the ground truth corresponding to anchor i . s and s * represent the binary matrix of the predicted mask and ground truth, respectively. , , and log denote the pixel-based product, summation, and logarithm, respectively.
Finally, the segmentation quality of each target can be expressed by mask score S m a s k :
S m a s k = S c l s S m a s k i o u .
where S c l s represent the classification confidence obtained from R-CNN branch, and S m a s k i o u is the output of MaskIoU branch.

3. Results

This section presents the different performances of two deep neural network models on forest-fire image classification and segmentation. All the experiments are based on Python 3.6 and Pytorch using the Windows system. The hardware used is AMD R7-5800H and an NVIDIA RTX 3070 with 16 GB memory.

3.1. Fire Image Classification

3.1.1. Accuracy Assessment

To observe the performance of the model at different training-set proportions, we compared five existing deep-classification networks (VGGNet [46], GoogleNet [47], ResNet [29], and SE-ResNet50 [36]) with our DSA-ResNet50. The performance of these models trained with different proportions (20%, 40%, 60%, and 80%) of training images is shown in Figure 6. We evaluate the classification performance using four metrics: accuracy (Acc), Kappa coefficient (K), Omission Error (OE), and Commission Error (CE). According to Figure 6, it can be noticed that with the increase of training sets, the Acc and K keep growing while the OE and CE decrease. This is because as the training set increases, the model can learn more relevant features. In addition, OE is slightly higher than CE, which indicates that some of the fine fire points are highly similar to the surrounding soil or obscured by vegetation, making it difficult for the model to classify them accurately. It is worth noting that our DSA-ResNet50 is superior to other models under different proportions of training images. The calculation formulas are as follows:
A c c = T P + T N T P + T N + F P + F N ,
K = p o p e 1 p e ,
C E = F P T P + F P ,
O E = F N T P + F N .
where T P and F P represent the number of fire or no-fire images classified as fire label, respectively; F N and T N represent the number of fire or no-fire images classified as no-fire label, respectively; p o is the overall classification accuracy, and p e is the accidental consistency.
Table 3 demonstrates the results of comparison models trained with 80% images, where our DSA-ResNet50 performs the best (Acc = 93.65%, K = 0.864, OE = 20.59%, and CE = 4.23%). In addition, with a slight increase in network parameters (1.8 million), the addition of the DSA module increased Acc and K by 2.37%, 0.025, and decreased OE and CE by 9.28%, 4.12%, respectively, suggesting that the proposed attention mechanism can capture more features and thus improve the classification ability of the model.

3.1.2. Visualization Analysis

To better understand CNN’s decision on image classification, the visualization method Gradient-weighted Class Activation Mapping (Grad-CAM) [48] was used to generate a heatmap for evaluating important regions in each input image. Figure 7 shows the Grad-CAM visualizations of forest-fire images taken from different UAV angles (shown in the first and third columns), based on DSA-ResNet50 model trained with 80% training images. According to each input image and its corresponding visualization, it can be seen that DSA-ResNet50 can easily focus on areas with fire points (marked with red boxes), indicating that the DSA module enhances the network’s ability to recognize fire areas.

3.2. Fire Detection and Segmentation

3.2.1. Evaluation Metrics

The accuracy of segmentation results is often measured by IoU, which represents the overlap rate between the predicted result and its corresponding ground truth label, and the closer the value is to 1, the better the segmentation performance is. To be fair, we adopt the mean value of IoUs on the testing set to measure the model, denoted as:
m I o U = 1 N i = 1 N P i G i P i G i ,
where P i and G i denote the predicted result and corresponding ground truth label for i-th image, respectively. In our experiment, if the IoU is 0.5 or above, the target is considered a positive sample, otherwise negative. In addition, the F1-score is used as another evaluation metric, denoted as:
f = 2 T P 2 T P + F P + F N ,
where T P , F P , and F N are defined in Section 3.1.1. Obviously, the larger the value of f , the better the accuracy of the model.

3.2.2. Performance Analysis and Comparison

The proposed MaskSU R-CNN is an instance segmentation model that implements parallel processing for object detection and segmentation. From the experimental results (Figure 8b), it can be found that our model can correctly identify forest-fire targets and achieve good segmentation results. Compared with the ground truth labels (Figure 8c), they remain almost the same except for some defects in flame details, and the minor differences are mainly caused by the translucency of forest fires and the interference of occlusions. Figure 9 demonstrates the loss curves over 120 epochs for both the training and validation sets. It can be seen that our model shows an overall smooth decreasing trend and gradually converges after about 80 epochs.
To demonstrate the superiority of MaskSU R-CNN on forest-fire segmentation, we compared our method with several DL-based semantic segmentation models, including SegNet [49], UNet [41], PSPNet [50], and DeepLabv3 [51]. Noticeably, the same dataset and configurations were used to train all models to make the predictions comparable. We selected a representative part of the images from the testing set for display, and the predicted results are shown in Figure 10. We can visually see that the segmentation results of our MaskSU R-CNN outperform other comparison models, especially on images that are hard for humans to recognize. In terms of apparent forest-fire targets that are significantly different from the background, most methods produced relatively accurate segmentation results. As for those forest fires with small targets and high concealment, as marked with green boxes in Figure 10a, most models generated some degree of under-segmentation, except for DeepLabv3 and our MaskSU R-CNN. It is worth noting that SegNet showed serious mis-segmentation (marked with blue boxes), which was mainly caused by the model without taking full advantage of contextual semantic relationships. Table 4 lists the results of the quantitative analysis with different comparison models. Our model obtained the highest f and mIoU on the testing set. Moreover, unlike the above segmentation methods, our model also achieves the differentiation of individual forest fires, which makes the fire segmentation more interpretable.
Furthermore, in order to verify the effectiveness of our improved model MaskSU R-CNN, we compared it with the original model Mask R-CNN and MS R-CNN, and some of the segmentation results are shown in Figure 11. It can be intuitively found that our MaskSU R-CNN achieves the best segmentation results, followed by the MS R-CNN. The main reason is that these two models both add a new branch MaskIoU on the basis of Mask R-CNN, and we further improve the quality of the predicted mask after reconstructing the MaskIoU branch using a U-shaped network. Therefore, the segmentation results are the most excellent. In particular, our model has a remarkable advantage in the correction of edge pixels in the fire regions, especially on inconspicuous fire images, such as fire points 1, 2, 6, and 8 in Figure 11e. As for those highly occluded fire targets in Figure 12, the segmentation confidence of our method is more reasonable, which is determined by the segmentation quality, rather than directly using the classification confidence. Meanwhile, the novel branch MaskIoU greatly improves the fine-grained characterization capability of our model.
To demonstrate the importance of the novel MaskIoU branch in our model, we conducted a series of ablation experiments. According to the ablation results in Table 5, it can be found that the MaskIoU branch can significantly improve the segmentation quality of forest fires (mIoU). In particular, after adding the novel MaskIoU branch with the U-shaped network, our mIoU reached 80.77%. Meanwhile, introducing the attention mechanism (DSA module) to the backbone ResNet can further mine the intrinsic information of the features and improve the performance of the model.
In addition to evaluating the segmentation performance of comparative methods, we also calculated the model size and running time. Our novel MaskIoU branch has about 0.63 G FLOPs compared with 0.39 G FLOPs in MS R-CNN. We use one 3070 GPU to test assess the running time (sec./frame). As for DSA-ResNet50, the speed is about 0.235 for Mask R-CNN, and 0.238 for both MS R-CNN and MaskSU R-CNN. Hence, the computation cost of MaskIoU branch is negligible.

4. Discussion

In contrast to other fixed-form objects, forest fires are dynamic objects with variable shapes and hard-to-depict textures [52]. Generally, a forest fire usually begins as a small-scale fire, develops into a medium-scale fire, and then becomes a large-scale fire. Typologically, it starts from ground fire, then spreads to the trunk, and finally to the tree crown [53,54]. Therefore, the detection of incipient fires appear to be particularly important. Unfortunately, there is little research on this aspect, and most of their research data sets are images of medium-scale or big-scale fires. We focused on this phenomenon and adopted the forest-fire data set based on UAV aerial photography, with minor fire points and strong flame concealment. This can have a high degree of simulating incipient fires. This study proposed a novel method by improving the existing instance segmentation model in order to provide more accurate fire-behavior data, which deeply explores the shallow information and deep higher-order semantics in image features and achieves high-precision recognition of incipient forest fires.
In terms of forest-fire recognition, previous methods have advantages in detecting fires with a faster speed and higher accuracy [55]. However, difficulties arise when complications occur, such as when the capture of fires from a drone’s perspective increases the misdetection rate, and inconspicuous fire points with a small target or high concealment are not easily discovered. To address the issues above, we reconstructed the MaskIoU branch of existing MS R-CNN model by adding a U-shaped network. Specifically, the improved branch cascades feature maps of the same size during encoding and decoding phase, allowing for better integration of pixel location features in the shallow network and pixel category features in the deep network, which provides some correction for edge pixels of forest fire targets.
In order to fully illustrate the rationality of the model in this paper, our MaskSU R-CNN is compared with the original models Mask R-CNN and Mask Scoring R-CNN from several perspectives. The convergence comparison in Figure 9 shows that the overall training loss of our method is slightly lower than the other two models with the same training samples. The visualization comparison in Figure 11 reveals that the segmentation mask of our method has the highest matching degree with the actual shape of the forest fire and has obvious advantages in processing forest-fire edge pixels. The quantitative comparison in Table 5 shows that our method achieves SOTA performance in terms of both detection accuracy and segmentation quality. In addition, the fixed structure of our MaskSU R-CNN allows for end-to-end training. Therefore, it is feasible to prune our method and deploy it to mobile devices under the premise of ensuring recognition accuracy.
Future research will focus on the model’s recognition capacity in satellite remote-sensing imageries, and the fusion of satellite multimodal data for forest-fire detection.

5. Conclusions

In this study, we present two solutions regarding the classification and segmentation of forest-fire images, with the main contributions as follows: (1) we design a novel attention mechanism (DSA module) to enhance the representation ability of feature channels and further improve the classification accuracy of incipient forest fires; (2) we merge the DSA module into ResNet as the backbone network of the instance segmentation model to improve the feature extraction capability; and (3) we reconstruct the MaskIoU branch of MS R-CNN using a U-shaped network, aiming to reduce the segmentation error. Experiments show that our MaskSU R-CNN outperforms many state-of-the-art segmentation models with a precision of 91.85%, recall 88.81%, F1-score 90.30%, and mIoU 82.31% in incipient forest-fire detection and segmentation. Our method, with its flexible structure and excellent performance, represents a shift toward the possibility of unmanned fire monitoring in a large area of forest.

Author Contributions

Conceptualization, Z.G. and D.G.; methodology, Z.G.; software, X.M. and Q.S.; validation, D.G. and Q.Y.; formal analysis, Z.G.; investigation, Z.G.; resources, Z.G. and Y.M.; data curation, Z.G.; writing—original draft preparation, Z.G.; writing—review and editing, D.G. and Q.Y.; visualization, X.M. and Y.M.; supervision, D.G.; project administration, Z.G. and D.G.; funding acquisition, D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Future Network Scientific Research Fund Project (grant number FNSRFP-2021-YB-17) and the Postgraduate Research & Practice Innovation Program of Jiangsu Province (grant number SJCX22_0312).

Data Availability Statement

This work uses the publicly available dataset FLAME, see reference [28] for data availability. More details about the data are available under Section 2.1.

Acknowledgments

We are very grateful to all the students assisted with data annotation and the experiments. We also thank the anonymous reviewers for helpful comments and suggestions to this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ryu, J.-H.; Han, K.-S.; Hong, S.; Park, N.-W.; Lee, Y.-W.; Cho, J. Satellite-based evaluation of the post-fire recovery process from the worst forest fire case in South Korea. Remote Sens. 2018, 10, 918. [Google Scholar] [CrossRef] [Green Version]
  2. Yun, T.; Jiang, K.; Li, G.; Eichhorn, M.P.; Fan, J.; Liu, F.; Chen, B.; An, F.; Cao, L. Individual tree crown segmentation from airborne LiDAR data using a novel Gaussian filter and energy function minimization-based approach. Remote Sens. Environ. 2021, 256, 112307. [Google Scholar] [CrossRef]
  3. Lucas-Borja, M.; Hedo, J.; Cerdá, A.; Candel-Pérez, D.; Viñegla, B. Unravelling the importance of forest age stand and forest structure driving microbiological soil properties, enzymatic activities and soil nutrients content in Mediterranean Spanish black pine (Pinus nigra Ar. ssp. salzmannii) Forest. Sci. Total Environ. 2016, 562, 145–154. [Google Scholar] [CrossRef] [PubMed]
  4. Burrell, A.L.; Sun, Q.; Baxter, R.; Kukavskaya, E.A.; Zhila, S.; Shestakova, T.; Rogers, B.M.; Kaduk, J.; Barrett, K. Climate change, fire return intervals and the growing risk of permanent forest loss in boreal Eurasia. Sci. Total Environ. 2022, 831, 154885. [Google Scholar] [CrossRef]
  5. Wu, Z.; He, H.S.; Keane, R.E.; Zhu, Z.; Wang, Y.; Shan, Y. Current and future patterns of forest fire occurrence in China. Int. J. Wildland Fire 2020, 29, 104. [Google Scholar] [CrossRef]
  6. Yang, X.; Chen, R.; Zhang, F.; Zhang, L.; Fan, X.; Ye, Q.; Fu, L. Pixel-level automatic annotation for forest fire image. Eng. Appl. Artif. Intell. 2021, 104, 104353. [Google Scholar] [CrossRef]
  7. Chai, J.; Zeng, H.; Li, A.; Ngai, E.W. Deep learning in computer vision: A critical review of emerging techniques and application scenarios. Mach. Learn. Appl. 2021, 6, 100134. [Google Scholar] [CrossRef]
  8. Fu, L.; Li, Z.; Ye, Q.; Yin, H.; Liu, Q.; Chen, X.; Fan, X.; Yang, W.; Yang, G. Learning robust discriminant subspace based on joint L2, p-and L2, s-Norm distance metrics. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 130–144. [Google Scholar] [CrossRef]
  9. Ye, Q.; Huang, P.; Zhang, Z.; Zheng, Y.; Fu, L.; Yang, W. Multiview learning with robust double-sided twin SVM. IEEE Trans. Cybern. 2021, 1–14. [Google Scholar] [CrossRef]
  10. Zhan, J.; Hu, Y.; Zhou, G.; Wang, Y.; Cai, W.; Li, L. A high-precision forest fire smoke detection approach based on ARGNet. Comput. Electron. Agric. 2022, 196, 106874. [Google Scholar] [CrossRef]
  11. Yu, Z.; Zhang, Y.; Jiang, B.; Yu, X. Fault-tolerant time-varying elliptical formation control of multiple fixed-wing UAVs for cooperative forest fire monitoring. J. Intell. Robot. Syst. 2021, 101, 48. [Google Scholar] [CrossRef]
  12. Peng, Y.; Wang, Y. Real-time forest smoke detection using hand-designed features and deep learning. Comput. Electron. Agric. 2019, 167, 105029. [Google Scholar] [CrossRef]
  13. Yan, Y.; Wu, X.; Du, J.; Zhou, J.; Liu, Y. Video fire detection based on color and flicker frequency feature. J. Front. Comput. Sci. Technol. 2014, 8, 1271–1279. [Google Scholar]
  14. Çelik, T.; Demirel, H. Fire detection in video sequences using a generic color model. Fire Saf. J. 2009, 44, 147–158. [Google Scholar] [CrossRef]
  15. Borges, P.V.K.; Izquierdo, E. A probabilistic approach for vision-based fire detection in videos. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 721–731. [Google Scholar] [CrossRef]
  16. Li, X.; Chen, Z.; Wu, Q.M.J.; Liu, C. 3D Parallel Fully Convolutional Networks for Real-Time Video Wildfire Smoke Detection. IEEE Trans. Circuits Syst. Video Technol. 2018, 30, 89–103. [Google Scholar] [CrossRef]
  17. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  18. Wang, Y.; Zhao, J.-H.; Zhang, D.-Y.; Ye, W. Forest fire image classification based on deep neural network of sparse autoencoder. Comput. Eng. Appl. 2014, 50, 173–177. [Google Scholar]
  19. Yin, M.; Lang, C.; Li, Z.; Feng, S.; Wang, T. Recurrent convolutional network for video-based smoke detection. Multimed. Tools Appl. 2018, 78, 237–256. [Google Scholar] [CrossRef]
  20. Friedlingstein, P.; Jones, M.W.; O’sullivan, M.; Andrew, R.M.; Hauck, J.; Peters, G.P.; Peters, W.; Pongratz, J.; Sitch, S.; le Quéré, C.; et al. Global carbon budget 2019. Earth Syst. Sci. Data 2019, 11, 1783–1838. [Google Scholar] [CrossRef] [Green Version]
  21. Huang, Q.; Razi, A.; Afghah, F.; Fule, P. Wildfire spread modeling with aerial image processing. In Proceedings of the 2020 IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM), Cork, Ireland, 31 August–3 September 2020; pp. 335–340. [Google Scholar]
  22. De Sousa, J.V.R.; Gamboa, P.V. Aerial forest fire detection and monitoring using a small UAV. KnE Eng. 2020, 5, 242–256. [Google Scholar] [CrossRef]
  23. Ciprián-Sánchez, J.F.; Ochoa-Ruiz, G.; Gonzalez-Mendoza, M.; Rossi, L. FIRe-GAN: A novel deep learning-based infrared-visible fusion method for wildfire imagery. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  24. Pan, H.; Badawi, D.; Zhang, X.; Cetin, A.E. Additive neural network for forest fire detection. Signal Image Video Process. 2019, 14, 675–682. [Google Scholar] [CrossRef]
  25. Zhang, J.; Zhu, H.; Wang, P.; Ling, X. ATT squeeze U-Net: A lightweight network for forest fire detection and recognition. IEEE Access 2021, 9, 10858–10870. [Google Scholar] [CrossRef]
  26. Yuan, C.; Liu, Z.; Zhang, Y. UAV-based forest fire detection and tracking using image processing techniques. In Proceedings of the 2015 International Conference on Unmanned Aircraft Systems (ICUAS), Denver, CO, USA, 9–12 June 2015; pp. 639–643. [Google Scholar]
  27. Sudhakar, S.; Vijayakumar, V.; Kumar, C.S.; Priya, V.; Ravi, L.; Subramaniyaswamy, V. Unmanned Aerial Vehicle (UAV) based Forest Fire Detection and monitoring for reducing false alarms in forest-fires. Comput. Commun. 2019, 149, 1–16. [Google Scholar] [CrossRef]
  28. Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P.Z.; Blasch, E. Aerial imagery pile burn detection using deep learning: The FLAME dataset. Comput. Netw. 2021, 193, 108001. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Huang, Z.; Huang, L.; Gong, Y.; Huang, C.; Wang, X. Mask scoring R-Cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6409–6418. [Google Scholar]
  31. Xu, Y.-H.; Li, J.-H.; Zhou, W.; Chen, C. Learning-empowered resource allocation for air slicing in UAV-assisted cellular V2X communications. IEEE Syst. J. 2022, 1–4. [Google Scholar] [CrossRef]
  32. Chelali, F.Z.; Cherabit, N.; Djeradi, A. Face recognition system using skin detection in RGB and YCbCr color space. In Proceedings of the 2015 2nd World Symposium on Web Applications and Networking (WSWAN), Sousse, Tunisia, 21–23 March 2015; pp. 1–7. [Google Scholar]
  33. Umar, M.M.; Silva, L.C.D.; Bakar, M.S.A.; Petra, M.I. State of the Art of Smoke and Fire Detection Using Image Processing. Int. J. Signal Imaging Syst. Eng. 2017, 10, 22–30. [Google Scholar] [CrossRef]
  34. Hackel, T.; Usvyatsov, M.; Galliani, S.; Wegner, J.D.; Schindler, K. Inference, learning and attention mechanisms that exploit and preserve sparsity in CNNs. Int. J. Comput. Vis. 2020, 128, 1047–1059. [Google Scholar] [CrossRef] [Green Version]
  35. Mnih, V.; Heess, N.; Graves, A. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst. 2014, 2, 2204–2212. [Google Scholar]
  36. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  37. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Li, Y.; Yuan, Y. Convergence analysis of two-layer neural networks with relu activation. Adv. Neural. Inf. Process. Syst. 2017, 30, 597–607. [Google Scholar]
  39. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  41. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  42. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  43. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft Coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  45. Huh, M.; Agrawal, P.; Efros, A.A. What makes ImageNet good for transfer learning? arXiv 2016, arXiv:1608.08614. [Google Scholar]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  47. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  48. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-Cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  49. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  50. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  51. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  52. Ba, R.; Chen, C.; Yuan, J.; Song, W.; Lo, S. SmokeNet: Satellite smoke scene detection using convolutional neural network with spatial and channel-wise attention. Remote Sens. 2019, 11, 1702. [Google Scholar] [CrossRef] [Green Version]
  53. Zhang, F.; Dong, Y.; Xu, S.; Yang, X.; Lin, H. An approach for improving firefighting ability of forest road network. Scand. J. For. Res. 2020, 35, 547–561. [Google Scholar] [CrossRef]
  54. Xu, R.; Lin, H.; Lu, K.; Cao, L.; Liu, Y. A forest fire detection system based on ensemble learning. Forests 2021, 12, 217. [Google Scholar] [CrossRef]
  55. Ye, Q.; Li, Z.; Fu, L.; Zhang, Z.; Yang, W.; Yang, G. Nonpeaked Discriminant Analysis for Data Representation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3818–3832. [Google Scholar] [CrossRef]
Figure 1. Frame samples of different videos: (a,b) normal spectrum palette; and (ce) heatmaps, including GreenHot, Fusion, and WhiteHot palettes.
Figure 1. Frame samples of different videos: (a,b) normal spectrum palette; and (ce) heatmaps, including GreenHot, Fusion, and WhiteHot palettes.
Remotesensing 14 03159 g001
Figure 2. Material of annotation result and ground truth for training: (a) raw image; (b) annotation of the forest-fire target; (c) ground truth label; and (d) raw image with the extracted mask.
Figure 2. Material of annotation result and ground truth for training: (a) raw image; (b) annotation of the forest-fire target; (c) ground truth label; and (d) raw image with the extracted mask.
Remotesensing 14 03159 g002
Figure 3. The proposed DSA module.
Figure 3. The proposed DSA module.
Remotesensing 14 03159 g003
Figure 4. Our proposed DSA-ResNet50 model for fire classification.
Figure 4. Our proposed DSA-ResNet50 model for fire classification.
Remotesensing 14 03159 g004
Figure 5. Our proposed MaskSU R-CNN model for fire segmentation.
Figure 5. Our proposed MaskSU R-CNN model for fire segmentation.
Remotesensing 14 03159 g005
Figure 6. Performances of the existing models (VGGNet16, GoogleNet22, ResNet50, and SE-ResNet50), and DSA-ResNet50 trained with four proportions (20%, 40%, 60%, and 80%) of training images. (a) Testing accuracy; (b) kappa coefficient; (c) omission error; and (d) commission error.
Figure 6. Performances of the existing models (VGGNet16, GoogleNet22, ResNet50, and SE-ResNet50), and DSA-ResNet50 trained with four proportions (20%, 40%, 60%, and 80%) of training images. (a) Testing accuracy; (b) kappa coefficient; (c) omission error; and (d) commission error.
Remotesensing 14 03159 g006
Figure 7. Grad-CAM visualization results on DSA-ResNet50. (af) show the different testing images and their corresponding Grad-CAM results, with the important regions highlighted in red.
Figure 7. Grad-CAM visualization results on DSA-ResNet50. (af) show the different testing images and their corresponding Grad-CAM results, with the important regions highlighted in red.
Remotesensing 14 03159 g007
Figure 8. Results of MaskSU R-CNN for forest-fire instance segmentation: (a) raw images; (b) predicted results; and (c) ground truth labels.
Figure 8. Results of MaskSU R-CNN for forest-fire instance segmentation: (a) raw images; (b) predicted results; and (c) ground truth labels.
Remotesensing 14 03159 g008
Figure 9. Loss curves of MaskSU R-CNN.
Figure 9. Loss curves of MaskSU R-CNN.
Remotesensing 14 03159 g009
Figure 10. Results on testing images. (a) Raw images and (b) ground truth labels. Segmented results by (c) SegNet, (d) UNet, (e) PSPNet, (f) DeepLabv3, and (g) our MaskSU R-CNN.
Figure 10. Results on testing images. (a) Raw images and (b) ground truth labels. Segmented results by (c) SegNet, (d) UNet, (e) PSPNet, (f) DeepLabv3, and (g) our MaskSU R-CNN.
Remotesensing 14 03159 g010
Figure 11. Predicted results of three different instance segmentation models: (a) mask R-CNN, (b) MS R-CNN, (c) our MaskSU R-CNN; (d) ground truth labels; and (e) segmentation details. Note that numbers 1 to 9 represent the order of fire points, and from left to right are the predicted results generated by Mask R-CNN, MS R-CNN, MaskSU R-CNN, and the corresponding ground truth label.
Figure 11. Predicted results of three different instance segmentation models: (a) mask R-CNN, (b) MS R-CNN, (c) our MaskSU R-CNN; (d) ground truth labels; and (e) segmentation details. Note that numbers 1 to 9 represent the order of fire points, and from left to right are the predicted results generated by Mask R-CNN, MS R-CNN, MaskSU R-CNN, and the corresponding ground truth label.
Remotesensing 14 03159 g011
Figure 12. Local visualization and segmentation confidence: (a) mask R-CNN, (b) MS R-CNN, and (c) our MaskSU R-CNN.
Figure 12. Local visualization and segmentation confidence: (a) mask R-CNN, (b) MS R-CNN, and (c) our MaskSU R-CNN.
Remotesensing 14 03159 g012
Table 1. Information about the datasets used for training, validation, and testing.
Table 1. Information about the datasets used for training, validation, and testing.
DatasetNumber of ImagesProportion
Training set160020%
320040%
480060%
640080%
Validation80010%
Testing set80010%
Total8000100%
Table 2. The details of the proposed MaskIoU branch with U-shaped network. C represents the number of classes in the dataset.
Table 2. The details of the proposed MaskIoU branch with U-shaped network. C represents the number of classes in the dataset.
OperationKernel/StrideOutput
Block1Conv + ReLU3 × 3 × 25614 × 14 × 256
Conv + ReLU3 × 3 × 25614 × 14 × 256
Conv + ReLU3 × 3 × 25614 × 14 × 256
Block2Maxpooling2 × 27 × 7 × 256
Conv + ReLU + Concat3 × 3 × 2567 × 7 × 512
Block3Up-sampling2 × 214 × 14 × 512
Conv + ReLU + Concat3 × 3 × 25614 × 14 × 512
Conv + ReLU + Concat3 × 3 × 25614 × 14 × 512
Conv + ReLU + Concat3 × 3 × 25614 × 14 × 512
Block4Conv + ReLU3 × 3 × 25614 × 14 × 512
Maxpooling2 × 27 × 7 × 256
Block5FC + ReLU/1024
FC + ReLU/1024
FC + ReLU/C (MaskIoU)
Table 3. Performances of different models on testing set using 60% training images.
Table 3. Performances of different models on testing set using 60% training images.
ModelLayersAcc (%)KOE (%)CE (%)Params (Million)
VGGNet1684.860.74337.5416.84138.53
GoogleNet2288.230.78434.5211.618.97
ResNet5091.280.83929.878.3526.85
SE-ResNet5092.460.85125.625.6228.65
DSA-ResNet (ours)5093.650.86420.594.2328.43
Table 4. Comparison results between different segmentation models.
Table 4. Comparison results between different segmentation models.
MethodMetrics
p (%) r (%) f (%)mIoU (%)
SegNet71.3741.3352.3535.45
UNet86.1885.9686.0777.85
PSPNet83.1281.2582.1769.74
DeepLabv390.9589.6490.2981.12
MaskSU R-CNN (ours)91.8588.8190.3082.31
Table 5. Ablation experiments between different instance segmentation variants.
Table 5. Ablation experiments between different instance segmentation variants.
ModelBackboneMaskIoUMetrics
p (%) r (%) f (%)mIoU (%)
Mask R-CNNResNet/85.6282.6184.0975.97
DSA-ResNet/87.9485.6986.8077.55
MS R-CNNResNetFCN88.9583.1685.9678.61
DSA-ResNetFCN90.1587.9489.0380.42
MaskSU R-CNN (ours)ResNetU-shaped
network
88.6388.8988.7680.77
DSA-ResNetU-shaped
network
91.8588.8190.3082.31
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guan, Z.; Miao, X.; Mu, Y.; Sun, Q.; Ye, Q.; Gao, D. Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model. Remote Sens. 2022, 14, 3159. https://doi.org/10.3390/rs14133159

AMA Style

Guan Z, Miao X, Mu Y, Sun Q, Ye Q, Gao D. Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model. Remote Sensing. 2022; 14(13):3159. https://doi.org/10.3390/rs14133159

Chicago/Turabian Style

Guan, Zhihao, Xinyu Miao, Yunjie Mu, Quan Sun, Qiaolin Ye, and Demin Gao. 2022. "Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model" Remote Sensing 14, no. 13: 3159. https://doi.org/10.3390/rs14133159

APA Style

Guan, Z., Miao, X., Mu, Y., Sun, Q., Ye, Q., & Gao, D. (2022). Forest Fire Segmentation from Aerial Imagery Data Using an Improved Instance Segmentation Model. Remote Sensing, 14(13), 3159. https://doi.org/10.3390/rs14133159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop