Next Article in Journal
Hydrological Characteristics Change of Hala Lake and Its Response to Climate Change, 1987–2018
Next Article in Special Issue
Nemo: An Open-Source Transformer-Supercharged Benchmark for Fine-Grained Wildfire Smoke Detection
Previous Article in Journal
Automatic Detection of Coseismic Landslides Using a New Transformer Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Landslide Detection Based on ResU-Net with Transformer and CBAM Embedded: Two Examples with Geologically Different Environments

1
National Institute of Natural Hazards, Ministry of Emergency Management, Beijing 100085, China
2
Key Laboratory of Compound and Chained Natural Hazards Dynamics, Ministry of Emergency Management of China, Beijing 100085, China
3
School of Engineering and Technology, China University of Geosciences (Beijing), Beijing 100083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(12), 2885; https://doi.org/10.3390/rs14122885
Submission received: 12 May 2022 / Revised: 13 June 2022 / Accepted: 14 June 2022 / Published: 16 June 2022
(This article belongs to the Special Issue Artificial Intelligence for Natural Hazards (AI4NH))

Abstract

:
An efficient method of landslide detection can provide basic scientific data for emergency command and landslide susceptibility mapping. Compared to a traditional landslide detection approach, convolutional neural networks (CNN) have been proven to have powerful capabilities in reducing the time consumed for selecting the appropriate features for landslides. Currently, the success of transformers in natural language processing (NLP) demonstrates the strength of self-attention in global semantic information acquisition. How to effectively integrate transformers into CNN, alleviate the limitation of the receptive field, and improve the model generation are hot topics in remote sensing image processing based on deep learning (DL). Inspired by the vision transformer (ViT), this paper first attempts to integrate a transformer into ResU-Net for landslide detection tasks with small datasets, aiming to enhance the network ability in modelling the global context of feature maps and drive the model to recognize landslides with a small dataset. Besides, a spatial and channel attention module was introduced into the decoder to effectually suppress the noise in the feature maps from the convolution and transformer. By selecting two landslide datasets with different geological characteristics, the feasibility of the proposed model was validated. Finally, the standard ResU-Net was chosen as the benchmark to evaluate the proposed model rationality. The results indicated that the proposed model obtained the highest mIoU and F1-score in both datasets, demonstrating that the ResU-Net with a transformer embedded can be used as a robust landslide detection method and thus realize the generation of accurate regional landslide inventory and emergency rescue.

Graphical Abstract

1. Introduction

As one of the most critical types of natural hazards, landslides are triggered by various external factors in most cases, including earthquakes, rainfall, variation of water level storms, and river erosion [1]. Once landslides develop into geological hazards, there is a great potential to cause devastating damage to natural structures and infrastructure, leading to human casualties and property damage [2]. A detailed landslide inventory records the location, date, and type of landslide in a given area, which can serve as a basis for further investigation and provide geologists with scientific data used to assess landslide risk [3,4]. Additionally, studies have shown that the legacy effect may lead to a higher chance of landslides occurring on previous landslide paths in the next decade. Therefore, generating landslide inventory is the first step in hazards management and susceptibility evaluation [5,6]. The development of advanced earth observation techniques provides unique opportunities for the comprehensive assessment of disaster losses [7], thus massive amounts of satellite images are gradually replacing field surveys as a low-cost resource for building a landslide database. As a consequence, how to detect landslides rapidly in large-scale remote sensing datasets is of great significance for disaster mitigation [8,9,10].
The main landslide detection methods based on remote sensing include (1) Pixel-based, (2) Object-oriented, (3) Machine Learning, and (4) Deep Learning. Among them, the adjacent pixels are not taken into account in the pixel-based method which mainly extracts landslides by comparing the images’ intensity or the band difference between two-phase images. For example, Nichol and Wong [11] attempted to extract landslide data using single-band difference, band ratio, and post-classification comparison. Zhao et al. [12] selected the normalized difference vegetation index (NDVI), enhanced vegetation index, and soil-adjusted vegetation index as features to classify the landslides. Mondini et al. [13] detected the rainfall-induced landslides by adopting four change detection. However, the pixel-based method has rigorous requirements for image acquisition and processing. Besides, the improved resolution and differences in the distribution of landslides may introduce pepper noise [3]. The object-oriented approach is a knowledge-driven approach (OOA), in compliance with the assumption that pixels with neighbors belong to the same class and treat landslides as collections of pixels [14]. For example, Lu et al. [15] used temporal principal component analysis, spectral angle mapper, Reed-Xiaoli detection, and textural analysis to derive object features as segmentation metrics and performed multilevel segmentation through a multiscale optimization approach. Rau et al. [16] adopted a multilevel segmentation and hierarchical semantic network for landslide detection. Eeckhaut et al. [17] only used derivatives of LiDAR for landslide detection in high vegetation cover areas. Although OOA takes the shape and texture properties of landslides into consideration to refine the result, the parameters, such as segmentation scale and features, are subjectively optimized by the user, resulting in a time-consuming process and low automation [18,19].
Machine learning [ML] has excellent potential in classification and capacity to handle data of high dimensionality and to map classes with very complex characteristics. Commonly used methods include support vector machines (SVM), decision trees (DTs), random forest (RF), and artificial neural networks (ANN) [20]. Among them, [21,22] have demonstrated that RF exhibited stable portability on new data compared to SVM and ANN. Piralilou et al. [23] proposed a landslide detection method connecting the result of OOA with logistic regression, multilayer perceptron, and RF by the Dempster-Shafer theory. Although machine learning is widely applied in the field of remote sensing, high-dimension data may overfit models or introduce noise, causing performance deterioration. Hence, uncertainties analysis in ML-based models is an indispensable part to select important variables, evaluate the sensibility of the input data and the generalization of the new samples [24,25]. For example, Adnan et al. [26] and Rossi et al. [27] combined the landslide susceptibility maps generated by models to address the uncertainty from the single model.
As a hot field in computer vision, deep learning (DL) has shown outstanding performance in image classification and segmentation [28,29,30]. In terms of landslide susceptibility mapping (LSM), Dao et al. [31] probed into the performance of the DL model which significantly improved the ability to predict future landslides compared to three ML models. Similarly, by introducing nine controlling factors related to landslides, Bui et al. [32] investigated the capability of the DL model in LSM compared to SVM, C4.5 and RF, and the results indicated that the DL model had a more stable classification performance. On the other hand, CNN equipped with VGGNet, ResNet, and DenseNet has been widely applied in landslide detection. For example, Cai et al. [33] and Liu et al. [34] incorporated controlling factors into samples to test the feasibility of Dense-Net in landslide extraction, while Nava et al. [35] explored the performance of the network when topographic factors were fused into SAR images. Furthermore, Ghorbanzadeh et al. [36] attempted to fuse landslide detection results generated from CNN trained with different datasets through the Dempster-Shafer model. Regarding ResU-Net, Prakash et al. [37] first explored the potential of U-Net with ResNet34 in landslide mapping, demonstrating the utility of deep learning based on EO data for regional landslide mapping. Furthermore, Qi et al. [38] proved that the U-Net with ResNet50 can improve the performance of the model on rainfall-induced landslide detection. Ghorbanzadeh et al. [39] firstly used the free Sentinel-2 data in landslide identification by evaluating the performance of U-Net and ResU-Net in three different landslide areas, the result indicated that ResU-Net obtained the highest F1-score. To alleviate the lack of generalization of the model caused by various landslide morphologies, Yi et al. [18] introduced an attention mechanism to assign weights for important feature maps in ResU-Net, the F1-score on the proposed model was improved by 7% compared to ResU-Net. Recently, Ghorbanzadeh et al. [40] designed a novel approach for landslide extraction. By integrating an OBIA-based model and ResU-Net, the proposed model addressed the fuzzy landslide boundaries problem by enhancing and refining the results generated by ResU-Net and possessed higher precision, recall, and F1-score. Besides, Ghorbanzadeh et al. [41] created a public landslide dataset for the landslide detection community, denoted as Landslide4Sense, which contains 3799 images fusing optical bands from Sentinel-2 with DEM and slope. Overall, CNN possesses robustness and scalability, and features are automatically extracted through hierarchical structure and convolution, avoiding the involvement of excessive domain knowledge [42]. However, studies have shown that the actual receptive field in CNN is much smaller than the theoretical size, which is not conducive to making full use of global information [43].
The proposal of a transformer provides an efficient parallel network to natural language processing (NLP) [44,45], and its architecture dominated by multi-head self-attention (MSA) abandons recursion and convolution. Different from CNN that expands the receptive field by stacking convolutions, MSA captures global contextual information at the beginning of feature extraction and establishes long-range dependencies on the target [46,47]. It has been confirmed that the inherent inductive biases, such as transition equivalence and locality, are excluded in the transformer. Among them, the former well respects the nature of the imaging process while the latter controls the model complexity by sharing parameters across space [48]. Nevertheless, the transformer still outperforms CNN when pretraining to a transformer with a large dataset [49,50].
The outstanding performance of the vision transformer (ViT) on classification confirmed that a transformer can be directly used in computer vision. For example, Bazi et al. [51] verified that ViT achieved high classification accuracy on four datasets. Reedha et al. [52] introduced ViT into vegetation classification with a small dataset, achieving higher F1-score compared to EfficienNet and ResNet. Recently, research by Park and Kim [53] showed that MSA will flatten the loss value, guiding the model to converge along a smooth trajectory. In addition, MSA is equivalent to a low-pass filter for aggregating feature maps, while convolution is similar to a high-pass filter, making feature maps more diverse, such phenomenon demonstrated that MSA and convolution complement each other when capturing features. As a consequence, numerous studies aiming to combine the CNN and transformer have appeared, for instance, Deng et al. [54] proposed CTNet, which introduced the ViT to explore global semantic features and employed CNN to compensate for the loss of local structural information. Zheng et al. [45] designed a segmentation network called SETR to model the global context of images based on the pure transformer. The proposed model could obtain excellent results only by combining with a simple decoder. Horry et al. [55] introduced a highly scalable and accurate ViT model into a low-parameter CNN for land use classification, aiming to cope with the growing amount of satellite data. Similarly, Xu et al. [56] proposed an end-to-end network, called ET-GSNet, to improve the generalization of the different tasks by knowledge distillation. Besides, transformer and convolution-based SAR image despeckling networks were proposed by Perera et al. [57] that also possessed superior performance compared to the pure CNN.
Given the strength of transformer in global context modeling, and that, as far as we know, there has been no research evaluating its potential in landslide detection. In this study, a segmentation network incorporating a transformer in ResU-Net proposed by Chen et al. [58] was selected to validate its effectiveness in landslide detection with small datasets, and the pre-trained weight on Imagenet21K was introduced to accelerate model convergence. To better fuse the feature maps from transformer and CNN, a spatial and channel attention module was embedded into the decoder. Finally, two landslide datasets with different characteristics were chosen to confirm the potential of the proposed model thoroughly.

2. Study Area

2.1. Iburi Dataset

The Iburi dataset is located in Iburi, Hokkaido, Japan, with a total area of 733.06 km2. The landforms in this area are mainly hills with moderate slopes and an altitude range from 0–624 m, as shown in Figure 1a. According to JMA: Hokkaido’s annual average temperature is 6–10 °C, and the annual precipitation is 800–1200 mm. At 3:08 local time on 6 September 2018, an Mw 6.6 earthquake struck this region, inducing a large number of coseismic landslides, most of which were shallow soil landslides, as shown in Figure 1b [59]. The Iburi dataset was interpreted by our team based on the pre-and post-earthquake Planet 3 m resolution images. To ensure the correct interpretation, we combined the high-resolution drone images provided by the Geospatial Information Authority of Japan, and finally obtained 9293 landslides.

2.2. Bijie Dataset

Considering the inadequacy of landslide-triggered landslide samples, a public landslide dataset, denoted as the Bijie dataset, was selected to evaluate the generalization of the proposed model. The Bijie dataset mainly consists of 770 landslide images which are all located in Bijie City, Guizhou Province, China. All landslides were extracted from TripleSat images with 0.8 m resolution based on visual interpretation and the majority of landslides were induced by rainfalls, earthquakes, and human activities [60,61], as shown in Figure 1c. Bijie City is situated in the transitional slope zone from the Tibet Plateau to the eastern hills with an altitude of 450–2869 m. Fragile geological environment, unstable geology, undulating terrain and abundant rainfall (annual average rainfall is 849~1399 mm) make it a high landslide-prone area [62], as shown in Figure 1d. In addition, frequent extreme weather has exacerbated the suddenness of geological disasters in this region. Therefore, there is a strong demand for automated and rapid landslide detection approaches for emergency rescue and risk evaluation in this region.

3. Methodology

3.1. Self-Attention and Multi-Head Attention

In the transformer, assuming that there is a processed sequence different from the CNN that generates the attention map through convolution and pooling [63], the goal of self-attention is to capture the interaction amongst all n entities by encoding each entity in terms of the global contextual information [47,52]. The implementation of the self-attention can be briefly described as a process of mapping queries (Q), keys (K), and values (V) to outputs [46], where queries, keys, and values are generated by projecting the input sequence X = x 1 , x 1 , x n n × d to three learnable weight matrices ( W q n × d q , W k n × d k , W v n × d v ) as shown in Equation (1). The specific attention score is computed by scaled dot-product-attention [46]. As shown in Figure 2a, the attention score is acquired through the following steps: (1) Q multiplies the transposed K (2) divided by the d k to reduce the weighting range (3) passes through the softmax function to obtain the final score. The detailed calculation equation of the attention score can be expressed as follows:
Q = X × W q , K = X × W k , V = X × W v
Attention ( Q , K , V ) = softmax ( Q · K T d k ) · V
Compared to CNN, self-attention increases the receptive field without increasing the computational cost associated with kernel sizes. In addition, self-attention is invariant to permutations and changes in the number of input points, hence it can easily operate on irregular inputs instead of the standard convolution that requires a grid structure [52,64]. In transformers, self-attention is extended to multi-head self-attention by computing h self-attention operations in parallel, in which h is called the head. Then, they are concatenated and projected to generate the final attention weights, as shown in Figure 2b. The purpose of using multi-head attention is to obtain semantic information in different subspaces so that the network can focus on multiple types of features, which is similar to the multi-channel in CNN. The detailed equation of MSA can be expressed as follows:
MSA ( Q , K , V ) = Concat ( head 1 , head h ) W v
head i = Attention ( Q W i q , K W i k , V W i v )

3.2. Vision Transformer

The vision transformer aims to introduce 3D pictures into the transformer structure [49]. It avoids the limitations in CNN that global information is obtained by expanding the receptive field layer-by-layer and helps the network model the global context. The complete ViT structure consists of an embedding layer, an encoder, and a head [51], as shown in Figure 3a.
The role of the embedding layer is to complete the conversion of 3D images to sequences to fit the input structure of the transformer. The detailed steps are listed as follows: Given a 3D image, it is first reshaped into a set of non-overlapping and flat 2D patches x n n × P 2 C , where H, W, C represent the height, width, and channel of the images, P represents the resolution of the patches, and n is the number of embedded patches. Subsequently, the patches are projected to the D dimension by a learnable embedding matrix E P 2 C × d [56]. Up to this point, an image is transformed into a sequence consisting of n patches which can also be called tokens. In the classification task (original article), an equally shaped and learnable patch is embedded into the sequence as the final classification representative of the model [54], while in the semantic segmentation task, this is not needed. Although global modeling facilitates the extraction of long-range semantic information, the lack of convolution makes it difficult for the model to capture positional features. Thus, features capable of representing relative or absolute positional information need to be embedded in the sequence [46,49], in ViT, a learnable 1D parameter with the same size as the sequence is summed to form the positional information. The above steps can be summarized as the following equation:
X = [ x 1 E ; x 2 E ; ; x n E ] + E p o s
The transformer is the encoder of ViT, it consists of L layers with the same structure. In each layer, MSA and MLP (multilayer perceptron) are the main components aiming to aggregate important tokens in the sequence [53], as shown in Figure 2b,c. The specific composition is shown in Figure 3b. First, the sequence normalized through the layer normalization is inputted into the MSA to calculate the importance weights among all tokens, and the original information is preserved by using the residual structure. Among them, MSA extracts feature information from different subspaces and then concatenates the result from h self-attention head, which enables the model to selectively assign more focus to certain positions of the image, and also helps the model to learn features faster and reduce the training cost [49,52]. Then, the obtained results are sequentially passed through an LN and an MLP, and the same residual structure is introduced to generate the final semantic information extracted by this layer.

3.3. Convolution Block Attention Module

Since the combination of ResNet-50 and a transformer was selected as the backbone and the pre-trained weights need to be loaded to accelerate the convergence, changing the encoder structure may be inappropriate and inconsistent with the aim of this work. In [65], the attention module Squeeze-and-Excitation (SE) was embedded into the segmentation head to selectively enhance the weight of important features and reduce the weight of less concerning features in channels. Inspired by this, we introduced an attention mechanism in skip connection for the feature maps generated by CNN and the transformer, driving the decoder to focus on more important information in both the global and local feature maps and suppress the unnecessary features. The Convolutional Block Attention Module (CBAM) is a simple and effective attention module that can be seamlessly integrated into any CNN with a negligible number of parameters [66]. As shown in Figure 4, the module derives attention weights along the space and channel, respectively, and then multiplies the weights with the feature maps for adaptive refinement [53].
The implementation of CBAM is depicted as follows: Given a feature map F H × W × C from the encoder, the CBAM derives a one-dimensional channel weight map C 1 × 1 × C and a two-dimensional spatial attention map S H × W × 1 by convolution and pooling. In the channel attention module, the channel weights are described as Equation (6). Specifically, the spatial information is compressed and aggregated into two C × 1 × 1 feature maps by global average pooling and global max pooling, then they are forwarded to a multilayer perceptron (MLP) and summed. Finally, the sigmoid function is applied to generate the final channel attention map.
C = Sigmoid ( MLP ( Maxpool ( F ) ) + MLP ( Avgpool ( F ) ) )
S = Sigmoid ( Conv ( Concat ( Maxpool ( F ) , Avgpool ( F ) ) ) )
In the spatial attention module, the weight values are calculated by Equation (7). In detail, F H × W × C are firstly fed into global max pooling and global average pooling along the channel axis to generate two 1 × H × W feature maps, and then are concatenated to generate an 2 × H × W feature map, in which a convolution operation is applied to reduce the dimensions to 1 × H × W . Finally, the spatial attention map is generated by a sigmoid function. The overall procedure of CBAM is written as:
F P refine = S ( C ( F P ) )

3.4. Attention TransU-Net

ResU-Net [67,68], as a classical network in semantic segmentation, has achieved great success in landslide detection tasks [34,38,39]. It has been demonstrated that the skip connection in U-Net effectively alleviates the detail loss problem due to convolution, and the residual structure prevents the vanishing gradient when mining deeper information [38]. To reduce the limitations of convolution in long-range information extraction, Chen et al. [58] designed a medical image segmentation network called TransUnet, introducing a transformer to model the global context of feature maps generated from ResU-Net, and the proposed model achieved higher accuracy compared to ResU-Net.
In this work, we built a modified TransUnet referring to [58], called CTransUnet (C represents CBAM), that suited the purpose of landslide detection and embedded the attention module proposed in Section 3.2. The proposed model consists of an encoder and decoder, the detailed structure is shown in Figure 5b. In the encoder, ResNet-50 is used to extract feature maps with the different receptive field, and then the introduced transformer aims to model the global context of the feature maps. Unlike the standard ResNet network, the Conv is converted to StdConv, same as in [49], and the batch normalization is replaced with group normalization which is more suitable for the transformer. Besides, the third and fourth stages in ResNet-50 are merged as shown in Figure 5a. The input image is finally converted to three feature maps in ResNet, in which all feature maps provide high-resolution information for the decoder, and the deepest feature map is selected as input for the transformer. Finally, the feature maps generated by transformer X n × d are reshaped into H P × W P × d . In the decoder, the feature maps are resampled to the full resolution by a cascaded upsampler which consists of multiple upsamling blocks. Each block is composed of a 2× upsamling operator, a 3 × 3 convolution operation, and a ReLu activation function successively [58]. In addition, after concatenating the feature maps at each layer, the CBAM module is embedded to further calculate weights for all features.

3.5. Evaluation

To evaluate the performance of the proposed model, the standard ResU-Net was selected as the benchmark in the experiments. In addition, we adopted five commonly used metrics to evaluate quantitatively, which include precision, recall, F1-score, IoU, and mIoU. Among them, precision and recall are defined as the following equations:
Precision = T P T P + F P
Recall = T P T P + F N
where TP, FP, and FN represent the true positive, false positive, and false negative respectively, as shown in Table 1:
F1-score is the harmonic mean of precision and recall [34] as written in Equation (11), it combines the capabilities of the two metrics and can better describe the quality of classification. IoU (Intersection over Union) and mIoU (mean IoU) are often used to evaluate the ability of the model in semantic segmentation [21]. The former describes the ratio of the intersection and union set of the positive results and the real results, and is computed by Equation (12), the latter represents the averaged value of summed IoU of each class.
F 1 - score = 2 precision recall precision + recall = 2 T P 2 T P + F P + F N
IoU = Predict Target Predict Target = T P T P + F P + F N

3.6. Dataset Preparation

In both datasets, a portion of the samples which were not involved in the training and validation process were selected as new data to assess the predictive power of both models [24,25]. For the 770 landslide images in the Bijie dataset, 70% were randomly selected for training with 20% for validation and 10% for testing. Considering the data volume and the distribution characteristics of the Iburi earthquake-induced landslides, a total of 296 images were cropped as the samples, of which 80% were randomly selected as training data and 20% as validation. Besides, 62 images were cropped for testing. The detailed parameters are shown in Table 2. To more perfectly match the pre-training weights, all samples were resized to 224 × 224 with RGB channels.
Finally, seven types of data augmentation strategies, such as salty noise, rotation, and flip, were processed for all samples to prevent overfitting and increase generalization referring to the previous studies [50,69], as shown in Figure 6.

3.7. Experimental Settings

For both encoders, the pre-trained weights on the ImageNet dataset [49] were used to accelerate downstream model convergence. We conducted three experiments in total. In the first set of experiments, the benchmark and the proposed model were trained separately using two datasets and then the performance was evaluated on the test datasets. Considering the parameter volume in ViT, we used base-level transformer parameters (i.e., patch size was set to 16, sequence length was 196, dimension of embedding patches was set to 768, dimension of the fully connected layer was set to 3072, head and layer were set to 12). In the second set of experiments, we evaluated the impact of data augmentation on both models. Finally, the influence of CBAM was evaluated to validate its potential in optimizing the decoder to process features maps.
In the training process, the trial-and-error method was adopted to select appropriate parameters. In detail, the SGD was selected as the optimizer with the momentum set to 0.9 and the weight decay set to 0.0001, the batch size was 48, and the drop-out in the transformer was set to 0.5 and 0.2 for the Iburi and Bijie datasets, respectively, the initial learning rate was set to 0.1, followed by a reduction proportion of 3/10 for every 20 epochs. Total epoch was set to 60. Considering the proportion of positive and negative pixels, the cross-entropy loss and dice loss were integrated to update the weight of the network, the former was less computational and helped the network to fit stably, while the latter mitigated the sample imbalance [70]. The loss function is expressed by the following equations:
L total = 1 L ce + 0.5 L dice
L ce = ( y ^ log ( y ) + ( 1 y ^ ) log ( 1 y ) )
L dice = 1 2 y ^ y + ς y ^ + y + ς
In Equation (14), y ^ represents the label (1 is landslide, 0 is no-landslide), y is the probability predicted by the model, while in Equation (15), y denotes the predicted class. ς represents the smooth coefficient, aiming to avoid that divisor is 0.
The codes were implemented on Pytorch 1.10.1 and all experiments were conducted on a Dell graphics workstation with the following specification: Central processing Unit: Intel Xeon Gold [email protected] GHz, a RAM of 512 GB, and two NVIDIA Quadro RTX 6000 (24 GB).

4. Results

4.1. Result Comparison

The proposed model and ResU-Net were trained with the same dataset for com-parison. Besides, both models were initialized with pre-trained weights, and the results in test samples were evaluated based on ground truth. Table 3 and Table 4 clearly illustrated that the proposed model had the best performance in two datasets compared to the ResU-Net in terms of the Recall, F1-score, IoU, and mIoU, demonstrating the segmentation potential of the proposed model on both datasets. Specifically, for the Bijie dataset, all metrics improved in the proposed model and the F1-score improved by 1.2%, reaching 87.73%, and mIoU improved by 1.04%, reaching 87.91%. For the Iburi dataset, the F1-score of the proposed model improved by 1.84% and mIoU improved by 0.99%. Regarding precision, we found that the proposed model only increased by 0.39% in the Bijie dataset while it decreased by 1.15% in the Iburi dataset. It is worthwhile to mention that recall and IoU in both datasets improved considerably compared to precision and mIoU, the former indicated that the proposed model was more effective in suppressing false negatives in semantic segmentation of landslides, while the latter proved that the proposed model improved apparently the recognition rate of positive samples. Overall, the proposed model had superior performance compared to ResU-Net.
To further explicitly evaluate the segmentation potential of the two models, we randomly selected five images to show the mapping results. By visual inspection, it appears that the proposed model outperformed ResU-Net in both datasets. Figure 7 indicated that both models detected the rough boundary of the landslides in the Bijie dataset, while the results of ResU-Net were easily affected by the landslide-like area or the bare soil area, such as Figure 7(aB,aC). Besides, the proposed model successfully discriminated the vegetation area on the landslide body while ResU-Net filtered the vegetation area on the body in Figure 7(aA). For the Iburi dataset, both models had defects on account of the complexity and variety of landslide shape. As shown in Figure 7(bB,bC), the models obtained poor results due to the complicated shapes of the landslides and the varied pixel values on the landslide body. Nevertheless, the proposed model identified more landslide pixels.

4.2. Impact of Data Augmentation and CBAM

In this section, we analyzed the impact of data augmentation and CBAM on the proposed model. Firstly, we repeated the first set of experiments without augmentation. In the experiment, the batch size was set to 12 considering the data volume was reduced by seven times. Figure 8 showed the results before and after applying the augmentation strategy, here we analyzed the variation in IoU and mIoU.
For the Bijie dataset, Figure 8a,b indicated that the mIoU has increased by 0.52% and IoU has increased by 0.83% after the augmentation in the proposed model, while the corresponding values were 0.99% and 1.68% in ResU-Net. In the Iburi dataset, the proposed model improved mIoU by 0.52% and IoU by 1.1%, while ResU-Net has im-proved by 0.62% and 1.15% as shown in Figure 8c,d. Compared with the Bijie dataset, the impact on the Iburi dataset was slightly less significant attributed to the distribution characteristics of landslides in the sample image as shown in Figure 6. Therefore, the flip and rotation may have little effect on the dataset. Overall, the promotion contributed by the augmentation strategy on the proposed model was lower than that on the pure FCN.
In theory, the addition of CBAM can filter the features that are unfavorable to the segmentation results and increase the weight of the interesting features to enhance the segmentation performance. Hence, we compared the results with and without the addition of CBAM. The experiments were conducted based on the first set of experiments with only the CBAM module removed. In Figure 9, the model with CBAM has improved in F1-score, IoU, and mIoU on two datasets to some extent. Specifically for the Bijie dataset, the attention module had less of an increase in results due to the fact that most of the landslides were single with obvious boundaries in the image. By comparison, for the Iburi dataset, the model exhibited a higher enhancement with a 2.27% increase in IoU. As shown in Figure 10(bA,bB), the model with CBAM could identify landslides with complex and fragmented morphology, proving that CBAM indeed assisted the network in better handling landslide information from multiple feature extraction layers. Additionally, the proposed model was more moderate in dealing with the area with apparent color differences on the landslide surface, maximizing the range of the landslide body.

5. Discussion

This study aimed to explore the potential of a transformer in landslide detection. As a newly proposed approach for image feature extraction, the unique attention mechanism in a transformer allows it to model long-range dependencies for global information on the feature map, while the inductive bias of convolutional networks can regulate the network, ensuring that the network can learn strong representations [53]. By combining CNN and a transformer, the performance of the model can be improved. Besides, it has been demonstrated that fine-tuning the downstream model by introducing the pre-trained transformer weight can accelerate the convergence, which compensates for the premise that a transformer requires large datasets to alleviate weak inductive bias [50]. Additionally, we embedded a spatial and channel attention module in the decoder to better fuse the features extracted through convolution and the transformer. Therefore, we designed a semantic segmentation network called CTransUnet for landslide extraction, in which ResU-Net was chosen as the base framework with a transformer and CBAM embedded into the encoder and decoder respectively. Furthermore, we tested the model performance using two datasets with different landslide characteristics.

5.1. Result Analysis

As shown in Section 4.1, both models obtained relatively higher IoU and F1-scores on different datasets, while the proposed model had a certain degree of improvement for all indicators compared to the pure convolution-based network, proving the feasibility and superiority of transformers. Note that the precision of the proposed model dropped slightly in the Ibrui dataset. To further analyze the reason, the statistics of the confusion matrix were compiled. As shown in Table 5, the proposed model classified more pixels into TP, even though it inevitably misclassified more FP. On the other hand, from the curves of recall and precision [71], the increase of precision may lead to the decrease of recall. Finally, we consider that the increased TP with the cost of slightly higher FP is acceptable for emergency rescue [37,72]. For example, we prefer the pro-posed model to detect more landslide pixels without significantly increasing the number of FN pixels (prefer higher recall rather than higher precision) when generating the distribution map of earthquake-triggered landslides. Although the proposed model classifies more bare soil, building, and road pixels as landslides (lower precision), it can be mitigated by overlaying curvature, slope, and DEM.
In Section 4.2, we found that even the incorporation of a transformer has led to the increased depth and complexity of the network, the proposed model also possessed a higher performance with a small dataset (e.g., only 237 images in the Iburi dataset) by introducing the pre-trained weight compared to the FCN which obviously increased the IoU after being trained with a large dataset. This is slightly different from the view that a transformer relies more on datasets to attenuate the effect of weak inductive bias [49,50]. According to a preliminary analysis, it is mainly because a transformer is not directly used for feature extraction but combined with CNN to better extract global and local semantic information of the feature maps, which is consistent with the findings in [53].
On the other hand, it can be seen from Figure 9 that the indicators of both landslide datasets have improved with the CBAM embedded, and it is worth noting that the proposed model had a higher improvement on the Iburi dataset. Such phenomenon can be ascribed to the difference of the two datasets. Unlike the Bijie dataset, the Iburi dataset suffered from noisy, low-resolution, and fragmented landslide images. However, the addition of CBAM helped the model pay better attention to important features as well as reducing the noise interference, allowing the model to handle landslide areas with large variations in pixel value.

5.2. Complexity Comparison of Models

As a landslide detection method aimed at providing auxiliary scientific data for emergency rescue, its complexity and time consumption need to be paid attention. Hence, the trade-offs between the performance and cost must be considered for practical application, considering that the complex algorithms tend to overload the hard-wires. To evaluate the complexity of the CTransUnet, the number of parameters and training time were compared. As can be seen from Table 6, the costs of introducing transformer into ResU-Net to improve performance are the increase in parameter volume and time consumption.
It is worth mentioning that since a transformer is a relatively large part of the proposed model, we reduce the layers of the transformer from standard 12 to 2 to compute its complexity. Table 6 indicated that when the number of layers of the transformer was reduced to two, it has a similar parameters volume as ResU-Net, nevertheless, the proposed model still cost much time to train. The main reason may be that the transformer may have a negative effect on model operations. To further analyze the proposed model performance with two layers, we repeated the first set of experiments on CTransUnet with two layers. As shown in Table 7, when we reduced the layers to two to ensure similar parameter volume with ResU-Net, the performance of the proposed model was still higher than that of the benchmark. Overall, CTransUnet achieved a relative trade-off between the model performance and complexity.

5.3. Deficiencies of Experiments

Although transformers are gradually becoming a popular and powerful approach in computer vision because of their unique attention mechanism and weak inductive bias, in a majority of related works focused on how to transplant transformers to classification and segmentation, fewer researchers explored the mechanisms between transformers and CNN. In addition, there are still some aspects not considered in this study.
  • How to effectively utilize and optimize the feature maps generated from CNN for the transformer, avoiding the loss of global contextual information.
  • How to reduce the reliance of models on pre-trained weights with a small training dataset, considering that the specialized images are different from the public datasets.
  • The efficient network architecture was not explored comprehensively. For instance, the position of the transformer in ResU-Net.
  • In order to match the pre-trained weights, the image size used in the model was fixed. Hence, multi-source remote sensing images were not taken into consideration for extracting more abundant information.
  • The number of samples needs to be quantitatively analyzed to evaluate its impact on the results.
To fully grasp the mechanism of transformers and their strength when fused with the CNN model, the above issues need to be explored in future research thoroughly, con-tributing to more flexible integration of transformers with various powerful CNN-based models. It has been demonstrated that a specific loss function can help ViT to outperform ResNet without data augmentation and pre-trained weights [73].

6. Conclusions

In this study, we first proposed a new DL-based approach denoted as CTransUnet for landslide detection with small training samples. It combines the strength of two sub-networks, namely ResU-Net and a transformer. The former was used to extract the local and sophisticated features while the latter was introduced to model the long-range dependence in the spatial relationships of the feature maps. Besides, a spatial-channel attention module (CBAM) was embedded into skip connection of the proposed model to increase the weight values of feature maps that were more important for segmentation. Finally, the pre-trained weights were introduced to accelerate the convergence. To evaluate the performance quantitatively, standard ResU-Net and five criteria were selected in experiments. The results showed that CTransUnet possessed optimum performance on both datasets with an IoU 65.74% and 78.15%. In addition, we explored the impact of data augmentation and CBAM on the proposed model. It was prominent that CTransUnet possessed satisfying results even without data augmentation and CBAM had the capability in optimizing the model to focus on the important features and process complex landslide morphologies. Overall, this study demonstrated the outstanding potential of transformers and CNN in landslide extraction, which can provide a scientific data source for emergency response and landslide susceptibility mapping. However, the transformer-based landslide detection approaches are still in infancy, the proposed issues in Section 5.3 should be investigate in further study.

Author Contributions

C.X. proposed the research concept, organized landslide interpretation work, and offered basic data. L.L. participated in the landslide interpretation and data preparation work. Z.Y. designed the framework of the research, conducted the experiments, and wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Institute of Natural Hazards, Ministry of Emergency Management of China (ZDJ2021-14), the Lhasa National Geophysical Observation and Research Station (NORSLS20-07), and the National Key Research and Development Program of China (2018YFC1504703).

Acknowledgments

We are grateful to the editors and anonymous reviewers for their constructive comments and suggestions that improved the quality of the manuscript.

Conflicts of Interest

The authors declare that there are no conflict of interest.

References

  1. Wang, H.; Liu, G.; Xu, W.; Wang, G. GIS-based landslide hazard assessment: An overview. Prog. Phys. Geogr. 2005, 29, 548–567. [Google Scholar] [CrossRef]
  2. Pardeshi, S.D.; Autade, S.; Pardeshi, S.S. Landslide hazard assessment: Recent trends and techniques. Springerplus 2013, 2, 1–11. [Google Scholar] [CrossRef] [Green Version]
  3. Guzzetti, F.; Mondini, A.; Cardinali, M.; Fiorucci, F.; Santangelo, M.; Chang, K. Landslide inventory maps: New tools for an old problem. Earth-Sci. Rev. 2012, 112, 42–66. [Google Scholar] [CrossRef] [Green Version]
  4. Wieczorek, G.F. Preparing a detailed landslide-inventory map for hazard evaluation and reduction. Bull. Assoc. Eng. Geol. 1984, 21, 337–342. [Google Scholar] [CrossRef]
  5. Samia, J.; Temme, A.; Bregt, A.; Wallinga, J.; Guzzetti, F.; Ardizzone, F.; Rossi, M. Do landslides follow landslides? Insights in path dependency from a multi-temporal landslide inventory. Landslides 2017, 14, 547–558. [Google Scholar] [CrossRef] [Green Version]
  6. Kamp, U.; Growley, B.J.; Khattak, G.A.; Owen, L.A. GIS-based landslide susceptibility mapping for the 2005 Kashmir earthquake region. Geomorphology 2008, 101, 631–642. [Google Scholar] [CrossRef]
  7. Martha, T.R.; Govindharaj, K.B.; Kumar, V.K. Damage and geological assessment of the 18 September 2011 Mw 6.9 earthquake in Sikkim, India using very high resolution satellite data. Geosci. Front. 2015, 6, 793–805. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, T.; Liu, M.; Zhang, H.; Jiang, X.; Huang, Y.; Jiang, X. Landslide Detection Based on Improved YOLOv5 and Satellite Images. In Proceedings of the 2021 4th IEEE International Conference on Pattern Recognition and Artificial Intelligence (PRAI), Yibin, China, 20–22 August 2021; pp. 367–371. [Google Scholar] [CrossRef]
  9. Nava, L.; Bhuyan, K.; Meena, S.R.; Monserrat, O.; Catani, F. Rapid Mapping of Landslides on SAR Data by Attention U-Net. Remote Sens. 2022, 14, 1449. [Google Scholar] [CrossRef]
  10. Ju, Y.; Xu, Q.; Jin, S.; Li, W.; Dong, X.; Guo, Q. Automatic object detection of loess landslide based on deep learning. Geomat. Inf. Sci. Wuhan Univ. 2020, 45, 1747–1755. [Google Scholar] [CrossRef]
  11. Nichol, J.; Wong, M.S. Satellite remote sensing for detailed landslide inventories using change detection and image fusion. Int. J. Remote Sens. 2005, 26, 1913–1926. [Google Scholar] [CrossRef]
  12. Zhao, W.; Li, A.; Nan, X.; Zhang, Z.; Lei, G. Post-earthquake landslides mapping from Landsat-8 data for the 2015 Nepal earthquake using a pixel-based change detection method. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1758–1768. [Google Scholar] [CrossRef]
  13. Mondini, A.C.; Guzzetti, F.; Reichenbach, P.; Rossi, M.; Gardinali, M.; Ardizzone, F. Semi-automatic recognition and mapping of rainfall induced shallow landslides using optical satellite images. Remote Sens. Environ. 2011, 115, 1743–1757. [Google Scholar] [CrossRef]
  14. Moosavi, V.; Talebi, A.; Shirmohammadi, B. Producing a landslide inventory map using pixel-based and object-oriented approaches optimized by Taguchi method. Geomorphology 2014, 204, 646–656. [Google Scholar] [CrossRef]
  15. Lu, P.; Stumpf, A.; Kerle, N.; Casagli, N. Object-oriented change detection for landslide rapid mapping. IEEE Geosci. Remote. Sens. Lett. 2011, 8, 701–705. [Google Scholar] [CrossRef]
  16. Rau, J.; Jhan, J.; Rau, R. Semiautomatic object-oriented landslide recognition scheme from multisensor optical imagery and DEM. IEEE Trans. Geosci. Remote Sens. 2013, 52, 1336–1349. [Google Scholar] [CrossRef]
  17. Van Den Eeckhaut, M.; Kerle, N.; Poesen, J.; Hervas, Javier. Object-oriented identification of forested landslides with derivatives of single pulse LiDAR data. Geomorphology 2012, 173, 30–42. [Google Scholar] [CrossRef]
  18. Yi, Y.; Zhang, W. A new deep-learning-based approach for earthquake-triggered landslide detection from single-temporal RapidEye satellite imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6166–6176. [Google Scholar] [CrossRef]
  19. Keyport, R.N.; Oommen, T.; Martha, T.R.; Sajinkumar, K.S.; Gierke, J.S. A comparative analysis of pixel-and object-based detection of landslides from very high-resolution images. Int. J. Appl. Earth Obs. Geoinf. 2018, 64, 1–11. [Google Scholar] [CrossRef]
  20. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of machine-learning classification in remote sensing: An applied review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef] [Green Version]
  21. Ghorbanzadeh, O.; Blaschke, T.; Gholamnia, K.; Meena, S.R.; Tiede, D.; Aryal, J. Evaluation of Different Machine Learning Methods and Deep-Learning Convolutional Neural Networks for Landslide Detection. Remote Sens. 2019, 11, 196. [Google Scholar] [CrossRef] [Green Version]
  22. Hu, Q.; Zhou, Y.; Wang, S.; Wang, F.; Wang, H. Improving the Accuracy of Landslide Detection in “Off-site” Area by Machine Learning Model Portability Comparison: A Case Study of Jiuzhaigou Earthquake, China. Remote Sens. 2019, 11, 2530. [Google Scholar] [CrossRef] [Green Version]
  23. Tavakkoli Piralilou, S.; Shahabi, H.; Jarihani, B.; Ghorbanzadeh, O.; Blaschke, T.; Gholamnia, K.; Meena, S.R.; Aryal, J. Landslide Detection Using Multi-Scale Image Segmentation and Different Machine Learning Models in the Higher Himalayas. Remote Sens. 2019, 11, 2575. [Google Scholar] [CrossRef] [Green Version]
  24. Guzzetti, F.; Reichenbach, P.; Ardizzone, F.; Cardinali, M.; Galli, M. Estimating the quality of landslide susceptibility models. Geomorphology 2006, 81, 166–184. [Google Scholar] [CrossRef]
  25. Sterlacchini, S.; Ballabio, C.; Blahut, J.; Masetti, M.; Sorichetta, A. Spatial agreement of predicted patterns in landslide susceptibility maps. Geomorphology 2011, 125, 51–61. [Google Scholar] [CrossRef]
  26. Adnan, M.S.G.; Rahman, M.S.; Ahmed, N.; Ahmed, B.; Rabbi, M.F.; Rahman, R.M. Improving Spatial Agreement in Machine Learning-Based Landslide Susceptibility Mapping. Remote Sens. 2020, 12, 3347. [Google Scholar] [CrossRef]
  27. Rossi, M.; Guzzetti, F.; Reichenbach, P.; Mondini, A.C.; Peruccacci, S. Optimal landslide susceptibility zonation based on multiple forecasts. Geomorphology 2010, 114, 129–142. [Google Scholar] [CrossRef]
  28. Pritt, M.; Chern, G. Satellite image classification with deep learning. In Proceedings of the 2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 10–12 October 2017; pp. 1–7. [Google Scholar] [CrossRef]
  29. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7. [Google Scholar] [CrossRef]
  30. Wang, R.; Lei, T.; Cui, R.; Zhang, B.; Meng, H.; Nandi, A.K. Medical Image Segmentation Using Deep Learning: A Survey. Wiley IET Image Process. 2020, 16, 1243–1267. [Google Scholar] [CrossRef]
  31. Dao, D.V.; Jaafari, A.; Bayat, M.; Mafi-Gholami, D.; Qi, C.; Moatedi, H.; Phong, T.V.; Ly, H.; Le, T.; Trinh, P.T.; et al. A spatially explicit deep learning neural network model for the prediction of landslide susceptibility. Catena 2020, 188, 104451. [Google Scholar] [CrossRef]
  32. Bui, D.T.; Tsangaratos, P.; Nguyen, V.T.; Liem, N.V.; Trinh, P.T. Comparing the prediction performance of a Deep Learning Neural Network model with conventional machine learning models in landslide susceptibility assessment. Catena 2020, 188, 104426. [Google Scholar] [CrossRef]
  33. Cai, H.; Chen, T.; Niu, R.; Plaza, A. Landslide detection using densely connected convolutional networks and environmental conditions. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 5235–5247. [Google Scholar] [CrossRef]
  34. Liu, T.; Chen, T.; Niu, R.; Plaza, A. Landslide Detection Mapping Employing CNN, ResNet, and DenseNet in the Three Gorges Reservoir, China. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 11417–11428. [Google Scholar] [CrossRef]
  35. Nava, L.; Monserrat, O.; Catani, F. Improving landslide detection on SAR data through deep learning. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  36. Ghorbanzadeh, O.; Meena, S.R.; Abadi, H.S.S.; Piralilou, S.T.; Lv, Z.; Blaschke, T. Landslide Mapping Using Two Main Deep-Learning Convolution Neural Network Streams Combined by the Dempster–Shafer Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 452–463. [Google Scholar] [CrossRef]
  37. Prakash, N.; Manconi, A.; Loew, S. Mapping Landslides on EO Data: Performance of Deep Learning Models vs. Traditional Machine Learning Models. Remote Sens. 2020, 12, 346. [Google Scholar] [CrossRef] [Green Version]
  38. Qi, W.; Wei, M.; Yang, W.; Xu, C.; Ma, C. Automatic Mapping of Landslides by the ResU-Net. Remote Sens. 2020, 12, 2487. [Google Scholar] [CrossRef]
  39. Ghorbanzadeh, O.; Crivellari, A.; Ghamisi, P.; Shahabi, H.; Blaschke, T. A comprehensive transferability evaluation of U-Net and ResU-Net for landslide detection from Sentinel-2 data (case study areas from Taiwan, China, and Japan). Sci. Rep. 2021, 11, 14629. [Google Scholar] [CrossRef]
  40. Ghorbanzadeh, O.; Gholamnia, K.; Ghamisi, P. The application of ResU-net and OBIA for landslide detection from multi-temporal sentinel-2 images. Big Earth Data 2022, 1–26. [Google Scholar] [CrossRef]
  41. Ghorbanzadeh, O.; Xu, Y.; Ghamis, P.; Kopp, M.; Kreil, D. Landslide4Sense: Reference Benchmark Data and Deep Learning Models for Landslide Detection. arXiv 2022, arXiv:2206.00515. [Google Scholar]
  42. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamara, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  43. Luo, W.; Li, Y.; Urtasun, R.; Zemel, R. Understanding the effective receptive field in deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar] [CrossRef]
  44. Wang, Q.; Li, B.; Xiao, T.; Zhu, J.; Li, C.; Wong, D.F.; Chao, L.S. Learning deep transformer models for machine translation. arXiv 2019, arXiv:1906.01787. [Google Scholar]
  45. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805v2. [Google Scholar]
  46. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst 2017, 30. [Google Scholar] [CrossRef]
  47. Cui, X.; Chen, X.; Zhou, J.; Lin, D. Transformer in image interpretation. In Proceedings of the 2021 International Conference on Computer Graphics, Artificial Intelligence, and Data Processing (ICCAID), Harbin, China, 18 March 2022; pp. 45–50. [Google Scholar] [CrossRef]
  48. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 6881–6890. [Google Scholar] [CrossRef]
  49. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  50. Steiner, A.; Kolesnikov, A.; Zhai, X.; Wightman, R.; Uszkoreit, J.; Beyer, L. How to train your vit? Data, augmentation, and regularization in vision transformers. arXiv 2021, arXiv:2106.10270. [Google Scholar]
  51. Bazi, Y.; Bashmal, L.; Rahhal, M.M.A.; Dayil, R.A.; Ajlan, N.A. Vision Transformers for Remote Sensing Image Classification. Remote Sens. 2021, 13, 516. [Google Scholar] [CrossRef]
  52. Reedha, R.; Dericquebourg, E.; Canals, R.; Hafiane, A. Transformer Neural Network for Weed and Crop Classification of High Resolution UAV Images. Remote Sens. 2022, 14, 592. [Google Scholar] [CrossRef]
  53. Park, N.; Kim, S. How Do Vision Transformers Work? arXiv 2022, arXiv:2202.06709v4. [Google Scholar]
  54. Deng, P.; Xu, K.; Huang, H. When CNNs meet vision transformer: A joint framework for remote sensing scene classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  55. Horry, M.; Chakraborty, S.; Pradhan, B.; Shukla, N.; Paul, S. 2-speed network ensemble for efficient classification of incremental land-use/land-cover satellite image chips. arXiv 2022, arXiv:2203.08267v1. [Google Scholar]
  56. Xu, k.; Deng, P.; Huang, H. Vision Transformer: An Excellent Teacher for Guiding Small Networks in Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Electron. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  57. Perera, M.V.; Bandara, W.G.C.; Valanarasu, J.M.J.; Patel, V.M. Transformer-based SAR Image Despeckling. arXiv 2022, arXiv:2201.09355v1. [Google Scholar]
  58. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306v1. [Google Scholar]
  59. Yamagishi, H.; Yanazaki, F. Landslides by the 2018 Hokkaido Iburi-Tobu Earthquake on September 6. Landslides 2018, 15, 2521–2524. [Google Scholar] [CrossRef] [Green Version]
  60. Zheng, G.; Xu, Q.; Ju, Y.; Li, W.; Zhou, X.; Peng, S. The Pusacun rockavalanche on August 28, 2017 in Zhangjiawan Nayongxian, Guizhou: Characteristics and failure mechanism. Eng. Geol. 2018, 26, 223–240. [Google Scholar] [CrossRef]
  61. Wu, C.; Qiao, J.; Wang, C.; Kong, J.; Chen, Z. Analysis on “12· 3” super large-scaled landslide in Zongling, Nayong, Guizhou. Res. Soil Water Conserv. 2006, 13, 100–102. [Google Scholar]
  62. Ji, S.; Yu, D.; Shen, C.; Li, W.; Xu, Q. Landslide detection from an open satellite imagery and digital elevation model dataset using attention boosted convolutional neural networks. Landslides 2020, 17, 1337–1352. [Google Scholar] [CrossRef]
  63. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  64. Kan, S.; Naseer, M.; Hatat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in vision: A survey. ACM Comput. Surv. 2021. [Google Scholar] [CrossRef]
  65. Zhang, C.; Jiang, W.; Zhang, Y.; Wang, W.; Zhao, Q.; Wang, C. Transformer and CNN Hybrid Deep Neural Network for Semantic Segmentation of Very-high-resolution Remote Sensing Imagery. IEEE Trans. Geosci Remote Sens. 2022, 60, 1–20. [Google Scholar] [CrossRef]
  66. Woo, S.; Park, J.; Lee, J.; Kweon, I. CBAM: Convolutional block attention module. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef] [Green Version]
  67. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–27 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  68. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  69. Soares, L.P.; Dias, H.C.; Grohmann, C.H. Landslide segmentation with U-Net: Evaluating different sampling methods and patch sizes. arXiv 2020, arXiv:2007.06672v1. [Google Scholar]
  70. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DIMPBT), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef] [Green Version]
  71. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23th International Conference on Machine Learning (ICML), Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar] [CrossRef] [Green Version]
  72. Brunetti, M.T.; Melillo, M.; Peruccacci, S.; Ciabatta, L.; Brocca, L. How far are we from the use of satellite rainfall products in landslide forecasting? Remote Sens. Environ. 2018, 210, 65–75. [Google Scholar] [CrossRef]
  73. Chen, X.; Hsieh, C.; Gong, B. When vision transformers outperform ResNets without pre-training or strong data augmentations. arXiv 2022, arXiv:2106.01548v3. [Google Scholar]
Figure 1. The geographical distribution of landslide datasets. (a) is the landslide distribution and topography in the Iburi district. (b) is the detailed landslide shapes in the Iburi dataset, the majority of landslides are continuous small- and medium-sized shallow detrital slides. (c) is the detailed landslide shapes in the Bijie dataset. An image contains a landslide. The size of both datasets is 224 × 224 pixels, and the resolutions are 3 m and 0.8 m respectively. (d) is the landslide distribution and topography in Bijie city.
Figure 1. The geographical distribution of landslide datasets. (a) is the landslide distribution and topography in the Iburi district. (b) is the detailed landslide shapes in the Iburi dataset, the majority of landslides are continuous small- and medium-sized shallow detrital slides. (c) is the detailed landslide shapes in the Bijie dataset. An image contains a landslide. The size of both datasets is 224 × 224 pixels, and the resolutions are 3 m and 0.8 m respectively. (d) is the landslide distribution and topography in Bijie city.
Remotesensing 14 02885 g001
Figure 2. The structure of Self Attention (a), Multi-Head Self Attention (b), Multi-Layer Perceptron (c). (a) represents the standard calculation of the self-attention score, (b) is the calculation of self-attention score in transformer. (c) is a part of transformer.
Figure 2. The structure of Self Attention (a), Multi-Head Self Attention (b), Multi-Layer Perceptron (c). (a) represents the standard calculation of the self-attention score, (b) is the calculation of self-attention score in transformer. (c) is a part of transformer.
Remotesensing 14 02885 g002
Figure 3. The structure of vision transformer (a), transformer (b). In (a), conv2d aims to reshape the images into nine parts. Position Embedding gives patch the location information. In (b), *L represents the layer number, default is 12.
Figure 3. The structure of vision transformer (a), transformer (b). In (a), conv2d aims to reshape the images into nine parts. Position Embedding gives patch the location information. In (b), *L represents the layer number, default is 12.
Remotesensing 14 02885 g003
Figure 4. The structure of the Convolutional Block Attention Module. The upper represents the channel attention and the lower represents the spatial attention, the feature maps first pass through the upper and then the lower.
Figure 4. The structure of the Convolutional Block Attention Module. The upper represents the channel attention and the lower represents the spatial attention, the feature maps first pass through the upper and then the lower.
Remotesensing 14 02885 g004
Figure 5. The structure of modified ResNet50 (a), CTansUnet (b). *3,4,9 represents the cycles of blocks in each stage of ResNet-50.
Figure 5. The structure of modified ResNet50 (a), CTansUnet (b). *3,4,9 represents the cycles of blocks in each stage of ResNet-50.
Remotesensing 14 02885 g005
Figure 6. Data augmentation. *-1 represents the Iburi dataset, *-2 represents the Bijie dataset. All converts are implemented through OpenCV. * represents A–H.
Figure 6. Data augmentation. *-1 represents the Iburi dataset, *-2 represents the Bijie dataset. All converts are implemented through OpenCV. * represents A–H.
Remotesensing 14 02885 g006
Figure 7. The result comparison of two models in the Bijie dataset (a) and the Iburi dataset (b). Black color represents the background, white color represents the landslide area. Red polygons show the areas that were not fully identified by the two models. A–D are images randomly selected from the testing samples for visual comparison.
Figure 7. The result comparison of two models in the Bijie dataset (a) and the Iburi dataset (b). Black color represents the background, white color represents the landslide area. Red polygons show the areas that were not fully identified by the two models. A–D are images randomly selected from the testing samples for visual comparison.
Remotesensing 14 02885 g007
Figure 8. The performance of two models on the Bijie and Iburi datasets before and after data augmentation. (a,c) represent the proposed model, (b,d) represent ResU-Net.
Figure 8. The performance of two models on the Bijie and Iburi datasets before and after data augmentation. (a,c) represent the proposed model, (b,d) represent ResU-Net.
Remotesensing 14 02885 g008
Figure 9. The performance and variation of the proposed model before and after embedding CBAM. (a) represents the Bijie dataset, (b) represents the Iburi dataset. The dashed line represents the increment after embedding CBAM.
Figure 9. The performance and variation of the proposed model before and after embedding CBAM. (a) represents the Bijie dataset, (b) represents the Iburi dataset. The dashed line represents the increment after embedding CBAM.
Remotesensing 14 02885 g009
Figure 10. The resulting comparison of two models with and without CBAM in the Bijie dataset (a) and the Iburi dataset (b). Black color represents the background, white color represents the landslide area. Red polygons show the areas that were not fully identified by the models. A–C are images randomly selected from the testing samples for visual comparison.
Figure 10. The resulting comparison of two models with and without CBAM in the Bijie dataset (a) and the Iburi dataset (b). Black color represents the background, white color represents the landslide area. Red polygons show the areas that were not fully identified by the models. A–C are images randomly selected from the testing samples for visual comparison.
Remotesensing 14 02885 g010
Table 1. Confusion matrix for binary classification, 1 represents landslide pixel, 0 represents background pixel.
Table 1. Confusion matrix for binary classification, 1 represents landslide pixel, 0 represents background pixel.
Prediction FalsePrediction Truth
Ground FalseTNFP
Ground TruthFNTP
Table 2. Information of the datasets.
Table 2. Information of the datasets.
DatasetSizeBandTrain/Val/TestResolution
Bijie224 × 224RGB525/150/750.8 m
Iburi237/59/623 m
Table 3. Performance of models on the Bijie dataset. The bold means the variation between the proposed model and ResU-Net.
Table 3. Performance of models on the Bijie dataset. The bold means the variation between the proposed model and ResU-Net.
ModelPrecisionRecallF1-ScoreIoUmIoU
Proposed87.2488.2387.7378.1587.91
(↑0.39)(↑2.01)(↑1.2)(↑1.88)(↑1.04)
ResU-Net86.8586.2286.5376.2786.87
Table 4. Performance of models on the Iburi dataset. The bold means the variation between the proposed model and ResU-Net.
Table 4. Performance of models on the Iburi dataset. The bold means the variation between the proposed model and ResU-Net.
ModelPrecisionRecallF1-ScoreIoUmIoU
Proposed77.0981.7079.3365.7480.23
(↓1.15)(↑3.99)(↑1.36)(↑1.84)(↑0.99)
ResU-Net78.2477.7177.9763.9079.24
Table 5. Confusion matrix of two models on the Iburi dataset, ResU-Net (up), proposed (down).
Table 5. Confusion matrix of two models on the Iburi dataset, ResU-Net (up), proposed (down).
Prediction FalsePrediction Truth
Ground False2,684,93875,702
Ground Truth78,089272,183
Ground False2,675,61485,026
Ground Truth64,094286,178
Table 6. The comparison of models on time consumption and parameters.
Table 6. The comparison of models on time consumption and parameters.
ModelParameterTraining Time
proposed-Layer12105.4 M1.68 h
proposed-Layer234.6 M1.25 h
ResU-Net32.5 M0.87 h
Table 7. The performance of the proposed model with transformer Layer = 2. The bold means the improvement in the proposed model.
Table 7. The performance of the proposed model with transformer Layer = 2. The bold means the improvement in the proposed model.
ModelF1-ScoreIoUmIoU
BiJieproposed-layer287.5477.8487.75
(↑1.01)(↑1.57)(↑0.88)
ResU-Net86.5376.2786.87
Ibruiproposed-layer278.2264.2379.45
(↑0.25)(↑0.33)(↑0.21)
ResU-Net77.9763.9079.24
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, Z.; Xu, C.; Li, L. Landslide Detection Based on ResU-Net with Transformer and CBAM Embedded: Two Examples with Geologically Different Environments. Remote Sens. 2022, 14, 2885. https://doi.org/10.3390/rs14122885

AMA Style

Yang Z, Xu C, Li L. Landslide Detection Based on ResU-Net with Transformer and CBAM Embedded: Two Examples with Geologically Different Environments. Remote Sensing. 2022; 14(12):2885. https://doi.org/10.3390/rs14122885

Chicago/Turabian Style

Yang, Zhiqiang, Chong Xu, and Lei Li. 2022. "Landslide Detection Based on ResU-Net with Transformer and CBAM Embedded: Two Examples with Geologically Different Environments" Remote Sensing 14, no. 12: 2885. https://doi.org/10.3390/rs14122885

APA Style

Yang, Z., Xu, C., & Li, L. (2022). Landslide Detection Based on ResU-Net with Transformer and CBAM Embedded: Two Examples with Geologically Different Environments. Remote Sensing, 14(12), 2885. https://doi.org/10.3390/rs14122885

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop