Next Article in Journal
Potential of Satellite Spectral Resolution Vegetation Indices for Estimation of Canopy Chlorophyll Content of Field Crops: Mitigating Effects of Leaf Angle Distribution
Next Article in Special Issue
Detection of Multitemporal Changes with Artificial Neural Network-Based Change Detection Algorithm Using Hyperspectral Dataset
Previous Article in Journal
Climate Change and CO2 Fertilization Have Played Important Roles in the Recent Decadal Vegetation Greening Trend on the Chinese Loess Plateau
Previous Article in Special Issue
An Unsupervised Saliency-Guided Deep Convolutional Neural Network for Accurate Burn Mapping from Sentinel-1 SAR Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

STCD-EffV2T Unet: Semi Transfer Learning EfficientNetV2 T-Unet Network for Urban/Land Cover Change Detection Using Sentinel-2 Satellite Images

1
School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran 14174-66191, Iran
2
Deutsches Zentrum für Luft- und Raumfahrt (DLR), Institut für Methodik der Fernerkundung (IMF), 82234 Wessling, Germany
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(5), 1232; https://doi.org/10.3390/rs15051232
Submission received: 17 January 2023 / Revised: 20 February 2023 / Accepted: 21 February 2023 / Published: 23 February 2023
(This article belongs to the Special Issue Advanced Artificial Intelligence for Environmental Remote Sensing)

Abstract

:
Change detection in urban areas can be helpful for urban resource management and smart city planning. The effects of human activities on the environment and ground have gained momentum over the past decades, causing remote sensing data sources analysis (such as satellite images) to become an option for swift change detection in the environment and urban areas. We proposed a semi-transfer learning method of EfficientNetV2 T-Unet (EffV2 T-Unet) that combines the effectiveness of composite scaled EfficientNetV2 T as the first path or encoder for feature extraction and convolutional layers of Unet as the second path or decoder for reconstructing the binary change map. In the encoder path, we use EfficientNetV2 T, which was trained by the ImageNet dataset. In this research, we employ two datasets to evaluate the performance of our proposed method for binary change detection. The first dataset is Sentinel-2 satellite images which were captured in 2017 and 2021 in urban areas of northern Iran. The second one is the Onera Satellite Change Detection dataset (OSCD). The performance of the proposed method is compared with YoloX-Unet families, ResNest-Unet families, and other well-known methods. The results demonstrated our proposed method’s effectiveness compared to other methods. The final change map reached an overall accuracy of 97.66%.

Graphical Abstract

1. Introduction

Remote sensing data are used in the detection of the changes unfolding in an area over a period of time [1]. This procedure, which is named change detection, has the capability to capture worthwhile information for environmental research, particularly research that is concentrated on urban development and smart cities, hydrological cycles, climate change, city management, and deforestation; therefore, it can be an essential factor for local and global development project plans [2]. In recent decades, human activities in the world have caused huge changes on the earth, causing remote sensing resources such as satellite images to become an important tool for simple, swift, and accurate change detection over vast areas. The methods which are used to gain information and detect changes through remote sensing data are categorized into two parts: semi-automatic and automatic [2].
The change detection techniques in images which are known as two-dimensional remote sensing data are generally grouped into two categories: supervised and unsupervised methods. The supervised methods are identified to be resistant to atmospheric condition changes, brightness, and weak sensor calibration during data capturing, which means that the supervised methods are not sensitive to radiometric correction but have a significant weakness when used on terrestrial reference data because collecting the reference data is time-consuming, costly, and complicated [2]. The unsupervised methods do not require terrestrial reference data as they compare between pixels directly; however, they are sensitive to radiometric correction [2]. Table 1 briefly summarizes traditional change detection approaches and their advantages and limitations [3,4,5,6].
Given the merits and demerits of traditional change detection methods outlined in Table 1, it would be ideal to find a method which uses the features contained in images for change detection and thus improve on the drawbacks of these methods. In recent years, automatic deep learning networks have been highly effective in extracting high-level features from images for change detection purposes. Generally, deep learning methods for change detection include five categories: Convolutional Neural Networks (CNNs), Autoencoders (AEs) or stack Autoencoders (SAEs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), and Deep Belief Networks (DBNs) [16]. This study focuses on urban change detection of multispectral satellite images. The previous research on deep learning change detection methods and their advantages and limitations of multispectral remote sensing datasets are investigated briefly in Table 2.
Regarding Table 2, most studies have focused on AEs and CNNs for the change detection of multispectral remote sensing data. CNNs have some advantages for change detection purposes such as high accuracy, high performance, extract robust features, and end-to-end training; however, they have limitations such as requiring a time-consuming and huge amount of training data. One of the main networks in this category is the Siamese network. CNNs change detection networks are used for urban land cover or land use, building change detection, and water land use. AEs are simple models and employ end-to-end training which extracts deep features efficiently. AEs have some limitations such as computational complexity, a requirement for a large amount of training data, and poor performance in boundary detection. Unet is one of the main AE networks for change detection. AEs are usually applied to urban, land cover, land use, forest, and building change detection. GANs are usually used for training data generation and can significantly improve the performance of deep learning networks, depending on the amount of training data. According to Table 2, the main limitations of previous deep learning methods include model complexity, time-consuming networks, poor performance in detecting the shape and boundary of objects, and small changes. Considering the limitations of previous studies, we propose a method based on semi-transfer learning. This method uses a pre-trained EfficientNetV2 T network [47] with faster training speed and efficient parameters as an encoder part for feature extraction and convolutional layers of Unet as a decoder part for the reconstruction binary change map. The Semi Transfer Learning Change Detection EfficientNetV2 T-Unet (STCD-EffV2T Unet) method not only has less complexity but also has a faster training speed. Besides, most previous studies have concentrated on urban and building change detection datasets with less complexity and diversity, while the area in our study was selected from northern Iran, in which the urban areas are highly complex. This area consists of three parts: the first part is the coastline and urban area near water or wetland, the second part is the urban area which is located on foothills and mountainous terrain, and the last part is the urban areas which are covered mainly by vegetation or are surrounded by forest and agricultural land. Our proposed method can detect changes with high precision in these complex urban areas. Furthermore, the performance of the proposed method is evaluated using OSCD, which is one of the most famous datasets for urban change detection.
The main purpose of this study is to detect changes in these three kinds of areas that have always been challenging, using our proposed method to overcome this challenge. In comparison with other studies, our proposed method is not time-consuming or complex. Moreover, even using a focal loss function that is not effective in terms of change detection results, its performance will be adequate. The advantages of using transfer learning with different inputs make our method stronger than the others.
The rest of this study is organized as follows. The Section 2 contains the datasets that are used in this research and the properties of the study areas. In the Section 3, our proposed method is investigated in detail and the results are demonstrated in the Section 4. The comparisons between the performances of the methods are collected in Section 5. The conclusion is included in the Section 6.

2. Materials and Datasets

2.1. Sentinel-2 Satellite Image of the Northern Iran Dataset

The Sentinel-2 satellite images from northern Iran’s urban areas, especially located in the Gilan and Mazandaran provinces, are used in this study. The urban areas in these provinces are chiefly located on the margins of forests and farmlands as well as on the coastline, near water or wetland, and also on mountainous terrains or foothills. Most of the changes in these areas are located near forests and agricultural lands [48,49]. The changes are related to the expansion of urban areas horizontally due to the construction of tourism facilities, countryside houses, and resorts.
This research uses Sentinel-2 satellite images captured on 2 July 2017, and 10 August 2021, and their corresponding ground truth produced by using Google Earth images taken simultaneously. Change polygons are drawn employing ENVI and ArcGIS software. The Sentinel-2 images include 13 bands in the short-wave infrared, near-infrared, and ultraviolet parts of the spectrum, with resolutions ranging between 10 and 60 m (the multispectral satellite images are processed from level-0 to level-2A by Payload Data Ground Segment (PDGS). Only level-1C and level-2A products are released to users. In this study, we use the user available products as dataset). Figure 1 illustrates the study area. This dataset will be uploaded here: [http://rslab.ut.ac.ir, accessed on 20 February 2023].

2.2. Onera Sentinel-2 Satellite Change Detection (OSCD) Dataset

The second dataset adopted in this research is the OSCD dataset, which is a dataset of urban changes that can help to demonstrate the performance of our proposed method regarding complex urban changes. The OSCD dataset has been captured using Sentinel-2 images taken from places with various levels of urbanization in several other countries that have experienced rapid urban development and growth. This dataset includes the data of 24 important urban areas worldwide with ground truths. Figure 2 presents the RGB images of the five cities of Rio, Montpellier, Rennes, Chongqing, and Beihai at two times and their location in the world. The OSCD dataset is also available at https://rcdaudt.github.io/oscd/, accessed on 6 June 2020. In Table 3, the descriptions of the two datasets are illustrated.

3. Proposed Method

In this section, the proposed method is described in detail. As Figure 3 shows, the general procedure of the proposed method includes three main steps: (1) pre-processing, (2) data augmentation and train test split, and (3) encoder-decoder architecture. The main steps will be described in the following sub-sections.

3.1. Pre-Processing

In the pre-processing step, the RGB channels of two datasets are normalized using Equation (1), then they are divided into 128 × 128 patches, and, finally, the patches which contain the higher portion of the target in urban areas are chosen. Moreover, the ground truth is converted into 128 × 128 patches so that it is suitable as an input of the STCD-EffV2T Unet.
x d min d max + d min
In Equation (1),  d min  is the minimum gray value in each channel,  d max   is the maximum gray value in each channel, and x is the gray value of each pixel.

3.2. Train and Test Split and Augmentation

After pre-processing, the evaluation and training data are separated from the test data. A total of 67% of all the data are considered as training and evaluation, and the remaining 33% are applied to testing. In the next step, augmentation is used for the training dataset, as a result of which the neural network will resist the changes in the augmentation domain and boost the network’s efficiency [50]. Figure 4 shows data augmentation for a single image. The augmentation of this study includes rotation with three angles (+90°, −90°, and 180°).
In the North of Iran dataset, the training patches raise from 1545 to 4140 and in OSCD they raise from 495 to 1980, in which surplus is the result of augmentation.

3.3. Encoder-Decoder Architecture

In this section, we describe the EfficientNetV2 T-Unet as an encoder-decoder architecture for binary change detection. EfficientNetV2 T is used as a feature extractor in the encoder stage and the convolutional layers of the Unet network are used as decoders.

3.3.1. EfficientNet Encoder

The EfficientNet network families were proposed for the first time by [51]. In previous studies, the EfficientNetV1 as an encoder and convolutional layers of Unet as a decoder were used for semantic segmentation at environments [52], disease detection with CT scan and colonoscopy images [53,54], surface defect detection [55], blood vessel segmentation [56], and neuron instance segmentation [57]. In this study, we use the EfficientNetV2 as an encoder, which was proposed for the first time by [47]. In the EfficientNetV2, a combination of training-aware neural architecture reaches and scaling is used to optimize training speed and parameter efficiency. The EfficientNetV2 has versions B0, B1, B2, B3, L, M, S, and T. To obtain the maximum advantages of the EfficientNetV2 networks, we evaluate them as encoders of our proposed method (the results are presented in Section 4). Since EfficientNetV2 T comprises an efficient number of parameters and a high speed in the training process, it is chosen as the encoder to extract features. In addition, it reached better accuracy.
The fundamental building blocks of EfficientNetV2 are the mobile inverted bottleneck convolution (MBConv) [52] and fused mobile inverted bottleneck convolution (Fused-MBConv) [47]. The structures of these two blocks are shown in Figure 5.
The EfficientNetV2 networks include seven blocks. In blocks 1–3, MBConv is replaced with Fused-MBConv; however, in the EfficientNetV1, all blocks are MBConv. Fused-MBConv improves training speed with a slight overhead on parameters and FLOPs. If, in all blocks (1–7), MBConv is replaced with Fused-MBConv, the training speed will reduce and increase the FLOPs and model complexity. The network architecture of the EfficientNetV2 T used in this study is illustrated in Figure 6.

3.3.2. Transfer Learning with Different Input Channels

Most models that are similar to EfficientNetV2 were trained by datasets such as ImageNet and have three channels as input. For change detection, the pre-trained model should be converted from three channels into six channels (three RGB Time1, three RGB Time2). In previous studies, the three methods that were used are as follows:
(1) Expand the weight dimensions to account for additional channels and randomly initialize the value.
(2) Similar to the first method; however, instead of using a random value for filling, using the mean of other values.
(3) A second parallel network is created which has the same architecture as the pre-trained network but has different input channels. This new parallel network performs feature extraction on the remaining channels. Then, the output of this network is concatenated with the output of the original pre-trained network. In this method, the parallel network learns the representation specific to additional channels, and we still take advantage of using the pre-trained model as well.
In this study, we tried all three methods. For some networks, such as EfficientNet and Yolox, the second method has better performance; however, for DeepLabv3+ and VGG19, the third method works properly. In this study, we use the “Semi-Transfer Learning” technique in the sense that the encoder path (EfficientNet V2T) was pre-trained by the ImageNet dataset, although the decoder path (convolutional layers of Unet) was not trained.

3.3.3. Unet Decoder

Unet is a U-shaped, fully convolutional neural network which was used for the first time by Olaf and Ronneberger to segment medical images [58]. The Unet network includes the encoder and decoder paths. The task of the encoder path is to detect the backgrounds of images. The output of the encoder is smaller than the input; however, it is expanded in the second path, called the decoder. The second path detects the exact position of features by transpose convolution [59]. The decoder incorporates the high-level features by a subsequence of transpose and concatenation with the corresponding feature maps from the encoder path. The feature maps captured from the encoder path contain low-level features. In contrast, these low-level features contain helpful information about complex scenes with multiple objects and their relative configuration. They are combined with intermediate high-level feature maps from Unet, the decoder path, and low-level feature maps from EfficientNetV2. The network boosts context information on higher resolution layers by using many feature channels in the up-sampling part [52].
Our proposed method uses EfficientNetV2 T as an encoder instead of a set of convolutional layers and the decoder part is the convolutional layers of Unet. Figure 6 illustrates the proposed architecture in detail.

3.3.4. Loss Function

We implement the focal binary loss function which is usually used for unbalanced binary classification [60]. The focal loss can significantly decrease the training and validation loss during the training process. This loss function is defined using the following equation:
FL ρ = α 1 ρ γ log ρ .
In Equation (2),  ρ 0 , 1  is the model-estimated probability for the class with label 1 and  1 ρ  for the class with label 0, and  α  and  γ  are two hyperparameters for which  α  shows a weight balancing factor for class 1 (default is 0.25 as mentioned in [60]) and  γ  is a focusing parameter that is used to compute the focal factor (default is 2.0 as mentioned in [60]). In this study, we consider  α = 0.1  and  γ = 2.0  since the performance of the loss function with  α = 0.1  is better than the default value. Moreover, the Adam optimizer with a learning rate of  1   ×   10 4  and decay rate of   1   ×   10 6  is considered to adjust network parameters.

3.3.5. Accuracy Assessment

Accuracy assessment is a non-separable part of any remote sensing task. In this study, the final results of the proposed method are compared with the ground truth and other change detection networks quantitatively and qualitatively. The quantitative comparison is based on the metrics which are described subsequently (Table 4).

3.3.6. Comparative Methods

To compare the efficiency and speed of our proposed method, we compare our method with other EfficientNetV2 series including B0, B1, B2, B3, L, M, and S. Other EfficientNet series have a different number of parameters; however, they have a nearly similar general architecture. In addition, the following networks are used to confirm the efficiency of STCD-EffV2T Unet. These approaches are YoloX [61], ResNest [62], VGG19 [63], DeepLabv3+ [64], and U2Net [65], which are briefly described below:
  • YoloX [61]: the Yolo family networks are generally used for object detection. These networks are fast and accurate, and trained on the COCO dataset. In this study, we use YoloX series such as Nano, Tiny, S, and X, which are trained on the COCO dataset as an encoder, and the convolutional layers of Unet as a decoder.
  • ResNest [62]: this network, named the split-attention network, includes four series: ResNest50, ResNest101, ResNest200, and ResNest269. The number of parameters increases according to the number of these networks. In this study, we use ResNest50, ResNest101, and ResNest200, which were trained by the ImageNet dataset as the encoder part and convolutional layers of Unet as the decoder part.
  • VGG19 [63]: this network is one of the most famous networks for many remote sensing tasks. In this study, we use VGG19 which was pre-trained by ImageNet as an encoder path and convolutional layers of Unet as a decoder path.
  • DeepLabV3+ [64]: the last modification of the DeepLab network is DeepLabV3+, uploaded at http://keras.io. This network is used for multiclass segmentation. In the architecture of the DeepLabV3+ network, the ResNet50 is the backbone which is pre-trained by ImageNet. In this study, we use DeepLabV3+ for binary change detection, and we change the input channels into six channels and share the weight with the third method in Section 3.3.4.
  • U2Net [65]: this network is one of the newest Unet network forms proposed for salient object detection. U2Net is a two-level nested U-structure. In the structure of this network, different residual U blocks are used. We compare the performance of this approach with our proposed method.

4. Experimental Results

In this study, the STCD-EffV2T Unet is implemented using the RGB channels of two datasets. The system configuration used for this study is Intel (R) core (TM)i7-7800X CPU 3.5 GHz, 32.0 GB installed RAM, and NVIDIA GeForce GTX 1050Ti. All networks are trained using the TensorFlow 2.10.0 platform and python 3.8.
One of the main goals of this study is to develop a fast network. Our proposed method can generate an accurate binary change detection map in a reasonable time. The comparison between training time and several other parameters of this method with other methods is shown in Table 5. The metrics’ results, introduced in Section 3.3.5, are also demonstrated.
Table 5 shows that the STCD-EffV2T Unet achieves an accuracy of 97.66% and KC achieves an accuracy of 0.67 after 100 epochs, which are the best results among other networks. Despite some exceptions, the training time increases by increasing the number of parameters. YoloXNano Unet, YoloXTiny Unet, and EffV2 B0 Unet have the shortest training time and their parameters are less than 5 M. However, the final change detection results show that these networks cannot overcome the complexity of the study areas. The precession, KC, and other metrics show that the performance of the STCD-EffV2T Unet is better than these three networks. In Table 5, we see that, when the number of the parameters exceeds eight million, the training of all the networks becomes time-consuming, whereas the final change detection map will not improve. The networks such as EffV2L Unet, EffV2M Unet, YoloXX Unet, U2Net, and ResNest family networks are time-consuming in terms of their training process, and the parameters and complexity of these networks are surplus for our binary change detection task. Among the YoloX series, YoloXNano, YoloXTiny, and YoloXS, which have the lowest training time, cannot detect binary changes. YoloXX has more than 20 million parameters; however, its performance is not satisfactory. Among the EffV2 family networks, the STCD-EffV2T Unet compromises between the number of parameters and the binary change detection accuracy. The EffV2 B0, EffV2 B1, EffV2 B2, and EffV2 B3 have fewer parameters than STCD-EffV2T Unet, and, therefore, cannot reach an accurate binary change map, whereas EffV2 L, M, and S have more parameters than STCD-EffV2T Unet; therefore, they are time-consuming and cannot reach an accurate final change map. Moreover, U2net and DeepLabV3+ are not as accurate as STCD-EffV2T Unet.

5. Discussion

The quantitative, qualitative, and visual analyses demonstrate that the STCD-EffV2T Unet works effectively in detecting changes in the urban areas of our datasets. The north of Iran dataset includes three parts: (1) coastal areas, which are near water and wetland, (2) urban areas which are covered with vegetation and surrounded by forest and agricultural land, and (3) urban areas which are located in foothills. In Figure 7, samples of these areas are considered in order to demonstrate the results. In Figure 7a,b, two coastal areas are considered. Detecting changes near water or the coast has always been challenging; however, our proposed method can dominate this challenge and detect these changes. Most of the changes in coastal areas are sparse and small; therefore, detecting them is difficult. However, our proposed method can detect them properly and preserve the edges almost correctly. Figure 7c–e shows the results of three urban area samples covered with vegetation or surrounded by forest or agricultural land. These kinds of areas cover most of the north of Iran; in fact, the majority of the changes took place in these areas. The changes in these urban areas are vast and most of them are the constructions of truism facilities and countryside houses on the border of urban areas; this trend has grown seriously in recent years. The changes in these areas are dense and our proposed method can detect most of them and properly preserve the edges. In Figure 7a,e (Time1) there are some clouds in images; however, our proposed method, without any pre-processing, can detect changes. In Figure 7f, the result of urban areas which are located on the foothills is shown. The urban areas on the foothills form a minority of the study areas and the changes in this area are sparse. Most of this urban area is located in the mountain’s shadow; therefore, dividing between shadow and changes is hard. However, our proposed method can defeat this challenge and detect the changes and preserve the edges properly.
We also show the performance of STCD-EffV2T Unet on OSCD. The results of our proposed method for five cities of OSCD are shown in Figure 8. Although this network is slightly ill-suited to the detection of continued changes and those in edge or border areas (e.g., continuous changes in Montpellier City), it has considerable accuracy in tracking changes. As demonstrated by Figure 8, the network has effectively detected many tiny changes in Beihai and Rennes, where there have been many tiny changes in the ground truth. As shown in Figure 8, the proposed method also performed relatively well in detecting changes in coastal and port cities such as Rio.
Figure 9 compares the performance of STCD-EffV2T Unet with other EffV2 networks. It shows five coastal ports (a–e), five urban areas covered or surrounded by forest or agricultural land (f–j), and two urban areas located on foothills (k and l). These 12 areas are in the north of Iran. Among the EffV2 series, EffV2B3 and EffV2B2 can detect changes nearly properly; however, the STCD-EffV2T Unet can detect changes more accurately. As shown by Figure 9, the STCD-EffV2T Unet can preserve the edges and shapes of changes. There is the highest similarity between STCD-EffV2T Unet results and the ground truth.
Figure 10 compares the performance of STCD-EffV2T Unet with the other networks, including the YoloX series, ResNest series, VGG19, U2Net, and DeepLabV3+. In Figure 10, the same study areas are selected which were used in Figure 9. YoloXNano and YoloXTiny have better performance among other methods; however, the STCD-EffV2T Unet is superior. The other methods are not only time-consuming, as reported in Section 4 (experimental results), but also cannot reach the best performance.

6. Conclusions

In this study, we developed the STCD-EffV2T Unet network architecture, which benefits from the advantages of transfer learning for urban change detection with the northern area of Iran and the OSCD dataset. In our method, the RGB channels of two datasets were assigned as input of the STCD-EffV2T Unet network. The STCD-EffV2T Unet network can detect changes effectively in complex urban areas by unique feature extraction through the EfficientNetV2, pre-trained by the ImageNet dataset in the encoding path and convolutional layers of Unet as a decoding path. Taking into account the training speed of the STCD-EffV2T Unet network, it can produce an accurate binary change map in a reasonable time. It also can serve as a valuable network for generating maps of urban northern areas of Iran and updating the RS and GIS databases. In addition to the speed improvement, it achieves a 97.66% and 98.79% accuracy and F1-score, respectively. The STCD-EffV2T Unet successfully detected the diversity changes in northern Iran’s urban areas, including port cities and the coastal regions, urban areas adjacent to the forests and agricultural lands, and urban areas located on foothills. In this study, we applied the focal loss function, which is particularly useful for simultaneously detecting small and dense changes and also preserving the edges. In future studies, researchers can develop and use STCD-EffV2T Unet to analyze other urban areas with different challenges or three-dimensional changes. It may also be worthwhile to design other deep learning method architectures, such as encoder-decoder networks, for binary change detection.

Author Contributions

All authors contributed to the study’s conception and design. M.G., M.H. and P.R. performed material preparation, data collection, and analysis. M.G. wrote the first draft of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request.

Acknowledgments

The author would like to thank the Editor and reviewers for their valuable comments on our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jia, L.; Wang, J.; Ai, J.; Jiang, Y. A hierarchical spatial-temporal graph-kernel for high-resolution SAR image change detection. Int. J. Remote Sens. 2020, 41, 3866–3885. [Google Scholar] [CrossRef]
  2. Du, P.; Wang, X.; Chen, D.; Liu, S.; Lin, C.; Meng, Y. An improved change detection approach using tri-temporal logic-verified change vector analysis. ISPRS J. Photogramm. Remote Sens. 2020, 161, 278–293. [Google Scholar] [CrossRef]
  3. Goswami, A.; Sharma, D.; Mathuku, H.; Gangadharan, S.M.P.; Yadav, C.S.; Sahu, S.K.; Pradhan, M.K.; Singh, J.; Imran, H. Change Detection in Remote Sensing Image Data Comparing Algebraic and Machine Learning Methods. Electronics 2022, 11, 431. [Google Scholar] [CrossRef]
  4. Shafique, A.; Cao, G.; Khan, Z.; Asad, M.; Aslam, M. Deep learning-based change detection in remote sensing images: A review. Remote Sens. 2022, 14, 871. [Google Scholar] [CrossRef]
  5. Shi, W.; Zhang, M.; Zhang, R.; Chen, S.; Zhan, Z. Change detection based on artificial intelligence: State-of-the-art and challenges. Remote Sens. 2020, 12, 1688. [Google Scholar] [CrossRef]
  6. Bai, T.; Wang, L.; Yin, D.; Sun, K.; Chen, Y.; Li, W.; Li, D. Deep learning for change detection in remote sensing: A review. Geo-Spat. Inf. Sci. 2022, 1–27. [Google Scholar] [CrossRef]
  7. Bruzzone, L.; Prieto, D.F. Automatic analysis of the difference image for unsupervised change detection. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1171–1182. [Google Scholar] [CrossRef] [Green Version]
  8. Chen, J.; Gong, P.; He, C.; Pu, R.; Shi, P. Land-use/land-cover change detection using improved change-vector analysis. Photogramm. Eng. Remote Sens. 2003, 69, 369–379. [Google Scholar] [CrossRef] [Green Version]
  9. Gong, P. Change detection using principal component analysis and fuzzy set theory. Can. J. Remote Sens. 1993, 19, 22–29. [Google Scholar] [CrossRef]
  10. Celik, T. Unsupervised change detection in satellite images using principal component analysis and k-means clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
  11. Silván-Cárdenas, J.L.; Wang, L. On quantifying post-classification subpixel landcover changes. ISPRS J. Photogramm. Remote Sens. 2014, 98, 94–105. [Google Scholar] [CrossRef]
  12. Yuan, F.; Sawaya, K.E.; Loeffelholz, B.C.; Bauer, M.E. Land cover classification and change analysis of the Twin Cities (Minnesota) Metropolitan Area by multitemporal Landsat remote sensing. Remote Sens. Environ. 2005, 98, 317–328. [Google Scholar] [CrossRef]
  13. Bovolo, F.; Bruzzone, L.; Marconcini, M. A novel approach to unsupervised change detection based on a semisupervised SVM and a similarity measure. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2070–2082. [Google Scholar] [CrossRef] [Green Version]
  14. Healey, S.P.; Cohen, W.B.; Yang, Z.; Brewer, C.K.; Brooks, E.B.; Gorelick, N.; Hernandez, A.J.; Huang, C.; Hughes, M.J.; Kennedy, R.E. Mapping forest change using stacked generalization: An ensemble approach. Remote Sens. Environ. 2018, 204, 717–728. [Google Scholar] [CrossRef]
  15. Blaschke, T. Object based image analysis for remote sensing. ISPRS J. Photogramm. Remote Sens. 2010, 65, 2–16. [Google Scholar] [CrossRef] [Green Version]
  16. Khelifi, L.; Mignotte, M. Deep learning for change detection in remote sensing images: Comprehensive review and meta-analysis. IEEE Access 2020, 8, 126385–126400. [Google Scholar] [CrossRef]
  17. Cao, G.; Wang, B.; Xavier, H.-C.; Yang, D.; Southworth, J. A new difference image creation method based on deep neural networks for change detection in remote-sensing images. Int. J. Remote Sens. 2017, 38, 7161–7175. [Google Scholar] [CrossRef]
  18. Daudt, R.C.; Le Saux, B.; Boulch, A. Fully convolutional siamese networks for change detection. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP); IEEE: New York, NY, USA, 2018; pp. 4063–4067. [Google Scholar]
  19. Zhang, W.; Lu, X. The spectral-spatial joint learning for change detection in multispectral imagery. Remote Sens. 2019, 11, 240. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, M.; Tan, K.; Jia, X.; Wang, X.; Chen, Y. A deep siamese network with hybrid convolutional feature extraction module for change detection based on multi-sensor remote sensing images. Remote Sens. 2020, 12, 205. [Google Scholar] [CrossRef] [Green Version]
  21. Lin, Y.; Li, S.; Fang, L.; Ghamisi, P. Multispectral change detection with bilinear convolutional neural networks. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1757–1761. [Google Scholar] [CrossRef]
  22. Seydi, S.T.; Hasanlou, M.; Amani, M. A new end-to-end multi-dimensional CNN framework for land cover/land use change detection in multi-source remote sensing datasets. Remote Sens. 2020, 12, 2010. [Google Scholar] [CrossRef]
  23. Zhang, M.; Shi, W. A feature difference convolutional neural network-based change detection method. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7232–7246. [Google Scholar] [CrossRef]
  24. Yang, L.; Chen, Y.; Song, S.; Li, F.; Huang, G. Deep Siamese networks based change detection with remote sensing images. Remote Sens. 2021, 13, 3394. [Google Scholar] [CrossRef]
  25. Jiang, X.; Li, G.; Zhang, X.-P.; He, Y. A semisupervised Siamese network for efficient change detection in heterogeneous remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–18. [Google Scholar] [CrossRef]
  26. Chen, H.; Shi, Z. A spatial-temporal attention-based method and a new dataset for remote sensing image change detection. Remote Sens. 2020, 12, 1662. [Google Scholar] [CrossRef]
  27. Wang, D.; Chen, X.; Jiang, M.; Du, S.; Xu, B.; Wang, J. ADS-Net: An Attention-Based deeply supervised network for remote sensing image change detection. Int. J. Appl. Earth Obs. Geoinf. 2021, 101, 102348. [Google Scholar]
  28. Ji, S.; Wei, S.; Lu, M. Fully convolutional networks for multisource building extraction from an open aerial and satellite imagery data set. IEEE Trans. Geosci. Remote Sens. 2018, 57, 574–586. [Google Scholar] [CrossRef]
  29. Ma, J.; Shi, G.; Li, Y.; Zhao, Z. MAFF-Net: Multi-Attention Guided Feature Fusion Network for Change Detection in Remote Sensing Images. Sensors 2022, 22, 888. [Google Scholar] [CrossRef]
  30. Papadomanolaki, M.; Verma, S.; Vakalopoulou, M.; Gupta, S.; Karantzalos, K. Detecting urban changes with recurrent neural networks from multitemporal Sentinel-2 data. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium; IEEE: New York, NY, USA, 2019; pp. 214–217. [Google Scholar]
  31. Wiratama, W.; Lee, J.; Sim, D. Change detection on multi-spectral images based on feature-level U-Net. IEEE Access 2020, 8, 12279–12289. [Google Scholar] [CrossRef]
  32. Chen, H.; Wu, C.; Du, B. Towards deep and efficient: A deep Siamese self-attention fully efficient convolutional network for change detection in VHR images. arXiv 2021, arXiv:2108.08157. [Google Scholar]
  33. Zhang, H.; Wang, M.; Wang, F.; Yang, G.; Zhang, Y.; Jia, J.; Wang, S. A novel squeeze-and-excitation W-net for 2D and 3D building change detection with multi-source and multi-feature remote sensing data. Remote Sens. 2021, 13, 440. [Google Scholar] [CrossRef]
  34. Ding, Q.; Shao, Z.; Huang, X.; Altan, O. DSA-Net: A novel deeply supervised attention-guided network for building change detection in high-resolution remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102591. [Google Scholar] [CrossRef]
  35. Wang, Y.; Gao, L.; Hong, D.; Sha, J.; Liu, L.; Zhang, B.; Rong, X.; Zhang, Y. Mask DeepLab: End-to-end image segmentation for change detection in high-resolution remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2021, 104, 102582. [Google Scholar] [CrossRef]
  36. Wei, H.; Chen, R.; Yu, C.; Yang, H.; An, S. BASNet: A Boundary-Aware Siamese Network for Accurate Remote-Sensing Change Detection. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  37. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A deeply supervised image fusion network for change detection in high resolution bi-temporal remote sensing images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  38. Raza, A.; Huo, H.; Fang, T. EUNet-CD: Efficient UNet++ for change detection of very high-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  39. Papadomanolaki, M.; Vakalopoulou, M.; Karantzalos, K. A deep multitask learning framework coupling semantic segmentation and fully convolutional LSTM networks for urban change detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7651–7668. [Google Scholar] [CrossRef]
  40. Ahangarha, M.; Shah-Hosseini, R.; Saadatseresht, M. Deep learning-based change detection method for environmental change monitoring using sentinel-2 datasets. Environ. Sci. Proc. 2020, 5, 15. [Google Scholar]
  41. Mou, L.; Zhu, X.X. A recurrent convolutional neural network for land cover change detection in multispectral images. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium; IEEE: New York, NY, USA, 2018; pp. 4363–4366. [Google Scholar]
  42. Gong, M.; Yang, Y.; Zhan, T.; Niu, X.; Li, S. A generative discriminatory classified network for change detection in multispectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 321–333. [Google Scholar] [CrossRef]
  43. Luo, X.; Li, X.; Wu, Y.; Hou, W.; Wang, M.; Jin, Y.; Xu, W. Research on change detection method of high-resolution remote sensing images based on subpixel convolution. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1447–1457. [Google Scholar] [CrossRef]
  44. Alvarez, J.L.H.; Ravanbakhsh, M.; Demir, B. S2-cGAN: Self-supervised adversarial representation learning for binary change detection in multispectral images. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium; IEEE: New York, NY, USA, 2020; pp. 2515–2518. [Google Scholar]
  45. Zhang, Y.; Deng, M.; He, F.; Guo, Y.; Sun, G.; Chen, J. FODA: Building change detection in high-resolution remote sensing images based on feature–output space dual-alignment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8125–8134. [Google Scholar] [CrossRef]
  46. Saha, S.; Mou, L.; Qiu, C.; Zhu, X.X.; Bovolo, F.; Bruzzone, L. Unsupervised deep joint segmentation of multitemporal high-resolution images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8780–8792. [Google Scholar] [CrossRef]
  47. Tan, M.; Le, Q. Efficientnetv2: Smaller models and faster training. In Proceedings of the 38th International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  48. Bigdeli, B.; Gomroki, M.; Pahlavani, P. Generation of digital terrain model for forest areas using a new particle swarm optimization on LiDAR data. Surv. Rev. 2020, 52, 115–125. [Google Scholar] [CrossRef]
  49. Gomroki, M.; Jafari, M.; Sadeghian, S.; Azizi, Z. Application of intelligent interpolation methods for DTM generation of forest areas based on LiDAR data. PFG-J. Photogramm. Remote Sens. Geoinform. Sci. 2017, 85, 227–241. [Google Scholar] [CrossRef]
  50. Khoshboresh-Masouleh, M.; Alidoost, F.; Arefi, H. Multiscale building segmentation based on deep learning for remote sensing RGB images from different sensors. J. Appl. Remote Sens. 2020, 14, 034503. [Google Scholar] [CrossRef]
  51. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  52. Baheti, B.; Innani, S.; Gajre, S.; Talbar, S. Eff-unet: A novel architecture for semantic segmentation in unstructured environment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Virtual, 14–19 June 2020; pp. 358–359. [Google Scholar]
  53. Niu, H.; Lin, Z.; Zhang, X.; Jia, T. Image Segmentation for pneumothorax disease Based on based on Nested Unet Model. In Proceedings of the 2022 3rd International Conference on Computer Vision, Image and Deep Learning & International Conference on Computer Engineering and Applications (CVIDL & ICCEA), Changchun, China, 20–22 May 2022; IEEE: New York, NY, USA, 2022; pp. 756–759. [Google Scholar]
  54. Thanh, N.C.; Long, T.Q. CRF-EfficientUNet: An Improved UNet Framework for Polyp Segmentation in Colonoscopy Images With Combined Asymmetric Loss Function and CRF-RNN Layer. IEEE Access 2021, 9, 156987–157001. [Google Scholar]
  55. Üzen, H.; Turkoglu, M.; Aslan, M.; Hanbay, D. Depth-wise Squeeze and Excitation Block-based Efficient-Unet model for surface defect detection. Vis. Comput. 2022, 1–20. [Google Scholar] [CrossRef]
  56. Mathews, M.R.; Anzar, S.M.; Krishnan, R.K.; Panthakkan, A. EfficientNet for retinal blood vessel segmentation. In Proceedings of the 2020 3rd International Conference on Signal Processing and Information Security (ICSPIS), Dubai, United Arab Emirates, 25–26 November 2020; IEEE: New York, NY, USA, 2020; pp. 1–4. [Google Scholar]
  57. Wu, H.; Souedet, N.; Jan, C.; Clouchoux, C.; Delzescaux, T. A general deep learning framework for neuron instance segmentation based on efficient UNet and morphological post-processing. Comput. Biol. Med. 2022, 150, 106180. [Google Scholar] [CrossRef]
  58. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  59. Peng, D.; Zhang, Y.; Guan, H. End-to-end change detection for high resolution satellite images using improved UNet++. Remote Sens. 2019, 11, 1382. [Google Scholar] [CrossRef] [Green Version]
  60. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE international Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  61. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  62. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R. Resnest: Split-attention networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2736–2746. [Google Scholar]
  63. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  64. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
Figure 1. Map of study location in Iran (red), Sentinel-2 images of study location in 2017 (top) and 2021 (bottom).
Figure 1. Map of study location in Iran (red), Sentinel-2 images of study location in 2017 (top) and 2021 (bottom).
Remotesensing 15 01232 g001
Figure 2. Multi-temporal image datasets. (a) Rio, (b) Montpellier, (c) Rennes, (d) Chongqing, and (e) Beihai.
Figure 2. Multi-temporal image datasets. (a) Rio, (b) Montpellier, (c) Rennes, (d) Chongqing, and (e) Beihai.
Remotesensing 15 01232 g002
Figure 3. Flowchart of the proposed method.
Figure 3. Flowchart of the proposed method.
Remotesensing 15 01232 g003
Figure 4. Original image (a); data augmentation with −90° (b), +90° (c), and 180° (d).
Figure 4. Original image (a); data augmentation with −90° (b), +90° (c), and 180° (d).
Remotesensing 15 01232 g004
Figure 5. The structure of MBConv and Fused-MBConv blocks.
Figure 5. The structure of MBConv and Fused-MBConv blocks.
Remotesensing 15 01232 g005
Figure 6. Top: the structure of EfficientNetV2 T, which contains seven blocks (represented with a different color); bottom: the structure of our proposed method using EfficientNetV2 T as an encoder and the convolutional layers of Unet as a decoder. The encoder is concatenated with the decoder at four different resolutions (Block2, Block3, Block4, and Block6).
Figure 6. Top: the structure of EfficientNetV2 T, which contains seven blocks (represented with a different color); bottom: the structure of our proposed method using EfficientNetV2 T as an encoder and the convolutional layers of Unet as a decoder. The encoder is concatenated with the decoder at four different resolutions (Block2, Block3, Block4, and Block6).
Remotesensing 15 01232 g006
Figure 7. Proposed method results for coastal areas (a,b), urban areas surrounded by forest or agricultural land (ce) and urban areas located on foothills (f) (top row contains Time 1 (2017) and Time 2 (2021) images, respectively; bottom row contains ground truth and proposed method results, respectively).
Figure 7. Proposed method results for coastal areas (a,b), urban areas surrounded by forest or agricultural land (ce) and urban areas located on foothills (f) (top row contains Time 1 (2017) and Time 2 (2021) images, respectively; bottom row contains ground truth and proposed method results, respectively).
Remotesensing 15 01232 g007aRemotesensing 15 01232 g007b
Figure 8. Results of the proposed method in the cities of Rio (a), Montpellier (b), Rennes (c), Chongqing (d), and Beihai (e) (from left to right: Time 1 image, Time 2 image, ground truth, and proposed method results).
Figure 8. Results of the proposed method in the cities of Rio (a), Montpellier (b), Rennes (c), Chongqing (d), and Beihai (e) (from left to right: Time 1 image, Time 2 image, ground truth, and proposed method results).
Remotesensing 15 01232 g008
Figure 9. Results of coastal areas (ae), results of areas located near the forest and agricultural land (fj), and results of areas which are located on the foothills (k,l).
Figure 9. Results of coastal areas (ae), results of areas located near the forest and agricultural land (fj), and results of areas which are located on the foothills (k,l).
Remotesensing 15 01232 g009aRemotesensing 15 01232 g009b
Figure 10. Results of coastal areas (ae), results of areas located near the forest and agricultural land (fj), and results of areas which are located on the foothills (k,l).
Figure 10. Results of coastal areas (ae), results of areas located near the forest and agricultural land (fj), and results of areas which are located on the foothills (k,l).
Remotesensing 15 01232 g010
Table 1. Summary of the literature on traditional change detection approaches.
Table 1. Summary of the literature on traditional change detection approaches.
CategorySubcategoryDefinitionModeAdvantagesLimitationsApplications
Visual
Analysis
_Generate a change map by visual interpretationSupervisedHighly reliable resultsDifficult to update, for large applications time-consuming, labor-intensiveUsed in different fields before [3]
Algebra-based methodsImage differencingA change map is generated by performing algebraic operations or transformations.UnsupervisedSimple and easy to implement and interprets, it decreases the impact of sunglasses’ topography shadow.Difficult to choose the proper threshold to identify the change map, Difficult to choose appropriate image bands, there is no “from-to” change information, and the Distribution of results is not normal.Urban land use [7], Urban land use and cover [8]
Image regression
Image rationing
Chang vector analysis
Vegetation index differencing
TransformationPrinciple component analysis (PCA)A change map is produced by using transformation methods; these methods are utilized to suppress correlated information and highlight variance.UnsupervisedReduce the redundancy between bands, emphasize different information in taken componentsDetailed change information cannot extractRural-urban land cover [9], Land cover [10]
Tasseled Cap
Chi-Square
Gramm-Schmidt
Classification methodsPost- classification comparisonA change map is generated by a classification methodSupervised unsupervised hybridProvide change information matrix, do not need atmospheric correctionSelecting training data is challengingLand cover [11], Urban land cover [12], Urban land cover [13], Forest change detection [14]
Spectral-Temporal combined analysis
EM- transformed (Expectation Maximization)
Unsupervised change detection methods
Hybrid change detection
Artificial Neural Networks
Advanced MethodLi-Strahler reflectance modeltransform the spectral reflectance values into physically based parametershybridMore straightforward to comprehend than the spectral signature. Can derive vegetation informationComplicated and time-consuming, developing the proper mode is challenging.Land cover [3]
Spectral mixture model
Biophysical parameter method
GIS techniqueIntegrate GIS and RS methodsUse different data sources for change detectionhybridLand use information can update directly in the GIS databaseThe quality of the result change map depends on different data typeForest change detection [15]
GIS method
Table 2. Literature review of deep learning change detection approaches.
Table 2. Literature review of deep learning change detection approaches.
CategorySub-CategoryModeDataAdvantagesLimitationsApplications
CNNDeep brief networkSupervisedMultispectral SPOT-5 and Landsat images, google earth imagesHigh accuracyTime of processingUrban land use and vegetation [17]
CNNFully convolutional Siamese networkSupervisedMultispectral Onera Sentinel-2 Satellite images (OSCD)Trained end to endHuge amount of training dataUrban land use [18]
CNNSpectral, spatial joint learning networkSupervised Multispectral Taizhou and Kunshan datasetHigh performanceHuge amount of training dataUrban land use [19]
CNNSiamese deep network with hybrid convolutional feature extraction moduleSupervisedMultispectral ZY-3 and GF-2 satellite imagesExtraction robust deep features Could not separate pixel from its neighbor for classificationUrban rural-urban non-urban land use [20]
CNNBilinear convolutional neural networkSupervisedMultispectral Lansat-8 satellite imagesEnd-to-end trainingGenerating label data is challengingRiver and Waterland use [21]
CNNMultidimensional CNNUnsupervisedOSCDEnd to endTime-consumingUrban land use land cover [22]
CNNFeature difference convolutional neural networkPre-trainedMultispectral Worldview-3, QuickBird and Ziyuan-3 satellite imagesPowerful robustness and generalization ability Require huge amount of pixel-level training samplesUrban land use [23]
CNNDeep Siamese semantic segmentation networkSupervisedRGB building imagesDecrease training sample issuePoor performance in detecting the exact boundary of the buildingUrban construction [24]
CNNSemi-supervised Siamese network based on transfer learning Pre-trainedHaiti earthquake QuickBird satellite imagesDecrease computational costError mapUrban land cover [25]
CNNAttention mechanism-based deep supervision networkSupervisedMultispectral LEVIR CD dataset [26]High performancemode complexUrban land cover [27]
CNNMulti-Attention guided feature fusion networkSupervisedLEVIRCD [26], WHUCD [28]Enhance deep feature extraction and fusionA large number of parametersBuilding change detection [29]
AEMultispectral UnetSupervisedOSCDEnd to endLow performanceUrban land use [4]
AE and RNNCombination of Unet and robust Recurrent Networks such as LSTMSupervisedOSCDEnd to endA large amount of training dataUrban land use [30]
AEUnetSupervisedMultispectral KompSAT-3 satellite imagesSolve spectral distortion issueComputational complexityUrban land use and forest [31]
AEDeep self-attention fully efficient convolutional Siamese network SupervisedGoogle Earth multispectral season varying dataset End to endComplexity of modelUrban land use [32]
AEDeveloped from Unet and SeNetSupervisedMultispectral Wuhan dataset from IKONOSEnd to endA large amount of training data2D and 3D building change detection [33]
AEIntensely supervised attention-guided networkSupervisedLEVIRCD [26]Extract deep features efficientlyTime-consumingBuilding change detection [34]
AEHierarchical self-attention augmented the Laplacian pyramid SupervisedHigres satellite imagesExtracting deep featuresComplex modelUrban land use [4]
AEFeature regularized mask DeepLabSupervisedLEVIRCD [26], GF-1 satellite imagesEnd to endCannot detect edge properlyBuilding change detection [35]
AEBoundary-aware Siamese networkSupervisedLEVIRCD [26]End to end, sharp boundaryComplex modelUrban land use [36]
AEEfficient Unet++SupervisedLEVIRCD [26], CD dataset [37]Minimize computational parametersTime-consumingUrban land use [38]
AEMultitask learning framework L-UnetSupervisedOSCDSolve the illumination differences problem and registration errorA poor performance, especially in preserving object shapeUrban change detection [39]
AEUnetSupervisedOSCDSimple model, easy to implementCannot recognize small changesUrban change detection [40]
RNNRecurrent convolutional Neural NetworkSupervisedMultispectral Taizhou datasetEnd to endCannot extract all deep featureUrban land use [41]
GANGenerative discriminatory classified networkSupervisedMultispectral Worldview-2 and GF-1 satellite imagesDecrease training sample issueComplexity of modelUrban land use and water [42]
GANDeep GAN with improved DeepLabV3+UnsupervisedOSCD, Landsat-8, and google earth satellite imageshigh performanceHuge amount of training dataUrban land use [43]
GANSelf-supervised conditional GANSemi-supervisedMultispectral Worldview-2 satellite imagesExtract features at multiple resolutions Model complexityUrban land use [44]
GANFeature output space dual alignmentSupervisedLEVIRCD [26], WHUCD [28]Address the problem of the pseudo changesSuper-parameters α and β are issueBuilding change detection [45]
DBNDeep joint segmentationUnsupervisedMultispectral Sentinel-2 and Pleiades imagesAny labeled training pixel is not requiredTime-consumingUrban land use [46]
Table 3. Information on datasets used in this study.
Table 3. Information on datasets used in this study.
DatasetsTimeBandsSpectrum Range (µm)Spatial Resolution (m)
Onera Sentinel-2 satellite images Change Detection (OSCD)Time12015Blue0.45~0.5210
Green0.52~0.59
Time22018Red0.63~0.69
North of Iran Sentinel-2 Satellite imagesTime12017Blue0.45~0.5210
Green0.52~0.59
Time22021Red0.63~0.69
Table 4. Information formulas for accuracy assessment metrics.
Table 4. Information formulas for accuracy assessment metrics.
MetricFormula
Precision   TP TP + FP
F1-score   2 × TP 2 × TP + FP + FN
IOU   TP TP + FP + FN
Accuracy   TP + TN TP + FN + TN + FP
Kappa Coefficient (KC)   2 × TP × TN FN × FP TP + FP × FP + TN + TP + FN × FN + TN
Table 5. Quantitative evaluation of the results captured for two datasets.
Table 5. Quantitative evaluation of the results captured for two datasets.
MethodAccuracy (%)Precision (%)F1-Score (%)IOU (%)Kappa Coefficient (KC)Time of Training
(h min s)
Parameters (Million)
STCD-EffV2T Unet (proposed method)
North of Iran dataset
97.6699.6198.7997.600.672 h 10 min 34 s6.6 M
STCD-EffV2T Unet (proposed method)
OSCD
97.3298.4497.0596.340.595 min 0 s6.6 M
EffV2 B0 Unet97.4599.6098.6897.400.591 h 15 min 26 s4.3 M
EffV2 B1 Unet97.5499.2998.7297.480.631 h 50 min 30 s4.8 M
EffV2 B2 Unet97.3198.9898.6097.250.631 h 56 min 3 s5.2 M
EffV2 B3 Unet97.2398.8998.5697.160.622 h 5 min 32 s6.2 M
EffV2 L Unet97.1799.0398.5397.100.603 h 30 min 45 s26.0 M
EffV2 M Unet97.0399.1498.4696.970.563 h 38 min 56 s13.7 M
EffV2 S Unet97.2399.3098.5697.170.582 h 38 min 44 s8.8 M
YoloXNano Unet97.2099.0098.5497.130.601 h 17 min 3 s2.1 M
YoloXTiny Unet97.3799.0498.6697.240.611 h 15 min 22 s2.9 M
YoloXX Unet97.1799.0798.4597.130.513 h 53 min 46 s20.4 M
YoloXS Unet97.3899.0898.6597.340.551 h 18 min 41 s3.5 M
VGG19 Unet97.0499.0798.6297.240.581 h 28 min 34 s18.2 M
ResNest50 Unet97.0599.0598.4797.150.623 h 21 min 26 s16.5 M
ResNest101 Unet97.2799.0998.5897.210.613 h 41 min 0 s34.8 M
ResNest200 Unet95.5797.1597.6795.460.509 h 36 min 34 s56.8 M
DeepLabV3+92.8295.0396.2192.700.292 h 30 min 2 s17.0 M
U2Net97.3799.0398.6097.300.606 h 2 min 18 s44.0 M
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gomroki, M.; Hasanlou, M.; Reinartz, P. STCD-EffV2T Unet: Semi Transfer Learning EfficientNetV2 T-Unet Network for Urban/Land Cover Change Detection Using Sentinel-2 Satellite Images. Remote Sens. 2023, 15, 1232. https://doi.org/10.3390/rs15051232

AMA Style

Gomroki M, Hasanlou M, Reinartz P. STCD-EffV2T Unet: Semi Transfer Learning EfficientNetV2 T-Unet Network for Urban/Land Cover Change Detection Using Sentinel-2 Satellite Images. Remote Sensing. 2023; 15(5):1232. https://doi.org/10.3390/rs15051232

Chicago/Turabian Style

Gomroki, Masoomeh, Mahdi Hasanlou, and Peter Reinartz. 2023. "STCD-EffV2T Unet: Semi Transfer Learning EfficientNetV2 T-Unet Network for Urban/Land Cover Change Detection Using Sentinel-2 Satellite Images" Remote Sensing 15, no. 5: 1232. https://doi.org/10.3390/rs15051232

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop