Skip to Content
Remote SensingRemote Sensing
  • Article
  • Open Access

16 September 2021

Change Detection from SAR Images Based on Convolutional Neural Networks Guided by Saliency Enhancement

,
and
1
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
2
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.

Abstract

Change detection is an important task in identifying land cover change in different periods. In synthetic aperture radar (SAR) images, the inherent speckle noise leads to false changed points, and this affects the performance of change detection. To improve the accuracy of change detection, a novel automatic SAR image change detection algorithm based on saliency detection and convolutional-wavelet neural networks is proposed. The log-ratio operator is adopted to generate the difference image, and the speckle reducing anisotropic diffusion is used to enhance the original multitemporal SAR images and the difference image. To reduce the influence of speckle noise, the salient area that probably belongs to the changed object is obtained from the difference image. The saliency analysis step can remove small noise regions by thresholding the saliency map, and interest regions can be preserved. Then an enhanced difference image is generated by combing the binarized saliency map and two input images. A hierarchical fuzzy c-means model is applied to the enhanced difference image to classify pixels into the changed, unchanged, and intermediate regions. The convolutional-wavelet neural networks are used to generate the final change map. Experimental results on five SAR data sets indicated the proposed approach provided good performance in change detection compared to state-of-the-art relative techniques, and the values of the metrics computed by the proposed method caused significant improvement.

1. Introduction

The synthetic aperture radar (SAR) imaging process is not affected by sunlight, clouds, or the atmosphere because of the microwave imaging principle. In the field of SAR image processing, change detection is a very important topic. SAR images are an important information resource for change detection when studying disaster relief, agricultural detection, and urban planning, especially when evaluating the damage caused by natural disasters [1,2,3,4]. Because of the interference of scattering echo, speckle noise will inevitably be generated; it has the nature of multiplicative noise, and it seriously affects the interpretation of SAR images [5]. Therefore, noise suppression is an important task in the process of change detection. In recent decades, many change detection methods utilizing SAR imagery have been introduced, and these approaches can be divided into two categories: coherent and incoherent change detection methods, depending on whether interferometric phase information is used. In this paper, we mainly discuss incoherent change detection methods [6].
In terms of incoherent SAR image change detection algorithms, supervised and unsupervised methods are the most used. The main problem associated with supervised methods is the lack of ground reference data, which often involves manual marking processes that are labor-intensive and time-consuming. Unsupervised SAR change detection generally includes three basic steps: speckle noise reduction, difference map generation, and classification [7].
There are problems that influence the effect of SAR image change detection. First, the inherent speckle noise in the SAR images may cause false positives. If we use the denoising method directly, we can also remove useful information in the denoising process. Second, the difference image (DI) influences the detection results, and the changed information may be lost when obtaining the DI. Finally, because the prior information is used in the supervised models to train the classifier, the supervised models may generate a performance superior to the unsupervised models. Nevertheless, prior information is usually achieved by manual annotation, which requires much work and affects the generalization of the model. Hence, in terms of the SAR image change detection, effective denoising, feature extraction, and prior information acquisition ought to be considered.
As the result of these findings, for the sake of suppressing the speckle noise and preserving interest information, one saliency detection method is used to extract the interesting regions that probably pertain to the changed objects. To extract the changed information, the convolutional-wavelet neural networks (CWNNs) model is utilized to learn features from the denoised images and the difference image. The speckle reducing anisotropic diffusion (SRAD) is used to enhance the input images and the difference image. The saliency detection is performed on the enhanced difference image, and the interest regions are extracted. The hierarchical fuzzy c-means clustering (HFCM) model is used for pre-classification. Finally, the CWNN model is used to generate the final change map. Experiments on qualitative and quantitative comparisons demonstrate the advantages and innovations of the proposed SAR image change detection algorithm. The major contributions of the proposed method are concluded as follows.
(1)
A saliency detection model is used in the proposed method, which aims to generate the salient regions that probably belong to the changed objects. The saliency detection model can extract attractive and compact salient areas from the difference image with a simple operation. It can remove background pixels and suppress noise.
(2)
A hierarchical fuzzy c-means clustering (HFCM) model is introduced in the proposed method and is used to select pixels with high probability of becoming changed or unchanged. The samples from the changed and unchanged parts are selected as the training set for the convolutional neural network.
(3)
A convolutional neural network based on dual-tree complex wavelet transform is constructed that aims to enhance the accuracy of change detection.

3. Proposed SAR Image Change Detection Method

In this section, we elaborate on the proposed unsupervised SAR image change detection method. The proposed method can be divided into the following steps: difference image, extraction of salient regions, pre-classification, and classification by CWNN model. The log-ratio operator is used to generate the initial difference image, then the saliency detection is utilized to obtain the salient regions. The HFCM model is used for pre-classification. Finally, the CWNN model is used to generate the final change map. The flow chart of the proposed automatic change detection algorithm is given in Figure 1.
Figure 1. The flow chart of the proposed SAR image change detection model.

3.1. Difference Image Generated by Log-Ratio Operator

Given two co-registered multi-temporal SAR images I1 and I2, which are obtained from the same region at different times, t1 and t2, respectively, the purpose of SAR image change detection is to produce a difference image that reflects the change information between t1 and t2.
The initial SAR images will be affected by noise, and the speckle reducing anisotropic diffusion (SRAD) [9] is used to suppress the noise in multi-temporal SAR images I1 and I2, generating the denoised images X1 and X2 corresponding to I1 and I2, respectively. Figure 2 shows the example of the Ottawa images denoised by SRAD, and the peak signal-to-noise ratio (PSNR) index is used to measure the denoising effect; we can see that the noise is suppressed effectively from the area with the red box.
Figure 2. The example of SRAD denoising for Ottawa images. (a) Time 1 of Ottawa; (b) denoised image of Time 1 (PSNR = 24.0372); (c) Time 2 of Ottawa; (d) denoised image of Time 2 (PSNR = 25.2647).
In this section, the initial difference image (DI1) is generated by the log-ratio (LR) operator, then SRAD is performed on the DI1, and the denoised difference image DI2 is obtained. The corresponding equation is:
D I 1 = log X 2 X 1 = log X 2 log X 1

3.2. Extraction of Salient Regions

Visual saliency regions contain information for visual image processing. In this section, the saliency detection theory is adopted to guide the change detection of SAR images. The initial difference image has a strong contrast region, and this is the salient region. We utilize the saliency detection model to locate the similar-change areas and optimize the proposed change detection task.
Suppose the X p denotes the intensity value of one pixel p in the image X. The saliency value V(p) of the pixel p can be calculated by the following [54]:
V p = X p X 1 + X p X 2 + + X p X N
where N is the total number of pixels in X. When two pixels have the same value of intensity, their saliency values are equal. Equation (2) can be modified as follows:
V p = j = 0 L 1 M j X p X j
where j depicts the pixel intensity, M j shows the number of pixels with an intensity equal to j , and L denotes the number of gray levels.
The saliency map of the denoised difference image DI2 is calculated by Equations (2) and (3), and we define the saliency map as DS.

3.3. Preclassification

In this section, the automatic threshold Otsu model is used to generate the binarized saliency map D s , and it is calculated by the following [40]:
D s x , y = 1 p x , y τ 0 p x , y τ
where p x , y denotes the gray value of the pixel in the salient area D s . The parameter τ shows the threshold computed by the Otsu model. In terms of D s , the value 1 indicates the salient pixels, and the value 0 indicates the non-salient pixels. The regions corresponding to the multi-temporal SAR images are extracted with the D s , and the following equation is given:
d s i = X i D s
where X i i = 1 ,   2 shows the matrix generated by the denoised multi-temporal SAR images. ⊙ denotes the dot product operator.
The new salient difference image D2 is generated by utilizing the log-ratio model, and it is calculated by the following:
D 2 = log d s 1 + 1 d s 2 + 1
When the D 2 is generated, the hierarchical FCM clustering (HFCM) model [55] is used to classify the D 2 into three components: the changed class Ω c , the unchanged class Ω u , and the intermediate class Ω i . The Ω c and Ω u are selected as the training samples, and Ω i is further classified by CWNN. More details of the hierarchical FCM clustering are contained in reference [55].

3.4. Classification by CWNN

CWNN was developed from convolutional neural networks (CNNs); it consists of convolutional layers, max-pooling layers, and fully connected layers [52,53]. In the CWNN model, the dual-tree complex wavelet transform (DTCWT) is introduced into the CNN model to reduce the effect of speckle noise in the SAR images. DTCWT has the advantages of good direction selectivity, limited redundancy, and a good reconstruction effect. DTCWT can decompose the layer preceding the pooling layer into eight components, including two low-frequency sub-bands LL1 and LL2, and the high-frequency sub-bands in six orientations, ±15°, ±45°, and ±75° (given by LH1, LH2, HL1, HL2, HH1, and HH2) [56]. We chose the average of the two low-frequency sub-bands as the output of the pooling layer. Firstly, low-frequency components maintain the structures of the input layer according to the specified rules to better represent the patch of the input image. Secondly, some noises are suppressed by losing the high-frequency components.
In the CWNN method, the input of a wavelet pooling layer presents the output of the previous convolutional layer. In terms of each input feature map x, the DTCWT is utilized to generate the sub-bands:
LL 1 ,   LL 2 ,   LH 1 ,   LH 2 ,   HL 1 ,   HL 2 ,   HH 1 ,   HH 2 = f x i
where f denotes the DTCWT function. The average of the low-frequency components is adopted as the output of the wavelet pooling layer, and the corresponding formula is defined as follows:
LL mean = 1 2 LL 1 + LL 2
where LL mean represents the output of the wavelet pooling layer.
Figure 3 shows an example of the wavelet pooling layer. x i denotes one feature map after the convolutional layer. The eight sub-bands are generated by the DTCWT performed on the feature map, including two low-frequency sub-bands and six high-frequency sub-bands. The output feature map is achieved by averaging the two low-frequency sub-bands.
Figure 3. Wavelet pooling layer.
The structure of CWNN is depicted in Figure 4. C2 and C4 denote the two convolutional layers, W3 and W5 present the wavelet pooling layer. Hence, the network can be depicted as {I1, C2, W3, C4, W5, F6, O7}. I1 shows the input layer, and all the input image patches are resampled to 28 × 14. C2 presents the convolutional layer with six convolutional kernels of 5 × 3. This layer generates six feature maps sized 24 × 12. W3 shows the wavelet pooling layer. In this layer, all the input feature maps are decomposed with one-level DTCWT. This wavelet pooling layer generates six feature maps with the size of 12 × 6. C4 depicts the convolutional layer with 12 convolutional kernels of 5 × 3. This layer produces 12 feature maps sized 8 × 4. W5 shows the wavelet pooling layer, generating 12 feature maps sized 4 × 2. F6 represents the fully connected layer with 96 units. O7 presents the output layer with two units.
Figure 4. The structure of CWNN.
In this section, the real samples and virtual samples are used as the training samples. When the CWNN training is finished, the image patches from Ω i are classified into changed and unchanged components, and the final change map is generated by the results of pre-classification and CWNN classification.

4. Experimental Results and Discussions

4.1. Data Set Descriptions

In this section, four real SAR image data sets and one simulated SAR image data set were used to demonstrate the effectiveness of the proposed SAR image change detection method. Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 show the SAR image data sets. Figure 5 is the Ottawa data set of two SAR images sized 290 × 350, captured by Radarsat-1, and they were acquired in May and August 1997. Figure 6 is the Coastline data set of two SAR images in Dongying, China, sized 450 × 280, captured by Radarsat-2; they were obtained in June 2008 and June 2009. Figure 7 shows the De Gaulle Airport data set of two SAR images acquired by ERS-1; they were taken in July 1997 and October 1998. The size of each SAR image is 240 × 370. Figure 8 is the Wenchuan data set of two SAR images sized 442 × 301, captured by ESA/ASAR on 3 March 2008 and 16 June 2008. These data mainly reflect the change caused by earthquake. The fifth data set in Figure 9 is the simulated data set of SAR images of the size 335 × 470 related to the village of Feltwell in the U.K. An image captured by the Daedalus 1268 Airborne Thematic Mapper (ATM) multispectral scanner was used as the reference image, and this image was assumed to be the Time 1 image of the data set. The Time 2 image was artificially generated from the reference one, and the land cover change was simulated by inserting some changes into the Time 1 image. The ground-truth images in Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 were produced by artificial tagging with prior information. A description of the five data sets used in the experiment is given in Table 1.
Figure 5. Ottawa data. (a) Image acquired in May 1997; (b) image acquired in August 1997; (c) ground-truth image.
Figure 6. Coastline data. (a) Image acquired in June 2008; (b) image acquired in June 2009; (c) ground-truth image.
Figure 7. De Gaulle Airport data. (a) Image acquired in July 1997; (b) image acquired in October 1998; (c) ground-truth image.
Figure 8. Wenchuan data. (a) Image acquired 3 March 2008; (b) image acquired 16 June 2008; (c) ground-truth image.
Figure 9. Simulated data. (a) Image acquired at Time 1; (b) image acquired at Time 2; (c) ground-truth image.
Table 1. The five data sets used in the experiment.

4.2. Experimental Settings

In this section, 10 closely related algorithms are compared: PCAKM [57], change detection using log-ratio and Otsu (LROtsu) [34], change detection using mean-ratio and Otsu (MROtsu) [34], change detection using log-ratio and FCM (LRFCM) [58], change detection utilizing Gabor wavelet and two-level clustering (GaborTLC) [59], LMT [60], PCANet [50], NRELM [55], change detection using neighborhood-based ratio and collaborative representation (NRCR) [61], convolutional-wavelet neural networks for change detection (CWNN) [53]. We used the relevant parameter values of the original articles proposed by the authors. In the proposed method, the parameter’s value setting in the CWNN model is consistent with that in reference [53]. A total of 10,000 pixels were randomly selected from Ω u and Ω c as the real samples, and 10,000 virtual samples were generated based upon these real samples. Table 2 shows the parameter settings of different methods.
Table 2. The parameter settings of different methods.
To help judge the results of the change detection, five objective indicators were used as measures, namely, false positives (FPs) [50,62], false negatives (FNs) [50,63], overall errors (OEs) [50,64], percentage correct classification (PCC) [50,65], and kappa coefficient (KC) [50,66]. In the binary ground-truth image, we calculated the actual number of pixels belonging to the unchanged class (Nu) and the changed class (Nc). FP depicts the number of pixels belonging to the unchanged class but are falsely classified as the changed class. FN shows the number of pixels belonging to the changed class but are falsely classified as the unchanged class. The OE is the sum of FP and FN, and it is defined as follows:
OE = FP + FN
PCC is calculated by the following:
PCC = N u + N c FP FN N u + N c × 100 %
KC is defined as follows:
KC = PCC PRE 1 PRE
where
PRE = N c FN + FP N c + N u FP + FN N u N c + N u N c + N u
The quantitative evaluations of the change results computed by the proposed and comparison algorithms are summarized in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8.
Table 3. Quantitative measures of different methods for the Ottawa data set.
Table 4. Quantitative measures of different methods for the Coastline data set.
Table 5. Quantitative measures of different methods for the De Gaulle Airport data set.
Table 6. Quantitative measures of different methods for the Wenchuan data set.
Table 7. Quantitative measures of different methods for the simulated data set.
Table 8. The average quantitative measures of different methods for the five data sets.

4.3. Results and Discussions

The corresponding experimental results of the Ottawa data set are reported in Table 3 and Figure 10. In Figure 10, we notice that the change map obtained by the proposed method outperformed other algorithms, and yielded better local consistency and fewer isolated pixels. The change maps generated by LROtsu, MROtsu, and LRFCM had many error pixels, and the three algorithms had the worst capacity to suppress the noise. Therefore, the FP values of LROtsu, MROtsu, and LRFCM were relatively high, the values were twice the FP value obtained by the proposed algorithm, as shown in Table 3. The change maps computed by PCAKM, GaborTLC, and LMT missed some changed regions, but the detection of the unchanged area was more accurate. Therefore, the FN values of the corresponding three methods were higher, while the FP values were lower, which was consistent with the data shown in Table 3. The PCANet, NRELM, NRCR, and CWNN techniques missed some unchanged regions. The OE value of the proposed algorithm for the Ottawa data set was reduced by 521, 229, 949, and 285 over PCANet, NRELM, NRCR, and CWNN, respectively. The PCC value of the proposed technique was improved by 0.51, 0.22, 0.93, and 0.28% over PCANet, NRELM, NRCR, and CWNN, respectively. KC was a comprehensive evaluation index; the KC value of the proposed algorithm was improved by 1.97, 1.03, 3.36, and 1.10% over PCANet, NRELM, NRCR, and CWNN, respectively.
Figure 10. Change detection results of the Ottawa data set. (a) PCAKM; (b) LROtsu; (c) MROtsu; (d) LRFCM; (e) GaborTLC; (f) LMT; (g) PCANet; (h) NRELM; (i) NRCR; (j) CWNN; (k) Proposed; (l) Reference.
The simulation results for the Coastline data set are shown in Figure 11 and Table 4. From Figure 11, we see that the PCAKM, LROtsu, MROtsu, LRFCM, GaborTLC, and NRCR had many noise spots, and the visual effect of change detection was poor with large FP. The change map generated by LMT was an improvement compared to the previous algorithms, but it missed some changed regions with large FN. Compared to the PCANet, NRELM, and CWNN methods, the OE value of the proposed method for the Coastline data set was reduced by 13,920, 5819, and 13,330, respectively. The PCC value of the proposed algorithm was improved by 11.04, 4.62, and 10.58% over PCANet, NRELM, and CWNN, respectively. The KC value of the proposed method was improved by 69.53, 55.20, and 69.40% over PCANet, NRELM, and CWNN, respectively. From the analysis of the results for the Coastline data set, the change map computed by the proposed method had a better change detection effect.
Figure 11. Change detection results of the Coastline dataset. (a) PCAKM; (b) LROtsu; (c) MROtsu; (d) LRFCM; (e) GaborTLC; (f) LMT; (g) PCANet; (h) NRELM; (i) NRCR; (j) CWNN; (k) Proposed; (l) Reference.
The simulation results for the De Gaulle Airport data set are given in Figure 12 and Table 5. From Figure 12, we see that the PCAKM, LROtsu, MROtsu, and LRFCM methods yielded much noise with large FP. The change results computed by GaborTLC, PCANet, NRELM, NRCR, and CWNN were poor. The LMT missed more changed regions with the largest FN value. Compared with the change maps generated by other algorithms, our method had a better performance in change detection on the De Gaulle Airport data set. In Table 5, we can observe that the values of OE, PCC, and KC generated by the proposed method were the best. The PCC value computed by the proposed method was improved by 12.93, 7.36, 15.52, and 9.07% over PCANet, NRELM, NRCR, and CWNN, respectively. The KC value computed by the proposed method was improved by 39.15, 31.52, 43.26, and 33.52% over PCANet, NRELM, NRCR, and CWNN, respectively. This means that the detection accuracy of our algorithm performed on the De Gaulle Airport data set was the highest compared to other approaches.
Figure 12. Change detection results of the De Gaulle Airport data set. (a) PCAKM; (b) LROtsu; (c) MROtsu; (d) LRFCM; (e) GaborTLC; (f) LMT; (g) PCANet; (h) NRELM; (i) NRCR; (j) CWNN; (k) Proposed; (l) Reference.
The simulation results for the Wenchuan data set are given in Table 6 and Figure 13. In Figure 13, we see that LROtsu, MROtsu, and LRFCM generated poor change detection results with some noise points; these three methods detected more unchanged areas as changed, so the corresponding FP values were large. The change detection results computed by PCAKM, GaborTLC, LMT, and NRELM were similar, but the results were still not ideal. In Table 6, we see that the proposed method had the least number of overall errors (OEs), and the values of PCC and KC were the best. The OE value of the proposed method for the Wenchuan data set was reduced by 1406, 3341, and 4900 over PCANet, NRCR, and CWNN, respectively. The PCC value of the proposed technique was improved by 1.05, 2.51, and 3.68% over PCANet, NRCR, and CWNN, respectively. The KC value of the proposed approach was improved by 4.82, 12.03, and 17.75% over PCANet, NRCR, and CWNN, respectively. The display of these data was obviously consistent with the analysis of the above image results. The change map generated by the proposed technique had higher detection accuracy and was closer to the reference true value image.
Figure 13. Change detection results of the Wenchuan data set. (a) PCAKM; (b) LROtsu; (c) MROtsu; (d) LRFCM; (e) GaborTLC; (f) LMT; (g) PCANet; (h) NRELM; (i) NRCR; (j) CWNN; (k) Proposed; (l) Reference.
The simulation results for the simulated data set are shown in Table 7 and Figure 14. Some isolated noise points were generated by the PCAKM, LROtsu, MROtsu, LRFCM, and LMT algorithms, and the FP values computed by the corresponding five methods were high. GaborTLC performed better than the previously mentioned change detection approaches. PCANet, NRELM, NRCR, and CWNN missed some changed regions with a large FN, although the four methods were effective for noise suppression. The proposed method effectively suppressed the noise and achieved a better detection effect that was closer to the true value image. The PCC value computed by the proposed method was improved by 0.83, 1.45, 0.79, and 1.24% over PCANet, NRELM, NRCR, and CWNN, respectively. The KC value generated by the proposed method was improved by 20.32, 41.38, 19.41, and 33.18% over PCANet, NRELM, NRCR, and CWNN, respectively.
Figure 14. Change detection results of the simulated dataset. (a) PCAKM; (b) LROtsu; (c) MROtsu; (d) LRFCM; (e) GaborTLC; (f) LMT; (g) PCANet; (h) NRELM; (i) NRCR; (j) CWNN; (k) Proposed; (l) Reference.
To evaluate the superiority of the algorithm more accurately, we averaged the experimental data of five groups of SAR images, as shown in Table 8. Figure 15 shows the objective performance of different change detection algorithms on five SAR data sets. For each metric, the scores computed by an approach to different SAR data sets were connected to obtain a curve, and the average score is given in the legend, so we can directly see the fluctuation of data. In Table 8, we see that the values of OE, PCC, and KC computed by the proposed method were the best. Although the MROtsu method had the smallest miss detection rate, it had a higher false detection rate, and the values of FP and OE were the highest. LMT had the lowest false detection rate, but it had the highest miss detection rate, that is to say, the FN was the highest. In qualitative and quantitative analyses, our algorithm had absolute advantages in detection efficiency for SAR images.
Figure 15. Objective performance of different change detection algorithms on five SAR data sets. (a) FP; (b) FN; (c) OE; (d) PCC; (e) KC.

5. Conclusions

In this paper, an effective SAR image change detection algorithm based on saliency-guided convolutional neural networks was proposed. The saliency map was adopted to guide the search for the interest regions in the initial difference image computed by a log-ratio operator, and the noise in the saliency map could be suppressed to some extent with the Otsu method. Then an enhanced difference image was generated by using the binarized saliency map and denoised input images. The hierarchical fuzzy c-means was used for pre-classification, and the final change map was obtained by the CWNN model. The experimental results demonstrated the effectiveness of the proposed change detection technique. Because the traditional saliency detection and the clustering model methods were used in the proposed algorithm, it was not an end-to-end deep learning model for SAR image change detection, so in future work, an end-to-end CWNN model for image change detection is what we need to construct.

Author Contributions

The experimental measurements and data collection were carried out by L.L. and H.M. The manuscript was written by L.L. with the assistance of H.M. and Z.J. All authors reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shanghai Aerospace Science and Technology Innovation Fund under Grant No. SAST2019-048.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, H.; Zhang, K.; Xiao, W. Building change detection in very high-resolution remote sensing image based on pseudo-orthorectification. Int. J. Remote Sens. 2021, 42, 2686–2705. [Google Scholar] [CrossRef]
  2. Wang, C.; Wang, X. Building change detection from multi-source remote sensing images based on multi-feature fusion and extreme learning machine. Int. J. Remote Sens. 2021, 42, 2246–2257. [Google Scholar] [CrossRef]
  3. Huo, J.-Y.; Mu, L. Fast change detection method for remote sensing image based on method of connected area labeling and spectral clustering algorithm. J. Appl. Remote Sens. 2021, 15, 016506. [Google Scholar] [CrossRef]
  4. Zhao, J.; Liu, S.; Wan, J.; Yasir, M.; Li, H. Change detection method of high resolution remote sensing image based on DS evidence theory feature fusion. IEEE Access 2021, 9, 4673–4687. [Google Scholar] [CrossRef]
  5. Dong, H.; Ma, W.; Jiao, L.; Liu, F.; Li, L. A multiscale self-attention deep clustering for change detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2021, 1–16. [Google Scholar] [CrossRef]
  6. Jung, J.; Yun, S.H. Evaluation of coherent and incoherent landslide detection methods based on synthetic aperture radar for rapid response: A case study for the 2018 Hokkaido Landslides. Remote Sens. 2020, 12, 265. [Google Scholar] [CrossRef] [Green Version]
  7. Yang, G.; Li, H.; Wang, W. Unsupervised change detection based on a unified framework for weighted collaborative repre-sentation with RDDL and fuzzy clustering. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8890–8903. [Google Scholar] [CrossRef]
  8. Deledalle, C.-A.; Denis, L.; Tupin, F. Iterative weighted maximum likelihood denoising with probabilistic patch-based weights. IEEE Trans. Image Process. 2009, 18, 2661–2672. [Google Scholar] [CrossRef] [Green Version]
  9. Yu, Y.; Acton, S.T. Speckle reducing anisotropic diffusion. IEEE Trans. Image Process. 2002, 11, 1260–1270. [Google Scholar]
  10. Farhadiani, R.; Homayouni, S.; Safari, A. Hybrid SAR speckle reduction using complex wavelet shrinkage and non-local PCA-based filtering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1489–1496. [Google Scholar] [CrossRef]
  11. Devapal, D.; Kumar, S.S.; Jojy, C. A novel approach of despeckling SAR images using nonlocal means filtering. J. Indian Soc. Remote Sens. 2016, 45, 443–450. [Google Scholar] [CrossRef]
  12. Hu, T.; Li, W.; Liu, N.; Tao, R.; Zhang, F.; Scheunders, P. Hyperspectral Image restoration using adaptive anisotropy total variation and nuclear norms. IEEE Trans. Geosci. Remote Sens. 2021, 59, 1516–1533. [Google Scholar] [CrossRef]
  13. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A nonlocal SAR image denoising algorithm based on LLMMSE wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2012, 50, 606–616. [Google Scholar] [CrossRef]
  14. Zhou, Y.; Shi, J.; Yang, X.; Wang, C.; Kumar, D.; Wei, S.; Zhang, X. Deep multi-scale recurrent network for synthetic aperture radar images despeckling. Remote Sens. 2019, 11, 2462. [Google Scholar] [CrossRef] [Green Version]
  15. Argenti, F.; Lapini, A.; Bianchi, T.; Alparone, L. A tutorial on speckle reduction in synthetic aperture radar images. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–35. [Google Scholar] [CrossRef] [Green Version]
  16. Zheng, Y.; Zhang, X.; Hou, B.; Liu, G. Using combined difference image and k-means clustering for sar image change detection. IEEE Geosci. Remote Sens. Lett. 2014, 11, 691–695. [Google Scholar] [CrossRef]
  17. Su, L.; Gong, M.; Sun, B. Change detection in synthetic aperture radar images based on non-local means with ratio similarity measurement. Int. J. Remote Sens. 2014, 35, 7673–7690. [Google Scholar] [CrossRef]
  18. Lou, X.; Jia, Z.; Yang, J.; Kasabov, N. Change detection in SAR images based on the ROF model semi-implicit denoising method. Sensors 2019, 19, 1179. [Google Scholar] [CrossRef] [Green Version]
  19. Wang, X.; Jia, Z.; Yang, J. Change detection in SAR images based on the logarithmic transformation and total variation de-noising method. Remote Sens. Lett. 2017, 8, 214–223. [Google Scholar] [CrossRef]
  20. Liu, L.; Jia, Z.; Yang, J. SAR image change detection based on mathematical morphology and the k-means clustering algorithm. IEEE Access 2019, 7, 43970–43978. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Wang, C.; Ji, Y.; Chen, J.; Deng, Y.; Chen, J.; Jie, Y. Combining segmentation network and nonsubsampled contourlet transform for automatic marine raft aquaculture area extraction from sentinel-1 images. Remote Sens. 2020, 12, 4182. [Google Scholar] [CrossRef]
  22. Li, L.; Ma, H.; Jia, Z.; Si, Y. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 12389–12409. [Google Scholar] [CrossRef]
  23. Li, L.; Ma, H. Pulse coupled neural network-based multimodal medical image fusion via guided filtering and WSEML in NSCT domain. Entropy 2021, 23, 591. [Google Scholar] [CrossRef]
  24. Kollem, S.; Reddy, K.; Rao, D. Improved partial differential equation-based total variation approach to non-subsampled con-tourlet transform for medical image denoising. Multimed. Tools Appl. 2021, 80, 2663–2689. [Google Scholar] [CrossRef]
  25. Li, L.; Si, Y.; Wang, L. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  26. Li, L.; Ma, H. Saliency-Guided nonsubsampled shearlet transform for multisource remote sensing image fusion. Sensors 2021, 21, 1756. [Google Scholar] [CrossRef] [PubMed]
  27. Liu, Y.; Li, S.; Zhang, H. Multibaseline interferometric phase denoising based on kurtosis in the NSST domain. Sensors 2020, 20, 551. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Chen, P.; Zhang, Y.; Jia, Z.; Yang, J.; Kasabov, N. Remote sensing image change detection based on nsct-hmt model and its application. Sensors 2017, 17, 1295. [Google Scholar] [CrossRef] [Green Version]
  29. Li, S.; Fang, L.; Yin, H. Multitemporal image change detection using a detail-enhancing approach with nonsubsampled con-tourlet transform. IEEE Geosci. Remote. Sens. Lett. 2012, 9, 836–840. [Google Scholar]
  30. Zhou, W.; Jia, Z.; Yu, Y. SAR image change detection based on equal weight image fusion and adaptive threshold in the NSST domain. Eur. J. Remote. Sens. 2018, 51, 785–794. [Google Scholar]
  31. Shen, F.; Wang, Y.; Liu, C. Synthetic aperture radar image change detection based on Kalman filter and nonlocal means filter in the nonsubsampled shearlet transform domain. J. Appl. Remote Sens. 2020, 14, 016517. [Google Scholar] [CrossRef]
  32. Hou, B.; Wei, Q.; Zheng, Y. Unsupervised change detection in SAR image based on Gauss-log ratio image fusion and com-pressed projection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 3297–3317. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Wang, S.; Wang, C.; Li, J.; Zhang, H. SAR image change detection using saliency extraction and shearlet transform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4701–4710. [Google Scholar] [CrossRef]
  34. Gong, M.; Zhou, Z.; Ma, J. Change detection in synthetic aperture radar images based on image fusion and fuzzy clustering. IEEE Trans. Image Process. 2011, 21, 2141–2151. [Google Scholar] [CrossRef] [PubMed]
  35. Ma, J.; Gong, M.; Zhou, Z. Wavelet fusion on ratio images for change detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2012, 9, 1122–1126. [Google Scholar] [CrossRef]
  36. Geng, J.; Ma, X.; Zhou, X.; Wang, H. Saliency-Guided deep neural networks for SAR image change detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7365–7377. [Google Scholar] [CrossRef]
  37. Zheng, Y.; Jiao, L.; Liu, H.; Zhang, X.; Hou, B.; Wang, S. Unsupervised saliency-guided SAR image change detection. Pattern Recognit. 2017, 61, 309–326. [Google Scholar] [CrossRef]
  38. Wang, X.; Gao, F.; Dong, J. Change detection for synthetic aperture radar images based on pattern and intensity distinctiveness analysis. In Ninth International Conference on Graphic and Image Processing; International Society for Optics and Photonics: Bellingham, WA, USA, 2018; Volume 10615, p. 1061548. [Google Scholar]
  39. Majidi, M.; Ahmadi, S.; Shah-Hosseini, R. A saliency-guided neighbourhood ratio model for automatic change detection of SAR images. Int. J. Remote Sens. 2020, 41, 9606–9627. [Google Scholar] [CrossRef]
  40. Li, M.; Li, M.; Zhang, P. SAR image change detection using PCANet guided by saliency detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 402–406. [Google Scholar] [CrossRef]
  41. Moser, G.; Serpico, S. Generalized minimum-error thresholding for unsupervised change detection from SAR amplitude im-agery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2972–2982. [Google Scholar] [CrossRef]
  42. Celik, T.; Ma, K.-K. Multitemporal image change detection using undecimated discrete wavelet transform and active contours. IEEE Trans. Geosci. Remote Sens. 2010, 49, 706–716. [Google Scholar] [CrossRef]
  43. Hao, M.; Shi, W.; Deng, K.; Feng, Q. Superpixel-based active contour model for unsupervised change detection from satellite images. Int. J. Remote Sens. 2016, 37, 4276–4295. [Google Scholar] [CrossRef]
  44. Jia, L.; Li, M.; Zhang, P.; Wu, Y.; Zhu, H. SAR Image change detection based on multiple kernel k-means clustering with local-neighborhood information. IEEE Geosci. Remote Sens. Lett. 2016, 13, 856–860. [Google Scholar] [CrossRef]
  45. Shang, R.; Wen, A.; Liu, Y.; Jiao, L.; Stolkin, R. Synthetic aperture radar image change detection based on improved bilateral filtering and fuzzy C mean. J. Appl. Remote Sens. 2016, 10, 046017. [Google Scholar] [CrossRef]
  46. Li, Z.; Jia, Z.; Liu, L.; Yang, J.; Kasabov, N. A method to improve the accuracy of SAR image change detection by using an image enhancement method. ISPRS J. Photogramm. Remote Sens. 2020, 163, 137–151. [Google Scholar] [CrossRef]
  47. Zhang, X.; Liu, G.; Zhang, C.; Atkinson, P.M.; Tan, X.; Jian, X.; Zhou, X.; Li, Y. Two-Phase object-based deep learning for multi-temporal SAR image change detection. Remote Sens. 2020, 12, 548. [Google Scholar] [CrossRef] [Green Version]
  48. Gao, Y.; Gao, F.; Dong, J.; Wang, S. Change detection from synthetic aperture radar images based on channel weighting-based deep cascade network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4517–4529. [Google Scholar] [CrossRef]
  49. Chan, T.-H.; Jia, K.; Gao, S.; Lu, J.; Zeng, Z.; Ma, Y. PCANet: A simple deep learning baseline for image classification? IEEE Trans. Image Process. 2015, 24, 5017–5032. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Gao, F.; Dong, J.; Li, B.; Xu, Q. Automatic change detection in synthetic aperture radar images based on PCANet. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1792–1796. [Google Scholar] [CrossRef]
  51. Liu, T.; Li, Y.; Cao, Y.; Shen, Q. Change detection in multitemporal synthetic aperture radar images using dual-channel convolutional neural network. J. Appl. Remote Sens. 2017, 11, 1. [Google Scholar] [CrossRef] [Green Version]
  52. Duan, Y.; Liu, F.; Jiao, L.; Zhao, P.; Zhang, L. SAR Image segmentation based on convolutional-wavelet neural network and markov random field. Pattern Recognit. 2017, 64, 255–267. [Google Scholar] [CrossRef]
  53. Gao, F.; Wang, X.; Gao, Y.; Dong, J.; Wang, S. Sea ice change detection in sar images based on convolutional-wavelet neural networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1240–1244. [Google Scholar] [CrossRef]
  54. Ma, J.; Zhou, Z.; Wang, B. Infrared and visible image fusion based on visual saliency map and weighted least square optimi-zation. Infrared Phys. Technol. 2017, 82, 8–17. [Google Scholar] [CrossRef]
  55. Gao, F.; Dong, J.; Li, B. Change detection from synthetic aperture radar images based on neighborhood-based ratio and ex-treme learning machine. J. Appl. Remote. Sens. 2016, 10, 046019. [Google Scholar] [CrossRef]
  56. Goyal, A.; Meenpal, T. Patch-Based dual-tree complex wavelet transform for kinship recognition. IEEE Trans. Image Process. 2021, 30, 191–206. [Google Scholar] [CrossRef]
  57. Celik, T. Unsupervised change detection in satellite images using principal component analysis and k-means clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
  58. Sharma, A.; Gulati, T. Change detection in remotely sensed images based on modified log ratio and fuzzy clustering. Blockchain Technol. Innov. Bus. Process. 2017, 84, 412–419. [Google Scholar] [CrossRef]
  59. Li, H.-C.; Celik, T.; Longbotham, N.; Emery, W.J. Gabor Feature based unsupervised change detection of multitemporal SAR images based on two-level clustering. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2458–2462. [Google Scholar] [CrossRef]
  60. Sumaiya, M.N.; Kumari, R.S.S. Logarithmic mean-based thresholding for SAR image change detection. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1726–1728. [Google Scholar] [CrossRef]
  61. Gao, Y.; Gao, F.; Dong, J.; Wang, S. Sea ice change detection in SAR images based on collaborative representation. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 7320–7323. [Google Scholar]
  62. He, Y.; Jia, Z.; Yang, J.; Kasabov, N. Multispectral image change detection based on single-band slow feature analysis. Remote Sens. 2021, 13, 2969. [Google Scholar] [CrossRef]
  63. Yang, L.; Chen, Y.; Song, S.; Li, F.; Huang, G. Deep Siamese networks based change detection with remote sensing images. Remote Sens. 2021, 13, 3394. [Google Scholar] [CrossRef]
  64. Shao, P.; Shi, W.; Liu, Z.; Dong, T. Unsupervised change detection using fuzzy topology-based majority voting. Remote Sens. 2021, 13, 3171. [Google Scholar] [CrossRef]
  65. Wu, T.; Luo, J.; Fang, J.; Ma, J.; Song, X. Unsupervised object-based change detection via a weibull mixture model-based binarization for high-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 2018, 15, 63–67. [Google Scholar] [CrossRef]
  66. Wu, T.; Luo, J.; Zhou, Y.; Wang, C.; Xi, J.; Fang, J. Geo-object-based land cover map update for high-spatial-resolution remote sensing images via change detection and label transfer. Remote Sens. 2020, 12, 174. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.