1. Introduction
Remote sensing satellites acquire useful data for environmental monitoring, such as land cover and land use surveys, urban studies and hazard management, collecting images that provide a global and rapid response of the involved areas [
1,
2,
3]. In the traditional processing chain, acquired images are downlinked to the ground for analysis and distribution operations. However, the increasing size of data produced by the Earth observation satellites requires innovative processing techniques. For this reason, intelligent on-board data processing and data-reduction procedures are becoming essential in current and planned missions [
4]. In fact, the high-resolution instruments in use produce huge sensor data volumes, which demand high on-board data storage capacity and communication resources, i.e., bandwidth, for data downlinking to end users on the ground. In order to face both the memory and bandwidth availabilities, many future satellite missions are working with the purpose of moving data processing from the ground segment to the space segment through on-board processing operations. In this context, a particular focus regards data management in small satellites. Small satellites have become increasingly popular due to their small size, low power consumption, short development cycle and low cost. However, they are subject to limited on-board transmitting power and limited communication band choices, as well as restricted storage space, so a new processing flow is required rather than the traditional one [
5]. Therefore, having the possibility to directly deliver the concerned higher-level information contained in the remote sensing images becomes crucial; this is indeed usually many orders of magnitude smaller than the remote sensing data themselves. Moreover, the processed results will greatly improve the downlink data transmission efficiency, allowing for reduction in the transmission bandwidth and enhancing the transmission rate. Additionally, by selecting only the useful information, the on-board platform memory cost can be reduced [
6,
7].
On-board processing requires several characteristics, such as flexibility, robustness, low computation burden. For this reason and also because of their ability to deal with complex data, such as remote sensing data, artificial intelligence models such as neural networks (NNs) represent a valid instrument [
8,
9]. NN algorithms are characterized by a collection of linked units that are aggregated in layers; the strength of each connection is adjusted by specific weights, or the coefficients of the model, which are determined during the NN learning process. The learning phase represents the most expensive task, but this can be performed off-line, with the ground stations capabilities, without affecting on-board processing operations. Once the coefficients are known, the model, if not too demanding in terms of storage, can be implemented on-board for the desired analysis. In this context, the main interest is in auto-associative neural networks (AANNs). AANNs are symmetrical networks in which the targets used to train the network are simply the input vectors themselves so that the network is attempting to map each input vector into itself. The middle layer is a bottleneck layer that forces a compressed knowledge representation of the original input [
10]. For their characteristics, AANNs have already been used to perform dimensionality reduction and to extract relevant features from remote sensing images [
9,
11,
12]. Moreover, AANNs, which are shallow networks, require low-level computational resources; thus, they are particularly suited for on-board operations.
In this work, the AANN is used to develop a low computational resource algorithm able to detect changes occurring between two images taken over the same area at different times. In remote sensing, change detection analysis is an important application that allows users to monitor phenomena such as urbanization, land cover change and natural disaster [
13]. NNs are one of the methods used to perform change detection, along with other approaches that include mathematical operations, principal component analysis (PCA), and geographic information systems (GIS) [
14,
15]. Particularly, in [
16], NNs with an encoding-decoding structure were used to obtain an encoded representation of the original multi-temporal image pair. Then, image subtraction was performed to detect seasonal and permanent changes that occurred over the time period. Additionally, in [
17], a deep NN model was used to extract image features that were later used to build the input dataset for another NN trained to classify between changed and unchanged images. In this context, on-board change detection introduces two interesting advantages: it reduces the amount of data transmitted to the ground by sending only the changed images, and it provides products for rapid decision-making.
Some examples regarding advanced on-board processing operations, mainly based on NN algorithms, can be found in the literature: Yao et al. proposed an on-board ship detection scheme to achieve near real-time on-board processing by small satellites computing platform [
18], while Zhang et al. suggested a similar technique for cloud detection [
19]. In [
20], Del Rosso et al. implemented an algorithm based on convolutional neural networks for volcanic eruptions detection based on Sentinel-2 and Landsat-7 optical data, while Diana et al. used SAR images to identify oil spills through a CNN algorithm [
21]. A procedure based on self-organizing maps is proposed in [
22] by Danielsen et al. to cluster hyperspectral images on-board, and, in [
23], Rapuano et al. successfully designed a hardware accelerator based on field programmable gate arrays (FPGAs) specifically made for the space environment.
Despite this, and although change detection applications in the ground segment are very important and largely investigated [
24], literature regarding on-board change detection is still very sparse and limited. When this paper was written, the only significant reference found was the work of Spigai and Oller [
25], who proposed a change detection-classification module specific to aircrafts based on very high spatial resolution (0.2 m and finer) SAR (synthetic aperture radar) and optical aerial images. The detection of changes was basically performed relying on the computation of a correlation coefficient, and NNs were merely used to classify the detected changes (e.g., plane-not plane). Next, they applied a compression ratio to uninteresting image sections to reduce the amount of data transmitted to the ground.
The change detection procedure here proposed aims at detecting whatever alteration eventually occurred by inspecting a representation of the input image pairs in a reduced space, which is obtained by means of an AANN. This method is considered suitable for the on-board execution since it does not require high computational resources. In fact, power consumption of the algorithm, detailed and described at the end of
Section 3, is aligned with the resources that might be available on-board [
26]. After the learning phase is completed, the core of the procedure is to encode the input image patches through the AANN at time t
1 and time t
2 into smaller representative vectors. Then, a dissimilarity measured is performed, which drives the final decision between change or no change. The dataset used in this paper focuses on Sentinel-2 satellite data because, besides being freely available, they are characterized by high spatial resolution and revisit time, which make them a good test for the images that will be provided by upcoming satellite missions with similar characteristics [
27,
28].
Another technique is also considered in this study, which plays the role of a benchmark for the performance analysis of the presented methodology. In fact, the results obtained with the AANN compression have been compared with those obtained with a discrete wavelet transform (DWT) compression [
29]. The latter is indeed widely recognized as very effective for lossy image compression.
This paper is organized as follows:
Section 2 describes the dataset used and the methodology by defining the AANNs, the DWT and the working flow. This section also provides a description of the dimensionality reduction process, comparing the results obtained with the two considered techniques. The change detection results are presented in
Section 3 and discussed in
Section 4. Finally,
Section 5 describes the conclusion and future developments.
3. Results
Although the procedure previously described was separately applied to the three bands R, G and B, in this section only the results of the R band will be shown. Indeed, this band is recognized to have, in general, greater sensitivity to variations, and the results achieved using the other two bands were less accurate [
24].
In
Figure 11, two images extracted from the test set are shown.
Figure 11a
1,a
2 represent an area mainly covered by vegetation. The main changed elements regard the land cover with some variations in buildings and roads. The ground truth is represented at two levels: in
Figure 11b, it relies on the OSCD dataset, while, in
Figure 11c, it stems from the CVA technique explained in
Section 2.2. In both cases, the changed areas are shown in white. Similarly,
Figure 11d
1,d
2,e,f represent another pair of images with the associated ground truth, but, in this latter case, the images show mainly built-up areas.
3.1. Qualitative Assessment
We have seen that the methodology is applied at the patch level. For the performance analysis, the patches obtained from the ground truth images were therefore divided into three categories: patches without changes, patches with more than 10% of pixels changed (i.e., out of 1024 total pixels, 102 are changed) and patches with less than 10% of pixels changed. The percentage of 10% was chosen because it represents the average extension of the changed areas within the patch. This allowed for analysis of the performance, either on patches with significant changes or on patches with smaller changes. Moreover, it should be specified that patches reporting less than 1% of pixels as changed (i.e., out of 1024 total pixels, 10 are changed) were considered no change patches. This choice allows for discarding of the patches with cropped changed areas from the OSCD ground truth and of the patches with irrelevant alterations from the CVA change ground truth. For each pair of patches, the ED values were normalized in the 0–1 interval and were compared with their respective ground truth labelling.
In
Figure 12, the graphs report, for the test image pairs shown in
Figure 11 and for both the AANN and DWT approaches, the ED values on the y-axis, while the patch’s identifying number is reported on the
x-axis. Each point is then colored according to the corresponding OSCD ground truth: patches without changes are shown in blue, patches with more than 10% of pixels changed are in red and patches with less than 10% of pixels changed are in green. We see that for both methods, the PVRs associated with no change patches are mainly localized in the lower part of the graph, while the PVRs related to a high number of changed pixels are in the upper area. Few PVRs in the green class (changed pixels below 10%) report ED values near zero, but this never happens for examples belonging to the red class. The misclassified samples seem to grow with the DWT dimensionality reduction. The graphs also point out the different response of the two considered images. Test image number 2 (couple (d) in
Figure 11) performed better with both the AANN and DWT, and a clearer distinction between blue and red markers can be observed. On the contrary, test image number 1 (couple (a) in
Figure 11) shows some no change PVRs at high ED values. However, it has to be remembered that the OSCD ground truth does not consider natural changes, such as vegetation growth.
In an ideal setting, the ED values related to a no change condition should be the lowest and grow as the number of changed pixels increases. This trend is evaluated in
Figure 13, which reports the correlation between the number of changed pixels in a ground truth patch and the respective ED. As for
Figure 12, the graphs are shown for the AANN and DWT algorithms. Indeed, the presence of a linear correlation can be observed, especially for the AANN method. It also has to be noted that the linear trend is disturbed by the significant number of the no change patches, and this situation is enhanced in test image number 1.
The same analysis was repeated considering the ground reference obtained with the CVA method. The corresponding graphs are shown in
Figure 14 and
Figure 15. Here, the EDs related to a condition of significant change (markers in red) are distinctly separated from the unchanged ones, especially in test image 2.
Figure 15, compared to the previous case, shows a higher linear correlation between the number of changed pixels and the ED values. The performance is generally better with respect to the OSCD ground-truth. This could be expected because, in this case, all types of changes are considered in the ground truth, and this is more consistent with the sensitivity characteristics expressed by the ED measurement. However, the qualitative assessment carried out so far suggests that the PVR can be effective for the distinction among patches with relevant changes and patches with small or no changes.
3.2. Quantiative Performance Analysis
Typical metrics for image classification quantitative performance analysis are:
where:
TP = true positives, i.e., samples labelled as changed (red markers) and correctly predicted by the model;
TN = true negatives, i.e., samples labelled as not changed (blue markers) and correctly predicted by the model;
FP = false positives, i.e., samples labelled as not changed (blue markers) and wrongly predicted as changed by the model;
FN = false negatives, i.e., samples labelled as changed (red markers) and wrongly predicted as not changed by the model.
In the metrics computations, samples related to patches with less than 10% of pixels changed (i.e., green markers) were discarded.
As highlighted in the ground truth images, the number of changed pixels is significantly lower than the number of unchanged ones, and this class imbalance was considered during the results evaluation. In the case of an unbalanced dataset, accuracy is not useful, while the F1 score represents a proper measure as a combination between precision and recall. The receiver operating characteristic (ROC) curve is also a valid method for evaluating the performance of an algorithm. It plots the recall against the FPR, and the area under the curve (AUC) measures the performance of the model for all possible change thresholds. The AUC ranges from 0.5 for a random classifier to 1 for a perfect classifier.
Figure 16 and
Figure 17 show the ROC curve obtained by validating with OSCD and CVA ground truth images, respectively, for test images 1 and 2 in
Figure 11.
Figure 16 highlights that the lowest AUC value occurs for test image number 2, in the DWT case. With the AANN algorithm, the value improves by approximately 0.2. Test image number 1 exhibits better AUC values; the best is 0.909 for the AANN algorithm. By validating with the CVA (
Figure 17), the values increase, reaching almost 1 for test image 2 with the AANN algorithm, but the other cases also reach high AUC values.
Table 2 summarizes the results obtained for test image number 1, shown in
Figure 11(a
1,a
2), validated with the OSCD and CVA reference images. The results obtained with the AANN and DWT techniques were compared; the table also reports the AUC values. The change detection with AANN reduction performed better than DWT, as already demonstrated in
Figure 12 and
Figure 14. F1 achieves 0.86 with CVA validation and AANN reduction, which is more than 0.2 higher than DWT. The result also increases by the same value when validating with OSCD reference image, although, in this latter case, the best absolute value of F1 is 0.6. Recall represents an important metric to analyze, since it provides a measure of the number of missed changes, and these are the main focus of a change detection algorithm. Recall doubles by performing reduction with the AANN, with both OSCD and CVA validations.
However, in an operational scenario, the threshold value should be previously defined, in order for the algorithm to run automatically. The choice of an optimum general value can be a very difficult task because it definitely depends on the land cover contained in the image; a scene with a prevalence of urban areas would certainly provide an optimum threshold value different from that of a rural one. To address this issue, we selected from the dataset the two most similar images in terms of land cover: test image number 1 (shown in
Figure 11a
1,a
2) was compared to test image number 3 (shown in
Figure 3). Based on the recall scores, we selected from image number 1 a threshold value of 0.4 as the optimum one, and we used the same value to evaluate the change detection results on image 3. The final scores on test image 3 are reported in
Table 3 for both the OSCD and the CVA datasets and always consider the performance obtained with the DWT as a benchmark.
The main change elements in test image number 3 are seasonal land cover variations that are not labelled in the OSCD ground truth image: This factor affects the results that reported lower performance, as shown in
Table 3. The metrics rise sharply considering the comparison with CVA image, as shown in the table. In fact, the ground truth image obtained via the CVA was specifically made to point out these alterations. In this case, F1 reaches 0.92 with an AANN reduction.
Finally, it must be underlined that the developed algorithm is not heavy in terms of computational requirements. It was performed on NVIDIA® GeForce® GTX 1650 Max Q GPU with 4 GB RAM. During the test phase, the time taken to open a couple of satellite images with size of 772 × 902 pixels, divide them in patches, normalize their values and return their encoded representation was 9.8 s and required 25.79 MB of RAM. The training weights of the AANN encoder part require less than 3 MB for storage.
4. Discussion
The method proposed here aims to develop a change detection technique with a compressed representation of the two considered images. The smaller representation can be obtained by means of an AANN or DWT, but PVRs extracted with an AANN led to better results in change detection.
The results were validated by means of OSCD ground truth, but, as highlighted in
Section 2.1, these images report only man-made changes, leaving out the natural changes. This characteristic negatively impacts the results, moreso for test images showing a natural environment than those with an urban setting. Therefore, validation of images obtained by means of CVA were also considered. As described in
Section 2.2, this method depends on the analyst’s sensitivity but was demonstrated to be effective in this case, overcoming the critical issues related to OSCD ground truth.
Compared to the study described in [
25], our work proposed a method focused on detecting any type of change and was tested on Sentinel-2 images that were easy to access and use. Additionally, the detection was performed on a compressed representation of the original acquisition. In [
16], Kalinicheva et al. used NNs and Spot-5 images at 10 m of spatial resolution, although the algorithm can be applied to any type of satellite acquisitions. The changed areas were highlighted through an image pixel-based subtraction. In this case, compression of the image is not performed to reduce the amount of data but to apply the subtraction method. Indeed, subtraction works fine on lower resolution images, but the pixelwise analysis may not be compliant with on-board processing requirements. Zhao et al. [
17] used multi-temporal hyperspectral images as the input dataset and two NN models. The features extracted by the first network are considered additional input for the second NN, which performs the change detection. They achieved more than 95% accuracy, but the complexity of the scheme does not seem developed to be run on-board.
The testing phase led to promising results, as shown in the figures in
Section 3.1, although it revealed the criticality related to the definition of a unique threshold to automatically perform the change detection. In fact, this value strictly depends on the scene depicted in the image. For this reason, numerical results were evaluated for two images of the test set that are similar in terms of land cover. By validating with the CVA, the AANN reached higher score values than the OSCD ground truth, and the gap with the DWT algorithm is sharply marked for test image 3. Generally, the results obtained by validating with the CVA were significantly better than the ones obtained with the OSCD.
5. Conclusions
In a scenario where the demand for on-board processing to be run on space payloads is increasing more and more to limit both bandwidth and data storage, this work proposes a new method for on-board change detection based on NN. In particular, an AANN extracts the image features, which are then used for a CVA. AANNs are characterized by a very light architecture; hence, they are particularly suitable for on-board operations.
In this study, the change detection algorithm was applied to Sentinel-2 images. Moreover, for performance analysis purposes, a standard image compressor based on DWT was also considered as a benchmark. The algorithms perform dimensionality reduction in multi-temporal input image patches by reducing their initial size into smaller representative vectors. A dissimilarity measure, i.e., the ED, was computed to indicate the probabilities of change.
The main contributions of this work can be summarized in three aspects: First, it shows the capability of AANNs in image compression and reconstruction. As detailed in
Section 2.5, the AANN achieved better results in reconstructing images than other well-known techniques, such as the DWT. Second, while most of the current change detection approaches use deep learning schemes [
38], our algorithm provides an innovative shallow neural network technique using compressed representation. Lastly, the procedure presented here was designed and developed to be run on-board, and the on-board processing is currently another relevant topic. The computing resources and the memory required to store the training weights are comparable to those typically available on-board, as described in
Section 1. The results demonstrated that for change detection purposes, image encoding by means of AANNs led to better achievements than the DWT and required low-level computing resources in the test phase, which could then be executed on-board. The results improved up to 0.2 in terms of the F1 score. The classification accuracy between the classes change and no change is strictly affected by the input dataset and the validation images. Indeed, datasets for change detection analysis are typically characterized by class imbalance and limited availability of ground truth images. Despite this, a promising distinction between EDs related to areas reporting significant changes and those for no change areas was shown, reaching an F1 score of 0.9.
Further research will investigate the tailoring of the method for SAR data, since they allow the implementation of the algorithm in any weather condition and during the night.