1. Introduction
SAR is a technology that uses active sensors to obtain high-resolution data from the Earth’s surface, by sending microwave signals to the target area and then processing the backscattered signal to generate images from the surface [
1]. One of the main advantages of using SAR imagery is the ability to obtain data day and night despite the environmental conditions since the microwave signal can be measured through clouds, smoke, fog, rain, snow and even penetrate the ground [
2]. One of the main disadvantages of the SAR technology is the speckle, a granular pattern present in all these types of images, that limits the availability of labeled datasets with ground-truth images to train supervised artificial intelligence models [
3], or to assessing the quality of the filtering process by comparing two images (ground truth and filtered). So, many authors use synthetically corrupted optical images with speckles and use the optical images as ground truth and the corrupted images as simulated SAR to be filtered. For this reason, it is very common that different publications assess the quality of the despeckling processes by using the optical reference, but when they are tested on actual SAR images there is no ground truth reference and the metrics and measurements cannot be calculated [
4,
5].
To quantitatively evaluate the despeckling process in SAR images, various metrics are utilized. One such metric is the Equivalent Number of Looks (ENL), which measures the noise level in a single image and is calculated as the square of the mean pixel value in a homogeneous region divided by the square of the standard deviation in the same region. The Mean Squared Error (MSE) assesses the average squared differences between two images’ pixel values, with a lower MSE indicating higher similarity [
6]. The Structural Similarity Index (SSIM) evaluates the similarity between two images in terms of luminance, contrast, and structure, with values close to 1 indicating high structural similarity. The Peak Signal-to-Noise Ratio (PSNR) quantifies the reconstruction quality of lossy compression codecs; higher PSNR values denote better filtering quality [
7]. Pratt’s Figure of Merit (PFOM) measures the accuracy of edge detection by comparing real and extracted edges in two images, with values close to 1 indicating perfect similarity in edge detection [
8]. The correlation measure proposed in [
9] and used in [
10] is a comprehensive indicator of despeckling quality since it uses edge detection to measure the correlation between two images. This
edge estimator was adapted in [
11] by including a radio edge detector to measure the geometric content remaining in a ratio image, called the
-Ratio estimator. The
-estimator in [
12] proposes that an evaluation based on a statistical measure of the quality of the remaining speckle is the first-order component of the quality measure.
In [
13] it is asserted that methodologies trained on synthetic datasets often exhibit subpar performance in practical applications, primarily attributable to the lack of clean SAR images. As a remedy, an innovative unsupervised despeckling technique, integrating online speckle generation with unpaired training, is introduced. In [
14] the construction of a diverse and authentic dataset utilizing a multicategory generalized Gaussian coherent SAR simulator is described. A markedly distinct methodology has recently surfaced [
15,
16,
17], wherein an enhanced SAR dataset is derived through the multitemporal averaging of SAR images to achieve a more authentic reference image.
In [
18] a novel protocol was introduced. This protocol involves generating ground truth data through the multitemporal fusion of multiple images depicting the same scene. This approach facilitated the creation of a labeled dataset and the training of new Deep Learning (DL) models aimed at reducing speckle noise using actual SAR images. This opens a new possibility to assess the quality of despeckling techniques applied to these actual SAR images. The main advantage of this protocol is the design of a labeled dataset with ground truth reference, which was split to train and validate the DL despeckling model.
The despeckling of SAR images employs a variety of filters to reduce speckle noise while preserving important image features. The traditional Lee filter adjusts the number of neighboring pixels within the sliding window to improve performance [
19]. The Fast Adaptive Nonlocal SAR (FANS) filter evolves from SAR-BM3D, utilizing nonlocal filtering and wavelet-domain shrinkage, with modifications tailored to SAR image characteristics, achieving superior results in terms of signal-to-noise ratio and perceived image quality [
20]. Among AI-based filters, the Autoencoder (AE) compresses and decompresses images through convolutional layers, effectively reducing noise during this process [
18,
21]. The multi-objective network (MONet) incorporates a modified cost function with the Kullback–Leibler divergence, demonstrating improved performance in metrics such as SSIM, SNR, and MSE over various nonlocal filters [
15]. The Semantic Conditional U-Net (SCUNet) combines semantic segmentation and image denoising in a unified framework, using an encoder-decoder structure with semantic information conditioning to enhance feature extraction and noise reduction [
22].
The innovative aspect of this paper lies in its comprehensive evaluation of traditional and AI-based despeckling filters applied to both synthetically corrupted optical images and actual SAR images using a novel method involving ratio images to measure residual speckles. Unlike prior studies that often rely solely on synthetically corrupted data or multitemporal fusion for ground truth, this work introduces a divergence measurement approach to quantitatively and visually assess the similarity between initial and remaining speckle. This method provides a more robust framework for evaluating the effectiveness of despeckling techniques, particularly highlighting the flexibility and limitations of AI-based models when applied to real-world SAR imagery. Additionally, the study underscores the importance of gamma distribution approximation in understanding the behavior of residual speckles, offering new insights and tools for future research in SAR image despeckling.
This paper is organized as follows. In
Section 2, the materials are presented, including the datasets used and the description of the metrics to assess the quality of the despeckling process. In
Section 3, the Speckle model and the proposed method to assess the despeckling processes are explained. In
Section 4 the experiments and measurements are presented, including all the processing steps, both on actual SAR and optical images. In
Section 5 a discussion and interpretation of the results is presented. Finally, in
Section 6 some relevant conclusions and future work are outlined.
5. Discussion
5.1. Optical Imagery
All the despeckling filters increased the ENL compared to the noisy image. The FANS filter exceeded the ENL observed in the ground truth images (
Table 1), which is undesirable as it indicates an over-smoothing effect, which can be confirmed in the third row of
Figure 3. The other filters remained within adequate limits, with the exception of SCUNet, which did not perform adequately in the optical dataset. It must be noticed that the FANS algorithm requires the number of looks as an input parameter when executed, so this could explain the adaptability that DL-based models lack.
The measurements of the MSE in
Table 2 represent the difference between the filtered images and the ground truth, with lower values indicating higher similarity. FANS consistently achieved the lowest MSE across most samples, demonstrating its effectiveness in preserving image details while reducing noise. SCUNet, conversely, showed higher MSE values, suggesting it was less effective in maintaining image fidelity.
SSIM (
Table 3) values range from 0 to 1, with higher values indicating greater similarity to the ground truth. FANS again performed the best, achieving the highest SSIM values across most samples. This indicates that FANS not only reduced noise effectively but also preserved the structural information of the images. SCUNet exhibited lower SSIM values, highlighting its comparative inefficiency in maintaining structural integrity during noise reduction.
PSNR (
Table 4) values indicate the quality of the filtered images, with higher values reflecting better image fidelity. FANS achieved the highest PSNR values across most samples, indicating its superior performance in noise reduction and image quality preservation. In contrast, SCUNet showed lower PSNR values, reinforcing the observations from the MSE and SSIM metrics regarding its lower effectiveness.
PFOM (
Table 6) values assess the filters’ performance in preserving image features while reducing noise. Higher PFOM values indicate better preservation of image features. The AE filter achieved the highest PFOM values in most samples, suggesting its excellent performance in feature preservation. FANS also performed well, consistently achieving high PFOM values, while SCUNet again showed relatively lower PFOM values, indicating less effective feature preservation.
The measurements of the MSE (
Table 2), SSIM (
Table 3), and PSNR (
Table 4) confirm the previous analysis, with FANS being the best in these metrics with optical images. The preservation of edges indicated by the PFOM (
Table 6) shows the AE has the best results in three out of the five samples.
In
Table 7, it is observed that the Jensen–Shannon Divergence (JSD) index of the Lee method performed the best in images 1, 3, and 4. However, the Lee method did not exhibit significant performance in terms of speckle suppression or edge preservation when compared to other filters. This discrepancy can be explained by the inherent characteristics of the JSD index and the nature of the Lee filter.
The JSD index focuses on the statistical preservation of noise characteristics rather than visual quality. The Lee filter may excel in maintaining the statistical properties of the speckle noise, resulting in low JSD values. However, this does not necessarily translate to effective speckle suppression or edge preservation, which are critical for visual quality and structural integrity.
The Lee filter’s performance is particularly strong in homogeneous regions where it can accurately preserve the local statistics. In such areas, the ratio image’s speckle remains close to the theoretical distribution, resulting in a lower JSD. This might explain its superior JSD performance in images 1, 3, and 4 if these images contain significant homogeneous regions. The Lee filter tends to blur edges and fine details because it averages pixel values within the local window. This limitation affects its performance in edge preservation and visual quality metrics like beta and PFOM, despite its ability to preserve the statistical properties of the speckle.
5.2. Actual SAR Imagery
Similarly to the optical case, all despeckling filters applied to the SAR dataset increased the ENL compared to the noisy image (
Table 8). Once again, the FANS filter exhibited the highest increase in ENL, indicating an over-smoothing effect, as demonstrated in the third row of
Figure 6. SCUNet struggled to adapt to the noise characteristics in these images, suggesting it was trained on a substantially different dataset, resulting in a rigid despeckling process.
Other metrics, such as MSE, SSIM, PSNR, beta and PFOM (
Table 9,
Table 10,
Table 11,
Table 12 and
Table 13, respectively), consistently indicate that the AE model outperforms other filters. This superior performance is attributable to the AE model’s ability to accurately represent the characteristics of speckle noise, as it was trained on a dataset comprising actual SAR images.
Table 9 displays MSE measures the difference between filtered images and ground truth. The AE filter consistently achieved the lowest MSE values across all samples, suggesting superior accuracy in noise reduction. Conversely, SCUNet had the highest MSE values, indicating less effective performance. The low MSE of AE suggests that this filter is highly accurate in approximating ground truth, surpassing other filters in precision. The relationship between AE and MONET is notable, as both show low MSE values, indicating they are good at preserving important details while reducing noise.
Table 10 presents SSIM measures structural similarity between filtered images and ground truth. The MONET filter achieved the highest SSIM scores for most samples, indicating a strong resemblance to ground truth. AE also performed well, with SSIM values close to those of MONET. The relationship between MONET and AE is particularly interesting, as both filters not only reduce noise but also preserve image structure. This suggests that both filters are effective not only in noise reduction but also in preserving crucial structural information of SAR images.
Table 11 displays PSNR measures image quality relative to ground truth. The AE filter consistently achieved the highest PSNR scores across all samples, indicating superior image quality. The high PSNR of AE reinforces its effectiveness in improving image quality, corroborating MSE and SSIM results. The relationship between AE and MONET again is notable, showing that both filters not only preserve image structure but also significantly improve overall image quality.
Table 13 presents PFOM indicates the degree of similarity between filtered images and ground truth. The AE filter consistently achieved the highest PFOM values, indicating the best performance. The consistency of AE in achieving the highest PFOM values suggests that it is the most effective filter in terms of improving overall image quality compared to ground truth. The relationship between AE and MONET shows that both filters are effective in improving image quality and similarity to ground truth.
5.3. Analysis of Gamma and Divergence in Ratio Images
In theory, ratio images should exhibit pure speckle under ideal conditions. However, after the application of a practical filter to a noisy image, patterns, structures and shapes persist within the resultant ratio image. The Gamma distribution obtained from the theoretical noise contrasted with the one obtained from the ratio image indeed exhibits disparities, measured through the JSD. This proposed methodology is a second-order measurement, that must be applied only when first-order evaluations are performed and the despeckling quality is found to be suitable. Scenarios such as in
Figure 7 where the last row corresponding to SCUNet exhibited the absence of structure as residual information corroborates this recommendation. This holds true despite the fact that the mean was not preserved and the metrics are suboptimal when compared to the other filters.
The ratio images JSD values observed for the optical dataset in
Figure 4 and
Figure 5, and for the SAR dataset in
Figure 7 and
Figure 8 exhibit behavior consistent with a Gamma distribution. These results highlight the expected differences between the analytically added and observed noise and the noise measured in the ratio images.
The JSD values documented in
Table 7,
Table 8,
Table 9,
Table 10,
Table 11,
Table 12,
Table 13 and
Table 14 align with observations derived from Gamma distributions. Certain filters, such as SCUNet, demonstrate challenges in accurately modeling the inherent randomness of the speckle. These results are further corroborated by the ratio images depicted in
Figure 4,
Figure 5,
Figure 6 and
Figure 7, where the images that display structures and shapes correspond to the ones with higher JSD values.
5.4. Evaluation of Methodology
The methodology employed in this study involves an evaluation of traditional and AI-based despeckling filters using both synthetically corrupted optical images and actual SAR images. This dual approach allows for a robust assessment of the filter performance across different types of data, which is crucial for understanding their practical applicability.
5.4.1. Analysis
The use of synthetic speckle on optical images provides a controlled environment to evaluate the filters, allowing for clear comparisons against known ground truth data. This method is appropriate as it enables the assessment of how well filters can remove speckle noise without sacrificing image details. The synthetic approach is particularly beneficial in initial testing phases, where controlled conditions are necessary to benchmark different techniques.
In addition to synthetic data, the study incorporates actual SAR images, ensuring that the findings are relevant to real-world scenarios. The multitemporal fusion technique used to generate ground truth references from SAR images is a notable strength of this methodology. This technique enhances the reliability of the ground truth by averaging multiple images of the same scene, which effectively reduces the noise inherent in individual SAR images. This approach is appropriate as it provides a realistic benchmark for evaluating despeckling filters in practical applications.
However, there are potential limitations to this methodology. Firstly, while the synthetic speckle approach allows for controlled comparisons, it may not capture all the complexities and variabilities present in actual SAR images. Real-world SAR data often contain additional noise sources and artifacts that are not fully replicated in synthetic scenarios. This limitation suggests that while synthetic tests are useful, they should be complemented by extensive testing on actual SAR data.
Another limitation is the reliance on multitemporal fusion to create ground truth references for SAR images. This method assumes that the scene remains unchanged across the multiple images used for fusion. Any changes in the scene (e.g., due to temporal dynamics) could introduce inaccuracies in the ground truth, potentially affecting the evaluation of despeckling performance. Therefore, while this approach is robust for many applications, it may not be suitable for highly dynamic environments where the scene changes significantly over time.
Additionally, the evaluation metrics used in the study, such as ENL, MSE, SSIM, PSNR, beta and PFOM, provide valuable quantitative assessments of filter performance. However, these metrics have their own limitations and might not capture all aspects of the visual quality and usability of the despeckled images. Future research could benefit from incorporating subjective assessments or additional metrics that better reflect the perceptual quality of the images.
5.4.2. Appropriateness and Limitations of AI-Based Models
AI-based models have shown significant promise in this study. These models leverage large datasets to learn complex patterns in speckle noise and effectively remove it while preserving image details. The flexibility of these models allows them to adapt to different noise characteristics, which is a major advantage over traditional methods.
However, AI-based models also have limitations. Their performance heavily depends on the quality and diversity of the training data. If the training data do not adequately represent the variety of scenarios encountered in actual SAR applications, the models may not generalize well. Additionally, AI models require substantial computational resources for training and may be sensitive to the choice of hyperparameters.
5.5. Feedback on Analysis
The analysis conducted in this study provides an evaluation of the despeckling filters, utilizing both quantitative metrics and qualitative visual inspections. The inclusion of multiple metrics such as ENL, MSE, SSIM, PSNR, beta, and PFOM offers a well-rounded assessment of filter performance, while the ratio analysis and divergence measurement provide innovative methods for evaluating residual speckle.
5.5.1. Strengths of the Analysis
Comprehensive Metrics: The use of multiple metrics ensures that different aspects of filter performance are captured. Metrics like SSIM and PSNR evaluate image quality and structural similarity, while ENL focuses on noise reduction effectiveness.
Dual Data Sources: By employing both synthetically corrupted optical images and actual SAR images, the study covers a wide range of scenarios. This dual approach strengthens the validity of the findings, as it shows how filters perform in controlled and real-world conditions.
Ratio Analysis and Divergence Measurement: These innovative methods provide a deeper understanding of residual speckle and its distribution, offering a new perspective on despeckling filter evaluation. The use of Gamma distribution approximation and Jensen–Shannon Divergence (JSD) adds robustness to the analysis.
5.5.2. Areas for Improvement
Subjective quality assessment: While quantitative metrics are essential, incorporating subjective quality assessments through user studies could provide additional insights. Human perception of image quality often differs from what metrics capture, and subjective assessments can highlight aspects of filter performance that metrics might miss.
Temporal dynamics in SAR images: The current methodology assumes static scenes for multitemporal fusion. Future studies should explore methods to account for temporal dynamics in SAR images, ensuring that ground truth generation remains accurate even in changing environments. This could involve advanced techniques for detecting and compensating for changes in the scene.
Real-world application scenarios: While the study provides a solid foundation, further research should investigate the performance of despeckling filters in specific real-world applications. For example, how do these filters impact subsequent tasks such as object detection, classification, or change detection in SAR imagery? Evaluating filters in the context of these applications would provide practical insights into their utility.
5.6. Implications and Contributions to the Literature
The findings from this study have several important implications for the field of SAR image despeckling. By comprehensively evaluating both traditional and AI-based despeckling filters, the study provides a robust framework for understanding their strengths and limitations. The key implications of the findings are as follows:
Efficacy of AI-based filters: The study demonstrates that AI-based filters, such as Autoencoder (AE) and Multi-Objective Network (MONET), can achieve superior performance in noise reduction and image quality preservation compared to traditional filters. This highlights the potential of AI techniques in advancing despeckling methodologies and sets a new benchmark for future research.
Importance of dataset diversity: The performance of AI-based filters heavily depends on the quality and diversity of the training datasets. The findings underscore the necessity for diverse and comprehensive training datasets that adequately represent the variability in real-world SAR imagery. This has implications for the design and curation of training datasets in future studies.
Ratio analysis and divergence measurement: The innovative use of ratio images and divergence measurement provides a novel method for evaluating residual speckle. This approach offers a more detailed understanding of speckle behavior post-filtering and can be used as a supplementary evaluation metric in future despeckling research.
Flexibility of traditional filters: Traditional filters like FANS showed adaptability across different types of images, performing well in both optical and actual SAR datasets. This flexibility is valuable for applications where AI-based models may not be feasible due to computational constraints or a lack of suitable training data.
This study makes several significant contributions to the existing literature on despeckling filters:
Comprehensive evaluation framework: By integrating both traditional and AI-based filters and employing a wide range of evaluation metrics, the study provides a comprehensive framework for assessing despeckling techniques. This holistic approach can serve as a reference for future studies aiming to evaluate and compare despeckling filters.
Novel analytical techniques: The introduction of ratio analysis and divergence measurement as tools for evaluating residual speckle represents a novel contribution to the field. These techniques offer new insights into the behavior of despeckled images and can complement traditional evaluation metrics in providing a more nuanced understanding of filter performance.
Benchmarking AI-based models: The study benchmarks several state-of-the-art AI-based despeckling models, providing detailed performance evaluations. This contributes to the growing body of literature on the application of AI in SAR image processing and highlights the potential and challenges of these techniques.
Insights into speckle behavior: The findings provide deeper insights into the behavior of speckle noise in SAR imagery and the effectiveness of different despeckling techniques. This contributes to the theoretical understanding of speckle noise and offers practical guidance for researchers and practitioners working on SAR image analysis.
Guidance for future research: The study identifies key areas for improvement and future research, such as the need for diverse training datasets, the exploration of hybrid filtering approaches, and the development of advanced ground truth generation techniques. These insights can guide future research efforts in the field.
6. Conclusions and Future Work
The protocol to generate a ground truth from a multitemporal fusion of SAR images allowed to have a dataset containing ground truth images to serve as a reference when assessing the quality of the despeckling filters. In the case of optical images, the original ones served as a reference while the synthetically corrupted ones served as noisy pairs.
Across all evaluated metrics, FANS emerged as the most effective filtering method for reducing noise and preserving image fidelity and structural integrity in synthetically speckled optical images. SCUNet, while less effective across these metrics, highlights the variability in the performance of different filtering methods. The AE filter, although not always the best in noise reduction, demonstrated strong performance in preserving important image features. These results underscore the importance of selecting appropriate filtering methods based on the specific requirements of image processing tasks.
The results of various evaluation metrics show that filters AE and MONET excel in multiple aspects. AE leads in terms of MSE, PSNR, beta, and PFOM, demonstrating precision and quality in noise reduction. MONET, on the other hand, exhibits high preservation of image structure, achieving the best results in SSIM and performing well in PSNR and PFOM. The FANS filter demonstrates extreme effectiveness in noise reduction according to ENL but may not be the best for maintaining image quality in terms of SSIM and PSNR. SCUNet, though less effective compared to other filters, shows mixed performance that may be useful in specific contexts. These in-depth analyses provide insight into the strengths and weaknesses of each filter, offering a solid foundation for selecting the most suitable filter based on specific SAR image processing needs.
The DL-based models are very efficient when applied on specific images, for instance, the AE outperforms actual SAR images, but it does not stand out when applied on optical ones, and the SCUNet was not able to attenuate the synthetic speckle added by using a Gamma model. On the other hand, the FANS filter could adapt to different levels of speckle and obtain better results in both, SAR and optical images.
The experiments conducted underscore the utility of ratio analysis in facilitating both visual inspection and quantitative evaluation of residual speckle within the filtered images. This innovative approach provides a comprehensive tool for understanding and assessing the efficacy of despeckling techniques in SAR images, enabling a direct comparison between the initial and residual speckle. By combining visual analysis and quantitative metrics, the study offers a more complete understanding of the effects of despeckling filters, allowing for a more precise assessment of their performance across different conditions and types of images. This approach has significant implications for the ongoing improvement of despeckling techniques and their application in a variety of contexts.
The proposed method of analysis of ratio images and divergence measurement is in line with the results derived from the other well-known metrics also calculated in this paper for both types of images. This is also in concordance with the findings after visually inspecting the filtered images.
Future research can be focused on using other levels of SAR imagery, or SAR data obtained from other satellites like RadarSAT or TerraSAR-X, which will require the design of new datasets by using a methodology similar to the one used in this work. Also, it can be possible to include more metrics to assess the despeckling quality. Other works may focus on improving the ratio analysis based on the Gamma approximation proposed in this paper.
The methodology used in this study is robust and appropriate for evaluating the effectiveness of despeckling filters. While there are potential limitations, such as the synthetic nature of some tests and the assumptions in multitemporal fusion, the approach provides a comprehensive framework for assessing both traditional and AI-based despeckling techniques. Future research should aim to address these limitations by incorporating more diverse datasets and exploring additional evaluation metrics.
Future Work
Adapting AI models to diverse data: AI-based models should be trained on more diverse and comprehensive datasets to improve their generalization capabilities. Research could focus on transfer learning techniques to adapt models trained on one type of data to perform well on another.
Exploring hybrid approaches: Combining traditional and AI-based methods could leverage the strengths of both approaches. For instance, using traditional filters to preprocess data before applying AI models might improve overall performance.
Advanced ground truth generation techniques: Developing more sophisticated methods for ground truth generation in SAR imagery, especially in dynamic environments, would enhance the reliability of despeckling filter evaluations.
Real-time despeckling: Researching methods to optimize despeckling filters for real-time processing could expand their applicability in operational settings, where timely data analysis is critical.